From 3d2800b8a7573788308668b71e2eadf27db4b437 Mon Sep 17 00:00:00 2001 From: pglorio Date: Wed, 11 Sep 2024 01:57:20 +0000 Subject: [PATCH] circleci fixes --- src/transformers/models/zamba/configuration_zamba.py | 2 +- tests/models/zamba/test_modeling_zamba.py | 2 -- 2 files changed, 1 insertion(+), 3 deletions(-) diff --git a/src/transformers/models/zamba/configuration_zamba.py b/src/transformers/models/zamba/configuration_zamba.py index b741546b02f46c..5f0c153703b334 100644 --- a/src/transformers/models/zamba/configuration_zamba.py +++ b/src/transformers/models/zamba/configuration_zamba.py @@ -135,7 +135,7 @@ def __init__( rms_norm_eps=1e-5, use_cache=True, num_logits_to_keep=1, - pad_token_id=None, + pad_token_id=0, bos_token_id=1, eos_token_id=2, max_position_embeddings=4096, diff --git a/tests/models/zamba/test_modeling_zamba.py b/tests/models/zamba/test_modeling_zamba.py index 086a651aa0838f..b1931254f58abd 100644 --- a/tests/models/zamba/test_modeling_zamba.py +++ b/tests/models/zamba/test_modeling_zamba.py @@ -646,8 +646,6 @@ def setUpClass(cls): model_id, torch_dtype=torch.bfloat16, low_cpu_mem_usage=True, use_mamba_kernels=False ) cls.tokenizer = AutoTokenizer.from_pretrained(model_id) - cls.tokenizer.add_special_tokens({"pad_token": "[PAD]"}) - cls.model.resize_token_embeddings(len(cls.tokenizer)) @slow def test_simple_generate(self):