Fix bos and pad token id defaults
Browse files- configuration_rnd.py +1 -1
configuration_rnd.py
CHANGED
@@ -17,7 +17,6 @@ from transformers.configuration_utils import PretrainedConfig
|
|
17 |
CONFIG_DEFAULTS = {
|
18 |
"attention_bias": False,
|
19 |
"attention_dropout": 0.0,
|
20 |
-
"bos_token_id": 151643,
|
21 |
"decoder_sparse_step": 1,
|
22 |
"eos_token_id": 151645,
|
23 |
"head_dim": 128,
|
@@ -36,6 +35,7 @@ CONFIG_DEFAULTS = {
|
|
36 |
"num_hidden_layers": 48,
|
37 |
"num_key_value_heads": 4,
|
38 |
"output_router_logits": False,
|
|
|
39 |
"rms_norm_eps": 1e-06,
|
40 |
"rope_scaling": False,
|
41 |
"rope_theta": 1000000.0,
|
|
|
17 |
CONFIG_DEFAULTS = {
|
18 |
"attention_bias": False,
|
19 |
"attention_dropout": 0.0,
|
|
|
20 |
"decoder_sparse_step": 1,
|
21 |
"eos_token_id": 151645,
|
22 |
"head_dim": 128,
|
|
|
35 |
"num_hidden_layers": 48,
|
36 |
"num_key_value_heads": 4,
|
37 |
"output_router_logits": False,
|
38 |
+
"pad_token_id": 151643,
|
39 |
"rms_norm_eps": 1e-06,
|
40 |
"rope_scaling": False,
|
41 |
"rope_theta": 1000000.0,
|