jeromeku commited on
Commit
750ddf0
·
verified ·
1 Parent(s): 887308e

Fix bos and pad token id defaults

Browse files
Files changed (1) hide show
  1. configuration_rnd.py +1 -1
configuration_rnd.py CHANGED
@@ -17,7 +17,6 @@ from transformers.configuration_utils import PretrainedConfig
17
  CONFIG_DEFAULTS = {
18
  "attention_bias": False,
19
  "attention_dropout": 0.0,
20
- "bos_token_id": 151643,
21
  "decoder_sparse_step": 1,
22
  "eos_token_id": 151645,
23
  "head_dim": 128,
@@ -36,6 +35,7 @@ CONFIG_DEFAULTS = {
36
  "num_hidden_layers": 48,
37
  "num_key_value_heads": 4,
38
  "output_router_logits": False,
 
39
  "rms_norm_eps": 1e-06,
40
  "rope_scaling": False,
41
  "rope_theta": 1000000.0,
 
17
  CONFIG_DEFAULTS = {
18
  "attention_bias": False,
19
  "attention_dropout": 0.0,
 
20
  "decoder_sparse_step": 1,
21
  "eos_token_id": 151645,
22
  "head_dim": 128,
 
35
  "num_hidden_layers": 48,
36
  "num_key_value_heads": 4,
37
  "output_router_logits": False,
38
+ "pad_token_id": 151643,
39
  "rms_norm_eps": 1e-06,
40
  "rope_scaling": False,
41
  "rope_theta": 1000000.0,