horiz94 commited on
Commit
0b67c52
1 Parent(s): 79c5cb9

eos token id modified

Browse files
Files changed (2) hide show
  1. config.json +2 -2
  2. configuration_flm.py +5 -5
config.json CHANGED
@@ -6,10 +6,10 @@
6
  "AutoModel": "modeling_flm.FLM",
7
  "AutoModelForCausalLM": "modeling_flm.FLM"
8
  },
9
- "bos_token_id": 100351,
10
  "cls_token_id": 100351,
11
  "embd_pdrop": 0.1,
12
- "eos_token_id": 100351,
13
  "initializer_range": 0.02,
14
  "input_mult": 1.0,
15
  "layer_norm_epsilon": 1e-05,
 
6
  "AutoModel": "modeling_flm.FLM",
7
  "AutoModelForCausalLM": "modeling_flm.FLM"
8
  },
9
+ "bos_token_id": 100256,
10
  "cls_token_id": 100351,
11
  "embd_pdrop": 0.1,
12
+ "eos_token_id": 100256,
13
  "initializer_range": 0.02,
14
  "input_mult": 1.0,
15
  "layer_norm_epsilon": 1e-05,
configuration_flm.py CHANGED
@@ -146,11 +146,11 @@ class FLMConfig(PretrainedConfig):
146
  summary_first_dropout=0.1,
147
  scale_attn_weights=True,
148
  use_cache=True,
149
- bos_token_id=50256,
150
- eos_token_id=50256,
151
- cls_token_id=50256,
152
- sep_token_id=50256,
153
- pad_token_id=50256,
154
  scale_attn_by_inverse_layer_idx=False,
155
  reorder_and_upcast_attn=False,
156
  relative_encoding=None,
 
146
  summary_first_dropout=0.1,
147
  scale_attn_weights=True,
148
  use_cache=True,
149
+ bos_token_id=None,
150
+ eos_token_id=None,
151
+ cls_token_id=None,
152
+ sep_token_id=None,
153
+ pad_token_id=None,
154
  scale_attn_by_inverse_layer_idx=False,
155
  reorder_and_upcast_attn=False,
156
  relative_encoding=None,