bachephysicdun commited on
Commit
bbb831b
·
verified ·
1 Parent(s): 5c15a05

Training in progress, step 36

Browse files
config.json CHANGED
@@ -17,6 +17,7 @@
17
  "n_inner": null,
18
  "n_layer": 12,
19
  "n_positions": 1024,
 
20
  "reorder_and_upcast_attn": false,
21
  "resid_pdrop": 0.1,
22
  "scale_attn_by_inverse_layer_idx": false,
 
17
  "n_inner": null,
18
  "n_layer": 12,
19
  "n_positions": 1024,
20
+ "pad_token_id": 50256,
21
  "reorder_and_upcast_attn": false,
22
  "resid_pdrop": 0.1,
23
  "scale_attn_by_inverse_layer_idx": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f7ed3156f1a947f522bd4e90c9058fbc93efd5d5fc7f4bafb989d03c9ed13198
3
  size 497774208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e66f9ff0694ec30b618a9e8d10d53248ff3417772db830548736ae05b2149db4
3
  size 497774208
special_tokens_map.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "bos_token": "<|endoftext|>",
3
  "eos_token": "<|endoftext|>",
4
- "pad_token": "<p>",
5
  "unk_token": "<|endoftext|>"
6
  }
 
1
  {
2
  "bos_token": "<|endoftext|>",
3
  "eos_token": "<|endoftext|>",
4
+ "pad_token": "<|endoftext|>",
5
  "unk_token": "<|endoftext|>"
6
  }
tokenizer_config.json CHANGED
@@ -14,7 +14,7 @@
14
  "clean_up_tokenization_spaces": true,
15
  "eos_token": "<|endoftext|>",
16
  "model_max_length": 1024,
17
- "pad_token": "<p>",
18
  "tokenizer_class": "GPT2Tokenizer",
19
  "unk_token": "<|endoftext|>"
20
  }
 
14
  "clean_up_tokenization_spaces": true,
15
  "eos_token": "<|endoftext|>",
16
  "model_max_length": 1024,
17
+ "pad_token": "<|endoftext|>",
18
  "tokenizer_class": "GPT2Tokenizer",
19
  "unk_token": "<|endoftext|>"
20
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7982434d1b7eb8d81f97233732f427d1fa54949d8b5749aa8d706257817e5299
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dde9e26f29c05c06278a9aa275e453c30f7cf5703ad8f7f0ad319cff3e619e8f
3
  size 5432