Falah commited on
Commit
a26fc0a
1 Parent(s): c2c316c
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[PAD]": 50257
3
+ }
config.json CHANGED
@@ -1,6 +1,5 @@
1
  {
2
- "_name_or_path": "distilgpt2",
3
- "_num_labels": 1,
4
  "activation_function": "gelu_new",
5
  "architectures": [
6
  "GPT2LMHeadModel"
@@ -9,21 +8,16 @@
9
  "bos_token_id": 50256,
10
  "embd_pdrop": 0.1,
11
  "eos_token_id": 50256,
12
- "id2label": {
13
- "0": "LABEL_0"
14
- },
15
  "initializer_range": 0.02,
16
- "label2id": {
17
- "LABEL_0": 0
18
- },
19
  "layer_norm_epsilon": 1e-05,
20
  "model_type": "gpt2",
21
  "n_ctx": 1024,
22
  "n_embd": 768,
23
  "n_head": 12,
24
  "n_inner": null,
25
- "n_layer": 6,
26
  "n_positions": 1024,
 
27
  "reorder_and_upcast_attn": false,
28
  "resid_pdrop": 0.1,
29
  "scale_attn_by_inverse_layer_idx": false,
 
1
  {
2
+ "_name_or_path": "gpt2",
 
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPT2LMHeadModel"
 
8
  "bos_token_id": 50256,
9
  "embd_pdrop": 0.1,
10
  "eos_token_id": 50256,
 
 
 
11
  "initializer_range": 0.02,
 
 
 
12
  "layer_norm_epsilon": 1e-05,
13
  "model_type": "gpt2",
14
  "n_ctx": 1024,
15
  "n_embd": 768,
16
  "n_head": 12,
17
  "n_inner": null,
18
+ "n_layer": 12,
19
  "n_positions": 1024,
20
+ "pad_token_id": 50256,
21
  "reorder_and_upcast_attn": false,
22
  "resid_pdrop": 0.1,
23
  "scale_attn_by_inverse_layer_idx": false,
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68f20421b5bfedab96ca2878bae6eb6edbe314ad48871b64d42838b7367dc71e
3
- size 655345093
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ddae377236a550bcd515d86b1f66e89edd80625045039104de02c73c2cb49da
3
+ size 995600901
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63d4a53c61fccbcb6e100c9a9dcbb3fbaff02e93f7823e70dba2d5225135381a
3
- size 333970169
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33dddbb7e8fec54387ffec7cd60eb2b1ebc80654ccd17a024c300c0ae3432c0e
3
+ size 510395581
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63bec005c023ab76f8a4cdb31c58474bf1b2bd45a410edad4663a2583b0d1cc1
3
- size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49dc98bb814bff9cc71afc4fa818112eb8c7a7d7f3476e261affbfeb1f41d2fc
3
+ size 13553
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b2a1447b37d4c2cbb42838ba63130a41ff9e79c581e529d3f8e2efe010a4ef0
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2350d076c30f249d9133ba20b2ce1b6f0798c07ab6eeae44685b28b0e30d03d
3
  size 627
special_tokens_map.json CHANGED
@@ -1,6 +1,24 @@
1
  {
2
- "bos_token": "<|endoftext|>",
3
- "eos_token": "<|endoftext|>",
4
- "pad_token": "<|endoftext|>",
5
- "unk_token": "<|endoftext|>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<|endoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "[PAD]",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
  }
tokenizer_config.json CHANGED
@@ -1,9 +1,33 @@
1
  {
 
2
  "add_prefix_space": false,
3
- "bos_token": "<|endoftext|>",
4
- "eos_token": "<|endoftext|>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "model_max_length": 1024,
 
6
  "special_tokens_map_file": null,
7
  "tokenizer_class": "GPT2Tokenizer",
8
- "unk_token": "<|endoftext|>"
 
 
 
 
 
 
 
9
  }
 
1
  {
2
+ "add_bos_token": false,
3
  "add_prefix_space": false,
4
+ "bos_token": {
5
+ "__type": "AddedToken",
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "eos_token": {
13
+ "__type": "AddedToken",
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "errors": "replace",
21
  "model_max_length": 1024,
22
+ "pad_token": null,
23
  "special_tokens_map_file": null,
24
  "tokenizer_class": "GPT2Tokenizer",
25
+ "unk_token": {
26
+ "__type": "AddedToken",
27
+ "content": "<|endoftext|>",
28
+ "lstrip": false,
29
+ "normalized": true,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ }
33
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:247870866df2f56c40d28a5a430019d354323cd203a93f1314acb13eff0d747b
3
  size 3579
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00d5ca618274f2126858ad7cbca1820d2d3104fa4836806661017227020b62b4
3
  size 3579
vocab.json CHANGED
The diff for this file is too large to render. See raw diff