deepparag commited on
Commit
5d483a6
1 Parent(s): 704340f
Files changed (2) hide show
  1. config.json +5 -33
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1,31 +1,11 @@
1
  {
2
- "_name_or_path": "EleutherAI/gpt-neo-2.7B",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPTNeoForCausalLM"
6
  ],
7
  "attention_dropout": 0,
8
  "attention_layers": [
9
- "global",
10
- "local",
11
- "global",
12
- "local",
13
- "global",
14
- "local",
15
- "global",
16
- "local",
17
- "global",
18
- "local",
19
- "global",
20
- "local",
21
- "global",
22
- "local",
23
- "global",
24
- "local",
25
- "global",
26
- "local",
27
- "global",
28
- "local",
29
  "global",
30
  "local",
31
  "global",
@@ -45,35 +25,27 @@
45
  "global",
46
  "local"
47
  ],
48
- 16
49
  ]
50
  ],
51
  "bos_token_id": 50256,
52
  "embed_dropout": 0,
53
  "eos_token_id": 50256,
54
  "gradient_checkpointing": false,
55
- "hidden_size": 2560,
56
  "initializer_range": 0.02,
57
  "intermediate_size": null,
58
  "layer_norm_epsilon": 1e-05,
59
  "max_position_embeddings": 2048,
60
  "model_type": "gpt_neo",
61
- "num_heads": 20,
62
- "num_layers": 32,
63
  "resid_dropout": 0,
64
  "summary_activation": null,
65
  "summary_first_dropout": 0.1,
66
  "summary_proj_to_labels": true,
67
  "summary_type": "cls_index",
68
  "summary_use_proj": true,
69
- "task_specific_params": {
70
- "text-generation": {
71
- "do_sample": true,
72
- "max_length": 50,
73
- "temperature": 0.9
74
- }
75
- },
76
- "tokenizer_class": "GPT2Tokenizer",
77
  "torch_dtype": "float32",
78
  "transformers_version": "4.14.1",
79
  "use_cache": true,
 
1
  {
2
+ "_name_or_path": "EleutherAI/gpt-neo-125M",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPTNeoForCausalLM"
6
  ],
7
  "attention_dropout": 0,
8
  "attention_layers": [
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  "global",
10
  "local",
11
  "global",
 
25
  "global",
26
  "local"
27
  ],
28
+ 6
29
  ]
30
  ],
31
  "bos_token_id": 50256,
32
  "embed_dropout": 0,
33
  "eos_token_id": 50256,
34
  "gradient_checkpointing": false,
35
+ "hidden_size": 768,
36
  "initializer_range": 0.02,
37
  "intermediate_size": null,
38
  "layer_norm_epsilon": 1e-05,
39
  "max_position_embeddings": 2048,
40
  "model_type": "gpt_neo",
41
+ "num_heads": 12,
42
+ "num_layers": 12,
43
  "resid_dropout": 0,
44
  "summary_activation": null,
45
  "summary_first_dropout": 0.1,
46
  "summary_proj_to_labels": true,
47
  "summary_type": "cls_index",
48
  "summary_use_proj": true,
 
 
 
 
 
 
 
 
49
  "torch_dtype": "float32",
50
  "transformers_version": "4.14.1",
51
  "use_cache": true,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:055b66a0b9805bab97fc993b5324d6cea6f690e58415594a7c52e67f84bd2f63
3
- size 10739628777
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8bfef9ef78d9921c26205adb9c8f2c1d940130438fbca3229f4c377216ca16d6
3
+ size 551192465