Cedille commited on
Commit
e50edc4
1 Parent(s): 985d998

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +9 -4
config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "activation_function": "gelu_new",
3
  "architectures": [
4
  "GPTJForCausalLM"
@@ -11,26 +12,30 @@
11
  "initializer_range": 0.02,
12
  "layer_norm_epsilon": 1e-05,
13
  "model_type": "gptj",
 
14
  "n_embd": 4096,
15
  "n_head": 16,
 
16
  "n_layer": 28,
17
  "n_positions": 2048,
 
18
  "rotary_dim": 64,
 
19
  "summary_activation": null,
20
  "summary_first_dropout": 0.1,
21
  "summary_proj_to_labels": true,
22
  "summary_type": "cls_index",
23
  "summary_use_proj": true,
24
- "transformers_version": "4.10.0.dev0",
25
- "tokenizer_class": "GPT2Tokenizer",
26
  "task_specific_params": {
27
  "text-generation": {
28
  "do_sample": true,
29
- "temperature": 1.0,
30
- "max_length": 50
31
  }
32
  },
 
33
  "torch_dtype": "float32",
 
34
  "use_cache": true,
35
  "vocab_size": 50400
36
  }
 
1
  {
2
+ "_name_or_path": ".",
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPTJForCausalLM"
 
12
  "initializer_range": 0.02,
13
  "layer_norm_epsilon": 1e-05,
14
  "model_type": "gptj",
15
+ "n_ctx": 2048,
16
  "n_embd": 4096,
17
  "n_head": 16,
18
+ "n_inner": null,
19
  "n_layer": 28,
20
  "n_positions": 2048,
21
+ "resid_pdrop": 0.0,
22
  "rotary_dim": 64,
23
+ "scale_attn_weights": true,
24
  "summary_activation": null,
25
  "summary_first_dropout": 0.1,
26
  "summary_proj_to_labels": true,
27
  "summary_type": "cls_index",
28
  "summary_use_proj": true,
 
 
29
  "task_specific_params": {
30
  "text-generation": {
31
  "do_sample": true,
32
+ "max_length": 50,
33
+ "temperature": 1.0
34
  }
35
  },
36
+ "tokenizer_class": "GPT2Tokenizer",
37
  "torch_dtype": "float32",
38
+ "transformers_version": "4.12.5",
39
  "use_cache": true,
40
  "vocab_size": 50400
41
  }