hails commited on
Commit
a10e8cb
1 Parent(s): 335c19c

Upload config.json

Browse files
Files changed (1) hide show
  1. config.json +2 -1
config.json CHANGED
@@ -3,6 +3,7 @@
3
  "GPTNeoXForCausalLM"
4
  ],
5
  "bos_token_id": 0,
 
6
  "eos_token_id": 0,
7
  "hidden_act": "gelu",
8
  "hidden_size": 5120,
@@ -17,7 +18,7 @@
17
  "rotary_pct": 0.25,
18
  "tie_word_embeddings": false,
19
  "torch_dtype": "float16",
20
- "transformers_version": "4.24.0",
21
  "use_cache": true,
22
  "use_parallel_residual": true,
23
  "vocab_size": 50688
 
3
  "GPTNeoXForCausalLM"
4
  ],
5
  "bos_token_id": 0,
6
+ "classifier_dropout": 0.1,
7
  "eos_token_id": 0,
8
  "hidden_act": "gelu",
9
  "hidden_size": 5120,
 
18
  "rotary_pct": 0.25,
19
  "tie_word_embeddings": false,
20
  "torch_dtype": "float16",
21
+ "transformers_version": "4.30.2",
22
  "use_cache": true,
23
  "use_parallel_residual": true,
24
  "vocab_size": 50688