chuanli-lambda commited on
Commit
eaaea8f
1 Parent(s): 24d8fbe

Upload GPTNeoXForCausalLM

Browse files
Files changed (2) hide show
  1. config.json +5 -5
  2. pytorch_model.bin +1 -1
config.json CHANGED
@@ -1,19 +1,19 @@
1
  {
2
- "_name_or_path": "/home/ubuntu/llm/outputs/ft-synthetic-instruct-gptj-pairwise-pythia2.8b-deepspeed/resume/checkpoint-6000",
3
  "architectures": [
4
  "GPTNeoXForCausalLM"
5
  ],
6
  "bos_token_id": 0,
7
  "eos_token_id": 0,
8
  "hidden_act": "gelu",
9
- "hidden_size": 2560,
10
  "initializer_range": 0.02,
11
- "intermediate_size": 10240,
12
  "layer_norm_eps": 1e-05,
13
  "max_position_embeddings": 2048,
14
  "model_type": "gpt_neox",
15
- "num_attention_heads": 32,
16
- "num_hidden_layers": 32,
17
  "rotary_emb_base": 10000,
18
  "rotary_pct": 0.25,
19
  "tie_word_embeddings": false,
 
1
  {
2
+ "_name_or_path": "/home/ubuntu/llm/outputs/ft-synthetic-instruct-gptj-pairwise-pythia1.4b-deepspeed/resume/checkpoint-6000",
3
  "architectures": [
4
  "GPTNeoXForCausalLM"
5
  ],
6
  "bos_token_id": 0,
7
  "eos_token_id": 0,
8
  "hidden_act": "gelu",
9
+ "hidden_size": 2048,
10
  "initializer_range": 0.02,
11
+ "intermediate_size": 8192,
12
  "layer_norm_eps": 1e-05,
13
  "max_position_embeddings": 2048,
14
  "model_type": "gpt_neox",
15
+ "num_attention_heads": 16,
16
+ "num_hidden_layers": 24,
17
  "rotary_emb_base": 10000,
18
  "rotary_pct": 0.25,
19
  "tie_word_embeddings": false,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e073bcc7d5b4046e4713e10760203fd6251d0d1ec0408d7db346ea93ad11a835
3
  size 5758948897
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c43c55421c925482e536ce0a68ea7a3c550e89234cb434788861328a42e27943
3
  size 5758948897