Pratye commited on
Commit
b45ef2b
1 Parent(s): e7b29ae

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +1 -18
config.json CHANGED
@@ -13,23 +13,6 @@
13
  "q_proj",
14
  "v_prol"
15
  ],
16
- "task_type": "CAUSAL_LM",
17
- "architectures": ["LLaMAForCausalLM"],
18
- "bos_token_id": 0,
19
- "eos_token_id": 1,
20
- "hidden_act": "silu",
21
- "hidden_size": 4096,
22
- "intermediate_size": 11008,
23
- "initializer_range": 0.02,
24
- "max_sequence_length": 2048,
25
- "model_type": "llama",
26
- "num_attention_heads": 32,
27
- "num_hidden_layers": 32,
28
- "pad_token_id": -1,
29
- "rms_norm_eps": 1e-06,
30
- "torch_dtype": "float16",
31
- "transformers_version": "4.27.0.dev0",
32
- "use_cache": true,
33
- "vocab_size": 32000
34
  }
35
 
 
13
  "q_proj",
14
  "v_prol"
15
  ],
16
+ "task_type": "CAUSAL_LM"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17
  }
18