Vipitis commited on
Commit
1aeea42
1 Parent(s): fb98268

Revert all changes

Browse files

back to where we started. might be more complicated than hoped :/

Files changed (1) hide show
  1. config.json +10 -7
config.json CHANGED
@@ -1,18 +1,21 @@
1
  {
2
- "_name_or_path": "bigcode/santacoder",
3
- "activation_function": "gelu_pytorch_tanh",
4
  "architectures": [
5
- "GPTBigCodeForCausalLM"
6
  ],
7
- "attention_softmax_in_fp32": false,
8
- "multi_query": true,
9
  "attn_pdrop": 0.1,
 
 
 
 
10
  "bos_token_id": 49152,
11
  "embd_pdrop": 0.1,
12
  "eos_token_id": 49152,
13
  "initializer_range": 0.02,
14
  "layer_norm_epsilon": 1e-05,
15
- "model_type": "gpt_bigcode",
16
  "n_embd": 2048,
17
  "n_head": 16,
18
  "n_inner": 8192,
@@ -20,7 +23,7 @@
20
  "n_positions": 2048,
21
  "reorder_and_upcast_attn": false,
22
  "resid_pdrop": 0.1,
23
- "scale_attention_softmax_in_fp32": false,
24
  "scale_attn_weights": true,
25
  "summary_activation": null,
26
  "summary_first_dropout": 0.1,
 
1
  {
2
+ "_name_or_path": "bigcode/santacoder",
3
+ "activation_function": "gelu_fast",
4
  "architectures": [
5
+ "GPT2LMHeadCustomModel"
6
  ],
7
+ "attention_head_type": "multiquery",
 
8
  "attn_pdrop": 0.1,
9
+ "auto_map": {
10
+ "AutoConfig": "configuration_gpt2_mq.GPT2CustomConfig",
11
+ "AutoModelForCausalLM": "modeling_gpt2_mq.GPT2LMHeadCustomModel"
12
+ },
13
  "bos_token_id": 49152,
14
  "embd_pdrop": 0.1,
15
  "eos_token_id": 49152,
16
  "initializer_range": 0.02,
17
  "layer_norm_epsilon": 1e-05,
18
+ "model_type": "gpt2",
19
  "n_embd": 2048,
20
  "n_head": 16,
21
  "n_inner": 8192,
 
23
  "n_positions": 2048,
24
  "reorder_and_upcast_attn": false,
25
  "resid_pdrop": 0.1,
26
+ "scale_attn_by_inverse_layer_idx": false,
27
  "scale_attn_weights": true,
28
  "summary_activation": null,
29
  "summary_first_dropout": 0.1,