Xu-Ouyang commited on
Commit
39ea302
1 Parent(s): 86b0c8c

Upload GPTNeoXForCausalLM

Browse files
config.json CHANGED
@@ -19,22 +19,31 @@
19
  "num_attention_heads": 40,
20
  "num_hidden_layers": 36,
21
  "quantization_config": {
 
22
  "bits": 8,
 
23
  "damp_percent": 0.1,
24
  "dataset": "wikitext2",
25
  "desc_act": false,
 
26
  "group_size": 128,
27
- "modules_in_block_to_quantize": null,
 
 
 
 
 
28
  "quant_method": "gptq",
29
  "sym": true,
30
- "true_sequential": true
 
31
  },
32
  "rope_scaling": null,
33
  "rotary_emb_base": 10000,
34
  "rotary_pct": 0.25,
35
  "tie_word_embeddings": false,
36
  "torch_dtype": "float16",
37
- "transformers_version": "4.41.2",
38
  "use_cache": true,
39
  "use_parallel_residual": true,
40
  "vocab_size": 50688
 
19
  "num_attention_heads": 40,
20
  "num_hidden_layers": 36,
21
  "quantization_config": {
22
+ "batch_size": 1,
23
  "bits": 8,
24
+ "block_name_to_quantize": "gpt_neox.layers",
25
  "damp_percent": 0.1,
26
  "dataset": "wikitext2",
27
  "desc_act": false,
28
+ "disable_exllama": false,
29
  "group_size": 128,
30
+ "model_seqlen": 4096,
31
+ "module_name_preceding_first_block": [
32
+ "gpt_neox.embed_in",
33
+ "gpt_neox.emb_dropout"
34
+ ],
35
+ "pad_token_id": null,
36
  "quant_method": "gptq",
37
  "sym": true,
38
+ "true_sequential": true,
39
+ "use_cuda_fp16": true
40
  },
41
  "rope_scaling": null,
42
  "rotary_emb_base": 10000,
43
  "rotary_pct": 0.25,
44
  "tie_word_embeddings": false,
45
  "torch_dtype": "float16",
46
+ "transformers_version": "4.44.0",
47
  "use_cache": true,
48
  "use_parallel_residual": true,
49
  "vocab_size": 50688
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
- "transformers_version": "4.41.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
+ "transformers_version": "4.44.0"
6
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39084e556fb09da3eac3ffeec946656fdd4dd80a868c590044c50156f07d3459
3
  size 4922816920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcb28d26bc1863251c39f3df06b54c2a36e1ae11523ebacbb536ce44f6d73c2e
3
  size 4922816920
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8afc27d42b72eec62feb592d7241908dbbd45a582b6b5a5fc110e0e8743e4e0
3
  size 4967652888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f58242672c6101ce5ef4e935ebcd2fb8823e68dda076316186c826044d8f1668
3
  size 4967652888
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff3ece0dc8b0c78352f7a4649f346d4985b6135f8aed34c2b14dddc4c7218375
3
  size 2747738936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af139823fc15d06897a5b57b1d73e02464facfd091b21c4cf66e2e20232e7aea
3
  size 2747738936