AlexWortega commited on
Commit
3b5b4b4
1 Parent(s): a001a21

Upload config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +14 -13
config.json CHANGED
@@ -1,20 +1,10 @@
1
  {
2
- "_name_or_path": "/extra_disk_1/blacksamorez/miqu-1-70b-AQLM-2Bit-1x16-hf/",
3
- "aqlm": {
4
- "in_group_size": 8,
5
- "nbits_per_codebook": 16,
6
- "num_codebooks": 1,
7
- "out_group_size": 1
8
- },
9
  "architectures": [
10
  "LlamaForCausalLM"
11
  ],
12
  "attention_bias": false,
13
  "attention_dropout": 0.0,
14
- "auto_map": {
15
- "AutoConfig": "configuration_llama_aqlm.LlamaConfig",
16
- "AutoModelForCausalLM": "modeling_llama_aqlm.LlamaForCausalLM"
17
- },
18
  "bos_token_id": 1,
19
  "eos_token_id": 2,
20
  "hidden_act": "silu",
@@ -22,18 +12,29 @@
22
  "initializer_range": 0.02,
23
  "intermediate_size": 28672,
24
  "max_position_embeddings": 32764,
25
- "model_type": "llama_aqlm",
26
  "num_attention_heads": 64,
27
  "num_hidden_layers": 80,
28
  "num_key_value_heads": 8,
29
  "pad_token_id": 0,
30
  "pretraining_tp": 1,
 
 
 
 
 
 
 
 
 
 
 
31
  "rms_norm_eps": 1e-05,
32
  "rope_scaling": null,
33
  "rope_theta": 1000000,
34
  "tie_word_embeddings": false,
35
  "torch_dtype": "float16",
36
- "transformers_version": "4.38.0.dev0",
37
  "use_cache": true,
38
  "vocab_size": 32000
39
  }
 
1
  {
2
+ "_name_or_path": "AlexWortega/miqu-1-70b-AQLM-2Bit-1x16-hf",
 
 
 
 
 
 
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
 
 
 
 
8
  "bos_token_id": 1,
9
  "eos_token_id": 2,
10
  "hidden_act": "silu",
 
12
  "initializer_range": 0.02,
13
  "intermediate_size": 28672,
14
  "max_position_embeddings": 32764,
15
+ "model_type": "llama",
16
  "num_attention_heads": 64,
17
  "num_hidden_layers": 80,
18
  "num_key_value_heads": 8,
19
  "pad_token_id": 0,
20
  "pretraining_tp": 1,
21
+ "quantization_config": {
22
+ "in_group_size": 8,
23
+ "linear_weights_not_to_quantize": [
24
+ "model.embed_tokens.weight",
25
+ "lm_head.weight"
26
+ ],
27
+ "nbits_per_codebook": 16,
28
+ "num_codebooks": 1,
29
+ "out_group_size": 1,
30
+ "quant_method": "aqlm"
31
+ },
32
  "rms_norm_eps": 1e-05,
33
  "rope_scaling": null,
34
  "rope_theta": 1000000,
35
  "tie_word_embeddings": false,
36
  "torch_dtype": "float16",
37
+ "transformers_version": "4.38.2",
38
  "use_cache": true,
39
  "vocab_size": 32000
40
  }