alexmarques commited on
Commit
67a9b67
1 Parent(s): c8e3503

Upload folder using huggingface_hub

Browse files
Files changed (4) hide show
  1. config.json +3 -9
  2. generation_config.json +1 -1
  3. model.safetensors +1 -1
  4. recipe.yaml +2 -4
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "microsoft__Phi-3-mini-128k-instruct",
3
  "architectures": [
4
  "Phi3ForCausalLM"
5
  ],
@@ -38,13 +38,7 @@
38
  ],
39
  "kv_cache_scheme": null,
40
  "quant_method": "compressed-tensors",
41
- "quantization_status": "frozen",
42
- "sparsity_config": {
43
- "format": "dense",
44
- "global_sparsity": 1.2496503239511723,
45
- "registry_requires_subclass": false,
46
- "sparsity_structure": "unstructured"
47
- }
48
  },
49
  "embd_pdrop": 0.0,
50
  "eos_token_id": 32000,
@@ -168,7 +162,7 @@
168
  "sliding_window": 262144,
169
  "tie_word_embeddings": false,
170
  "torch_dtype": "bfloat16",
171
- "transformers_version": "4.42.3",
172
  "use_cache": true,
173
  "vocab_size": 32064
174
  }
 
1
  {
2
+ "_name_or_path": "/root/.cache/huggingface/hub/models--microsoft--Phi-3-mini-128k-instruct/snapshots/d548c233192db00165d842bf8edff054bb3212f8",
3
  "architectures": [
4
  "Phi3ForCausalLM"
5
  ],
 
38
  ],
39
  "kv_cache_scheme": null,
40
  "quant_method": "compressed-tensors",
41
+ "quantization_status": "frozen"
 
 
 
 
 
 
42
  },
43
  "embd_pdrop": 0.0,
44
  "eos_token_id": 32000,
 
162
  "sliding_window": 262144,
163
  "tie_word_embeddings": false,
164
  "torch_dtype": "bfloat16",
165
+ "transformers_version": "4.44.0",
166
  "use_cache": true,
167
  "vocab_size": 32064
168
  }
generation_config.json CHANGED
@@ -7,5 +7,5 @@
7
  32007
8
  ],
9
  "pad_token_id": 32000,
10
- "transformers_version": "4.42.3"
11
  }
 
7
  32007
8
  ],
9
  "pad_token_id": 32000,
10
+ "transformers_version": "4.44.0"
11
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a94c12daa3595c34fcda567abbd4550d5351b45c5e72d52e5bb4ffca94053065
3
  size 4020365960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21a4e5d388f3be579989e33516a3140e625c832f522f0bac935f6c94c72a2512
3
  size 4020365960
recipe.yaml CHANGED
@@ -4,7 +4,5 @@ quant_stage:
4
  sequential_update: false
5
  dampening_frac: 0.01
6
  ignore: [lm_head]
7
- config_groups:
8
- group_0:
9
- targets: [Linear]
10
- weights: {num_bits: 8, type: int, symmetric: true, strategy: channel}
 
4
  sequential_update: false
5
  dampening_frac: 0.01
6
  ignore: [lm_head]
7
+ scheme: W8A16
8
+ targets: Linear