Upload folder using huggingface_hub
Browse files- config.json +3 -3
- generation_config.json +1 -1
- model.safetensors +2 -2
- recipe.yaml +2 -2
config.json
CHANGED
@@ -46,7 +46,7 @@
|
|
46 |
"quantization_status": "frozen",
|
47 |
"sparsity_config": {
|
48 |
"format": "dense",
|
49 |
-
"global_sparsity": 7.
|
50 |
"registry_requires_subclass": false,
|
51 |
"sparsity_structure": "unstructured"
|
52 |
}
|
@@ -66,8 +66,8 @@
|
|
66 |
"rope_scaling": null,
|
67 |
"rope_theta": 10000.0,
|
68 |
"tie_word_embeddings": false,
|
69 |
-
"torch_dtype": "
|
70 |
-
"transformers_version": "4.40.
|
71 |
"use_cache": true,
|
72 |
"vocab_size": 32000
|
73 |
}
|
|
|
46 |
"quantization_status": "frozen",
|
47 |
"sparsity_config": {
|
48 |
"format": "dense",
|
49 |
+
"global_sparsity": 7.81941733210164,
|
50 |
"registry_requires_subclass": false,
|
51 |
"sparsity_structure": "unstructured"
|
52 |
}
|
|
|
66 |
"rope_scaling": null,
|
67 |
"rope_theta": 10000.0,
|
68 |
"tie_word_embeddings": false,
|
69 |
+
"torch_dtype": "bfloat16",
|
70 |
+
"transformers_version": "4.40.2",
|
71 |
"use_cache": true,
|
72 |
"vocab_size": 32000
|
73 |
}
|
generation_config.json
CHANGED
@@ -3,5 +3,5 @@
|
|
3 |
"eos_token_id": 2,
|
4 |
"max_length": 2048,
|
5 |
"pad_token_id": 0,
|
6 |
-
"transformers_version": "4.40.
|
7 |
}
|
|
|
3 |
"eos_token_id": 2,
|
4 |
"max_length": 2048,
|
5 |
"pad_token_id": 0,
|
6 |
+
"transformers_version": "4.40.2"
|
7 |
}
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:166ff7480ffea0fab716c76f7865509f0cd0d2d3f998f148ba6960aeee610287
|
3 |
+
size 1231252716
|
recipe.yaml
CHANGED
@@ -5,6 +5,6 @@ quant_stage:
|
|
5 |
ignore: [lm_head]
|
6 |
config_groups:
|
7 |
group_0:
|
8 |
-
weights: {num_bits: 8, type: int, symmetric: true
|
9 |
-
input_activations: {num_bits: 8, type: int,
|
10 |
targets: [Linear]
|
|
|
5 |
ignore: [lm_head]
|
6 |
config_groups:
|
7 |
group_0:
|
8 |
+
weights: {num_bits: 8, type: int, symmetric: true}
|
9 |
+
input_activations: {num_bits: 8, type: int, dynamic: true, symmetric: true, strategy: token}
|
10 |
targets: [Linear]
|