Upload PhiForCausalLM
Browse files- config.json +1 -1
- model-00001-of-00003.safetensors +2 -2
- model-00002-of-00003.safetensors +2 -2
- model-00003-of-00003.safetensors +2 -2
- model.safetensors.index.json +5 -5
config.json
CHANGED
@@ -33,5 +33,5 @@
|
|
33 |
"torch_dtype": "float32",
|
34 |
"transformers_version": "4.39.1",
|
35 |
"use_cache": true,
|
36 |
-
"vocab_size":
|
37 |
}
|
|
|
33 |
"torch_dtype": "float32",
|
34 |
"transformers_version": "4.39.1",
|
35 |
"use_cache": true,
|
36 |
+
"vocab_size": 80979
|
37 |
}
|
model-00001-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:be10fcf4ff5c4ed03ff0cccde11177d35aabbf0fbff0b293fbbd17688a970f98
|
3 |
+
size 4998830376
|
model-00002-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:00b5645eb87e4496c1094919b5808ed2600afdabda06b71ce5ad6f2de689f717
|
3 |
+
size 4982541984
|
model-00003-of-00003.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5b114bea52001937245908f743bd55ce63d716a8e232637c2fb9257866eb35d9
|
3 |
+
size 1747405836
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"lm_head.bias": "model-00003-of-00003.safetensors",
|
@@ -84,8 +84,8 @@
|
|
84 |
"model.layers.13.mlp.fc1.weight": "model-00002-of-00003.safetensors",
|
85 |
"model.layers.13.mlp.fc2.bias": "model-00002-of-00003.safetensors",
|
86 |
"model.layers.13.mlp.fc2.weight": "model-00002-of-00003.safetensors",
|
87 |
-
"model.layers.13.self_attn.dense.bias": "model-
|
88 |
-
"model.layers.13.self_attn.dense.weight": "model-
|
89 |
"model.layers.13.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
|
90 |
"model.layers.13.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
91 |
"model.layers.13.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
|
@@ -324,8 +324,8 @@
|
|
324 |
"model.layers.29.mlp.fc2.weight": "model-00003-of-00003.safetensors",
|
325 |
"model.layers.29.self_attn.dense.bias": "model-00003-of-00003.safetensors",
|
326 |
"model.layers.29.self_attn.dense.weight": "model-00003-of-00003.safetensors",
|
327 |
-
"model.layers.29.self_attn.k_proj.bias": "model-
|
328 |
-
"model.layers.29.self_attn.k_proj.weight": "model-
|
329 |
"model.layers.29.self_attn.q_proj.bias": "model-00002-of-00003.safetensors",
|
330 |
"model.layers.29.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
331 |
"model.layers.29.self_attn.v_proj.bias": "model-00003-of-00003.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 11728728396
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"lm_head.bias": "model-00003-of-00003.safetensors",
|
|
|
84 |
"model.layers.13.mlp.fc1.weight": "model-00002-of-00003.safetensors",
|
85 |
"model.layers.13.mlp.fc2.bias": "model-00002-of-00003.safetensors",
|
86 |
"model.layers.13.mlp.fc2.weight": "model-00002-of-00003.safetensors",
|
87 |
+
"model.layers.13.self_attn.dense.bias": "model-00002-of-00003.safetensors",
|
88 |
+
"model.layers.13.self_attn.dense.weight": "model-00002-of-00003.safetensors",
|
89 |
"model.layers.13.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
|
90 |
"model.layers.13.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
|
91 |
"model.layers.13.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
|
|
|
324 |
"model.layers.29.mlp.fc2.weight": "model-00003-of-00003.safetensors",
|
325 |
"model.layers.29.self_attn.dense.bias": "model-00003-of-00003.safetensors",
|
326 |
"model.layers.29.self_attn.dense.weight": "model-00003-of-00003.safetensors",
|
327 |
+
"model.layers.29.self_attn.k_proj.bias": "model-00003-of-00003.safetensors",
|
328 |
+
"model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
|
329 |
"model.layers.29.self_attn.q_proj.bias": "model-00002-of-00003.safetensors",
|
330 |
"model.layers.29.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
|
331 |
"model.layers.29.self_attn.v_proj.bias": "model-00003-of-00003.safetensors",
|