voidful commited on
Commit
1a8193e
1 Parent(s): f44c293

Upload PhiForCausalLM

Browse files
config.json CHANGED
@@ -33,5 +33,5 @@
33
  "torch_dtype": "float32",
34
  "transformers_version": "4.39.1",
35
  "use_cache": true,
36
- "vocab_size": 70980
37
  }
 
33
  "torch_dtype": "float32",
34
  "transformers_version": "4.39.1",
35
  "use_cache": true,
36
+ "vocab_size": 80979
37
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bdda8684ea54eed60e454c6b472fe2445ae2f4568023cda7b5a172b8174c8510
3
- size 4922665472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be10fcf4ff5c4ed03ff0cccde11177d35aabbf0fbff0b293fbbd17688a970f98
3
+ size 4998830376
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4611cf8a3b6115b324cdcbc51b6b8ebc595f10453766b7e3c83e8951b44ed9d
3
- size 4982541992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00b5645eb87e4496c1094919b5808ed2600afdabda06b71ce5ad6f2de689f717
3
+ size 4982541984
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc88ae76118b844f76aeb260927891c77594cda92df1653649b59f336746d89b
3
- size 1618751200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b114bea52001937245908f743bd55ce63d716a8e232637c2fb9257866eb35d9
3
+ size 1747405836
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 11523908880
4
  },
5
  "weight_map": {
6
  "lm_head.bias": "model-00003-of-00003.safetensors",
@@ -84,8 +84,8 @@
84
  "model.layers.13.mlp.fc1.weight": "model-00002-of-00003.safetensors",
85
  "model.layers.13.mlp.fc2.bias": "model-00002-of-00003.safetensors",
86
  "model.layers.13.mlp.fc2.weight": "model-00002-of-00003.safetensors",
87
- "model.layers.13.self_attn.dense.bias": "model-00001-of-00003.safetensors",
88
- "model.layers.13.self_attn.dense.weight": "model-00001-of-00003.safetensors",
89
  "model.layers.13.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
90
  "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
91
  "model.layers.13.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
@@ -324,8 +324,8 @@
324
  "model.layers.29.mlp.fc2.weight": "model-00003-of-00003.safetensors",
325
  "model.layers.29.self_attn.dense.bias": "model-00003-of-00003.safetensors",
326
  "model.layers.29.self_attn.dense.weight": "model-00003-of-00003.safetensors",
327
- "model.layers.29.self_attn.k_proj.bias": "model-00002-of-00003.safetensors",
328
- "model.layers.29.self_attn.k_proj.weight": "model-00002-of-00003.safetensors",
329
  "model.layers.29.self_attn.q_proj.bias": "model-00002-of-00003.safetensors",
330
  "model.layers.29.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
331
  "model.layers.29.self_attn.v_proj.bias": "model-00003-of-00003.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 11728728396
4
  },
5
  "weight_map": {
6
  "lm_head.bias": "model-00003-of-00003.safetensors",
 
84
  "model.layers.13.mlp.fc1.weight": "model-00002-of-00003.safetensors",
85
  "model.layers.13.mlp.fc2.bias": "model-00002-of-00003.safetensors",
86
  "model.layers.13.mlp.fc2.weight": "model-00002-of-00003.safetensors",
87
+ "model.layers.13.self_attn.dense.bias": "model-00002-of-00003.safetensors",
88
+ "model.layers.13.self_attn.dense.weight": "model-00002-of-00003.safetensors",
89
  "model.layers.13.self_attn.k_proj.bias": "model-00001-of-00003.safetensors",
90
  "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00003.safetensors",
91
  "model.layers.13.self_attn.q_proj.bias": "model-00001-of-00003.safetensors",
 
324
  "model.layers.29.mlp.fc2.weight": "model-00003-of-00003.safetensors",
325
  "model.layers.29.self_attn.dense.bias": "model-00003-of-00003.safetensors",
326
  "model.layers.29.self_attn.dense.weight": "model-00003-of-00003.safetensors",
327
+ "model.layers.29.self_attn.k_proj.bias": "model-00003-of-00003.safetensors",
328
+ "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00003.safetensors",
329
  "model.layers.29.self_attn.q_proj.bias": "model-00002-of-00003.safetensors",
330
  "model.layers.29.self_attn.q_proj.weight": "model-00002-of-00003.safetensors",
331
  "model.layers.29.self_attn.v_proj.bias": "model-00003-of-00003.safetensors",