emrgnt-cmplxty commited on
Commit
43d1c96
1 Parent(s): b622277

Upload MistralForCausalLM

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "mistral-without-work-2222",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
@@ -16,11 +16,11 @@
16
  "num_hidden_layers": 32,
17
  "num_key_value_heads": 8,
18
  "rms_norm_eps": 1e-05,
19
- "rope_theta": 1000000,
20
- "sliding_window": 16384,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float32",
23
- "transformers_version": "4.36.2",
24
  "use_cache": false,
25
- "vocab_size": 32003
26
  }
 
1
  {
2
+ "_name_or_path": "mistral-owen-v2/tmp-checkpoint-34364",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
16
  "num_hidden_layers": 32,
17
  "num_key_value_heads": 8,
18
  "rms_norm_eps": 1e-05,
19
+ "rope_theta": 10000.0,
20
+ "sliding_window": 4096,
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float32",
23
+ "transformers_version": "4.37.0.dev0",
24
  "use_cache": false,
25
+ "vocab_size": 32000
26
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.36.2"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.37.0.dev0"
6
  }
model-00001-of-00006.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24ab55ce31e200f257edb538f850a8589df1ccae04d251a02d019162df5cc343
3
- size 4987246088
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87bd26a578887ce55f92736113d51843df65452d2022b512963a4589cc16f3ef
3
+ size 4987196936
model-00002-of-00006.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:298b32068fa9455ceb7c376fdb7a5faf88ee1e822b3d6f945294efc8d60c06cb
3
  size 4899116440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:768bc40c50832672eceeb6f5894805c8f30f2de8e1f9e46d7142bc18abc5f308
3
  size 4899116440
model-00003-of-00006.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:35c44f19a83541e65d3f02d6aeb4cda4b1c790d405d17ec63707868499ead64c
3
  size 4999813120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ada9193f11b2838a9b97edb7e9bff52b0b6696b5ffd6a10ebaff98bcc43c72f4
3
  size 4999813120
model-00004-of-00006.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6cc56daff6197648b9e042656eae34c19d6c939d6ca928c3853c2facd30f30c8
3
  size 4999813128
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e283e4197135f22ab5b5150f67b77b57853efec39787d0f062524e521e7492e
3
  size 4999813128
model-00005-of-00006.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9f840684e7370cc56da752ace87e2f38cfa1ce2ed3dce0a29f941f61474741d
3
  size 4832007496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ae295e3bf19d7d0cf1377e823e8c2d0c6628dda55166c49fc567289b8e01c3f
3
  size 4832007496
model-00006-of-00006.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa128efef337550d36233653b3092aed1147e606ee426c45e7f0c0ce3686de04
3
- size 4249064048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1eaedd828ea8536bb3d8476cd62fa1f322ed9a3c86d16b010c8f60f3a34a3db7
3
+ size 4249014896
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 28967026688
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00006-of-00006.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 28966928384
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00006-of-00006.safetensors",