angelahzyuan commited on
Commit
1969519
1 Parent(s): 48675af

Upload MistralForCausalLM

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
@@ -19,7 +19,7 @@
19
  "rope_theta": 1000000.0,
20
  "sliding_window": null,
21
  "tie_word_embeddings": false,
22
- "torch_dtype": "float32",
23
  "transformers_version": "4.36.2",
24
  "use_cache": false,
25
  "vocab_size": 32000
 
1
  {
2
+ "_name_or_path": "UCLA-AGI/Mistral7B-PairRM-SPPO-Iter2",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
19
  "rope_theta": 1000000.0,
20
  "sliding_window": null,
21
  "tie_word_embeddings": false,
22
+ "torch_dtype": "bfloat16",
23
  "transformers_version": "4.36.2",
24
  "use_cache": false,
25
  "vocab_size": 32000
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bdbec467a629b9edcdb8657b6c241e9a61d7f036f8fdf42daee669f293234274
3
- size 9886313368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f958f8578d17d07997c60b4aba97d512a94513be7fbaca97480eb83e31abdcc0
3
+ size 4943162336
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:58ccc01e971eabcf4dc15178ac7e4e0565bb953caa1149e4a4e00c5120f7b54a
3
- size 9999626240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f0ce64809a5d03130e0e36e87714bbd30cb42e7f33fcd911b655a8ac5fb1e9d
3
+ size 4999819336
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68b555c4781706a6396a88376d770afc4c8b92aa28b069290c063ac6e7b27679
3
- size 9081022384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b80edaf56717bf20568f655533ad28b5b7ab75aa2f18844b8289897542a35174
3
+ size 4540516344
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 28966928384
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00003-of-00003.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 14483464192
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00003-of-00003.safetensors",