thomadev0 commited on
Commit
32c0c78
·
verified ·
1 Parent(s): 86f1d73

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +10 -0
  2. checkpoint/config.json +30 -0
  3. checkpoint/generation_config.json +6 -0
  4. checkpoint/pytorch_model.bin/key_to_filename.json +3 -0
  5. checkpoint/pytorch_model.bin/p0.model.embed_tokens.weight +3 -0
  6. checkpoint/pytorch_model.bin/p1.model.layers.0.self_attn.q_proj.weight +3 -0
  7. checkpoint/pytorch_model.bin/p10.model.layers.0.mlp.up_proj.weight +3 -0
  8. checkpoint/pytorch_model.bin/p100.model.layers.7.mlp.gate_proj.weight +3 -0
  9. checkpoint/pytorch_model.bin/p1000.model.layers.76.input_layernorm.weight +3 -0
  10. checkpoint/pytorch_model.bin/p1001.model.layers.76.post_attention_layernorm.weight +3 -0
  11. checkpoint/pytorch_model.bin/p1002.model.layers.77.self_attn.q_proj.weight +3 -0
  12. checkpoint/pytorch_model.bin/p1003.model.layers.77.self_attn.q_proj.bias +3 -0
  13. checkpoint/pytorch_model.bin/p1004.model.layers.77.self_attn.k_proj.weight +3 -0
  14. checkpoint/pytorch_model.bin/p1005.model.layers.77.self_attn.k_proj.bias +3 -0
  15. checkpoint/pytorch_model.bin/p1006.model.layers.77.self_attn.v_proj.weight +3 -0
  16. checkpoint/pytorch_model.bin/p1007.model.layers.77.self_attn.v_proj.bias +3 -0
  17. checkpoint/pytorch_model.bin/p1008.model.layers.77.self_attn.o_proj.weight +3 -0
  18. checkpoint/pytorch_model.bin/p1009.model.layers.77.self_attn.o_proj.bias +3 -0
  19. checkpoint/pytorch_model.bin/p101.model.layers.7.mlp.up_proj.weight +3 -0
  20. checkpoint/pytorch_model.bin/p1010.model.layers.77.mlp.gate_proj.weight +3 -0
  21. checkpoint/pytorch_model.bin/p1011.model.layers.77.mlp.up_proj.weight +3 -0
  22. checkpoint/pytorch_model.bin/p1012.model.layers.77.mlp.down_proj.weight +3 -0
  23. checkpoint/pytorch_model.bin/p1013.model.layers.77.input_layernorm.weight +3 -0
  24. checkpoint/pytorch_model.bin/p1014.model.layers.77.post_attention_layernorm.weight +3 -0
  25. checkpoint/pytorch_model.bin/p1015.model.layers.78.self_attn.q_proj.weight +3 -0
  26. checkpoint/pytorch_model.bin/p1016.model.layers.78.self_attn.q_proj.bias +3 -0
  27. checkpoint/pytorch_model.bin/p1017.model.layers.78.self_attn.k_proj.weight +3 -0
  28. checkpoint/pytorch_model.bin/p1018.model.layers.78.self_attn.k_proj.bias +3 -0
  29. checkpoint/pytorch_model.bin/p1019.model.layers.78.self_attn.v_proj.weight +3 -0
  30. checkpoint/pytorch_model.bin/p102.model.layers.7.mlp.down_proj.weight +3 -0
  31. checkpoint/pytorch_model.bin/p1020.model.layers.78.self_attn.v_proj.bias +3 -0
  32. checkpoint/pytorch_model.bin/p1021.model.layers.78.self_attn.o_proj.weight +3 -0
  33. checkpoint/pytorch_model.bin/p1022.model.layers.78.self_attn.o_proj.bias +3 -0
  34. checkpoint/pytorch_model.bin/p1023.model.layers.78.mlp.gate_proj.weight +3 -0
  35. checkpoint/pytorch_model.bin/p1024.model.layers.78.mlp.up_proj.weight +3 -0
  36. checkpoint/pytorch_model.bin/p1025.model.layers.78.mlp.down_proj.weight +3 -0
  37. checkpoint/pytorch_model.bin/p1026.model.layers.78.input_layernorm.weight +3 -0
  38. checkpoint/pytorch_model.bin/p1027.model.layers.78.post_attention_layernorm.weight +3 -0
  39. checkpoint/pytorch_model.bin/p1028.model.layers.79.self_attn.q_proj.weight +3 -0
  40. checkpoint/pytorch_model.bin/p1029.model.layers.79.self_attn.q_proj.bias +3 -0
  41. checkpoint/pytorch_model.bin/p103.model.layers.7.input_layernorm.weight +3 -0
  42. checkpoint/pytorch_model.bin/p1030.model.layers.79.self_attn.k_proj.weight +3 -0
  43. checkpoint/pytorch_model.bin/p1031.model.layers.79.self_attn.k_proj.bias +3 -0
  44. checkpoint/pytorch_model.bin/p1032.model.layers.79.self_attn.v_proj.weight +3 -0
  45. checkpoint/pytorch_model.bin/p1033.model.layers.79.self_attn.v_proj.bias +3 -0
  46. checkpoint/pytorch_model.bin/p1034.model.layers.79.self_attn.o_proj.weight +3 -0
  47. checkpoint/pytorch_model.bin/p1035.model.layers.79.self_attn.o_proj.bias +3 -0
  48. checkpoint/pytorch_model.bin/p1036.model.layers.79.mlp.gate_proj.weight +3 -0
  49. checkpoint/pytorch_model.bin/p1037.model.layers.79.mlp.up_proj.weight +3 -0
  50. checkpoint/pytorch_model.bin/p1038.model.layers.79.mlp.down_proj.weight +3 -0
.gitattributes CHANGED
@@ -33,3 +33,13 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ compiled/39671fa6153a417c0059.neff filter=lfs diff=lfs merge=lfs -text
37
+ compiled/4004277e31e89dbc2cd2.neff filter=lfs diff=lfs merge=lfs -text
38
+ compiled/8273c4799b9ddf453584.neff filter=lfs diff=lfs merge=lfs -text
39
+ compiled/86668046903bd034a5a6.neff filter=lfs diff=lfs merge=lfs -text
40
+ compiled/891ea69f293f2939545b.neff filter=lfs diff=lfs merge=lfs -text
41
+ compiled/8c4b90826998854c36a0.neff filter=lfs diff=lfs merge=lfs -text
42
+ compiled/933237200bbf8dc1417b.neff filter=lfs diff=lfs merge=lfs -text
43
+ compiled/93e48a55161931f3aee7.neff filter=lfs diff=lfs merge=lfs -text
44
+ compiled/cf6f85a4d52f11817865.neff filter=lfs diff=lfs merge=lfs -text
45
+ compiled/d3b6c95e6bb346ee4b43.neff filter=lfs diff=lfs merge=lfs -text
checkpoint/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Smaug-72B-v0.1/",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": true,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 151643,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 8192,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 24576,
14
+ "max_position_embeddings": 32768,
15
+ "model_type": "llama",
16
+ "num_attention_heads": 64,
17
+ "num_hidden_layers": 80,
18
+ "num_key_value_heads": 64,
19
+ "pad_token_id": 151643,
20
+ "pretraining_tp": 1,
21
+ "rms_norm_eps": 1e-06,
22
+ "rope_scaling": null,
23
+ "rope_theta": 1000000,
24
+ "seq_length": 32768,
25
+ "tie_word_embeddings": false,
26
+ "torch_dtype": "float32",
27
+ "transformers_version": "4.36.2",
28
+ "use_cache": true,
29
+ "vocab_size": 152064
30
+ }
checkpoint/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "eos_token_id": 151643,
4
+ "pad_token_id": 151643,
5
+ "transformers_version": "4.36.2"
6
+ }
checkpoint/pytorch_model.bin/key_to_filename.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96b043d79e15b14221ff9be0360598ae29d60dc3bdff1d41ec9a92ca94ae65c7
3
+ size 94632
checkpoint/pytorch_model.bin/p0.model.embed_tokens.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf2ba4dff83df47c51598892b511a6d85d77ae3f711c042c648a28ec0dd6a3c8
3
+ size 4982834045
checkpoint/pytorch_model.bin/p1.model.layers.0.self_attn.q_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c38ac059caca2da46fd6f9832f39c1e79a0c7e383f0127b72899e6a9fa8cfc38
3
+ size 268436348
checkpoint/pytorch_model.bin/p10.model.layers.0.mlp.up_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dd88f35543815914d0b39b61962ffebe2ae768a182374126e71686ad5f626e8
3
+ size 805307248
checkpoint/pytorch_model.bin/p100.model.layers.7.mlp.gate_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92c50b02de0869b7c80cb16953b4c314a88fc1b1314f1193bf9a443d30c93d67
3
+ size 805307257
checkpoint/pytorch_model.bin/p1000.model.layers.76.input_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19bb3c6146e250c2407a4722815201388976a1df1c46af1c661fca8b2f0bc81a
3
+ size 33669
checkpoint/pytorch_model.bin/p1001.model.layers.76.post_attention_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d4d800221ad6143f6ab9251efe3a9730451a02b093332da0c9bbd861649040a
3
+ size 33696
checkpoint/pytorch_model.bin/p1002.model.layers.77.self_attn.q_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13035ca89111cda145c5399987ff6b9b4b492e5006f1cc53b5b7bc95f24a0c4b
3
+ size 268436360
checkpoint/pytorch_model.bin/p1003.model.layers.77.self_attn.q_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02ffe7f1e60cd04f776a237de2cce85eb10789c31a2de02d8574d1ccdcbe4f97
3
+ size 33672
checkpoint/pytorch_model.bin/p1004.model.layers.77.self_attn.k_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edbcf6ebc07dd0acfdf1819841ded195e5e1fb8cc3d9eafe9d87f32b41f39582
3
+ size 268436360
checkpoint/pytorch_model.bin/p1005.model.layers.77.self_attn.k_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34900538f22980f303d54a6e2b5b1e49ea36ed985be637af73d31c8e0b45d852
3
+ size 33672
checkpoint/pytorch_model.bin/p1006.model.layers.77.self_attn.v_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca578550b4d4d961f2d7bba0e076a50c4dbe5cbb8b16d99594e2cc41dc654c8a
3
+ size 268436360
checkpoint/pytorch_model.bin/p1007.model.layers.77.self_attn.v_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f1b0f95d39db390c677155ba2ec6f9a5a1537043759e14453658462e0ac2fcf
3
+ size 33672
checkpoint/pytorch_model.bin/p1008.model.layers.77.self_attn.o_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e38d3cf2e779558f117236f05b86e52e58c9ba5aa103fd2181d086dc1184659
3
+ size 268436360
checkpoint/pytorch_model.bin/p1009.model.layers.77.self_attn.o_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ee348eeb2703a35bf6d9e4b170fb0855ac8ad719cb48119baaef3e11b3f3dd8
3
+ size 33672
checkpoint/pytorch_model.bin/p101.model.layers.7.mlp.up_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c052eceb6937fae9c8ff34c51f5534ea36944a2a59965cfd812306d1f535e13a
3
+ size 805307251
checkpoint/pytorch_model.bin/p1010.model.layers.77.mlp.gate_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a08e113ca55fb0937034ad67cd08409a25d011d3632a84b964d54de656e0aeb
3
+ size 805307263
checkpoint/pytorch_model.bin/p1011.model.layers.77.mlp.up_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84e5a2a121d13c58e96f97c6d963915b99b5870d95ba8375a71cfea36bb930bc
3
+ size 805307257
checkpoint/pytorch_model.bin/p1012.model.layers.77.mlp.down_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7042573bac0b64307d1e52da23c52c1d94a135c8756b030220776237700a50db
3
+ size 805307263
checkpoint/pytorch_model.bin/p1013.model.layers.77.input_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6556366826ef92f791fd854c422a837e23ab12002b47b7a52767b6498f40448
3
+ size 33669
checkpoint/pytorch_model.bin/p1014.model.layers.77.post_attention_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e92b8afcb8454b148b8698f0ed41166b05b84440e9600c234005e5e2f7f1f82
3
+ size 33696
checkpoint/pytorch_model.bin/p1015.model.layers.78.self_attn.q_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:def16bdc409447f5b8686dc44c9564cc693b480a1b355b234c9a25ced4f44865
3
+ size 268436360
checkpoint/pytorch_model.bin/p1016.model.layers.78.self_attn.q_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:288532a9332460dba0cf6d702ec2c6df7f6cefa106b55a6aed4be88cd0265cd5
3
+ size 33672
checkpoint/pytorch_model.bin/p1017.model.layers.78.self_attn.k_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd07a7f67337533e714c131af0b8115960637398d83d790ba14d31ac4c10ce39
3
+ size 268436360
checkpoint/pytorch_model.bin/p1018.model.layers.78.self_attn.k_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cb8b18d4e40ec94b3117826e682b09ae98e6f5ac1f4c29037ee184239f91fad
3
+ size 33672
checkpoint/pytorch_model.bin/p1019.model.layers.78.self_attn.v_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6a0370e5e8ae0c85c895579f5efe50865e5835bd381049ca4b02cbd34e57380
3
+ size 268436360
checkpoint/pytorch_model.bin/p102.model.layers.7.mlp.down_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1b784afda55f8ecfb32d7c2090f0f50d9fd2a480f5c8d061342e0540314da9e
3
+ size 805307257
checkpoint/pytorch_model.bin/p1020.model.layers.78.self_attn.v_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff2f99a64c5dddc47157f2bb39658d560701649f0fe6c1b5c6f7abd023675bd9
3
+ size 33672
checkpoint/pytorch_model.bin/p1021.model.layers.78.self_attn.o_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13604c6e361891aefc71f1378a38a29a554ad923be116abfa346e3e2bbde9242
3
+ size 268436360
checkpoint/pytorch_model.bin/p1022.model.layers.78.self_attn.o_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:657aadf2e69572adddc88a4a3240708ad7d555e90ab71578ceef0d5fa193b2ea
3
+ size 33672
checkpoint/pytorch_model.bin/p1023.model.layers.78.mlp.gate_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88e146b9dd48639c765f574d168ca3531b5b2302e55f52fb54cf9ecd7faeba97
3
+ size 805307263
checkpoint/pytorch_model.bin/p1024.model.layers.78.mlp.up_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27c65d77479b717511b92c0510c78606edd2c095089e6ada45029ebedb0470f4
3
+ size 805307257
checkpoint/pytorch_model.bin/p1025.model.layers.78.mlp.down_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f652d93bd7527139330d16859b522494ab62f2bca0b48b9f5353fed77e2873a
3
+ size 805307263
checkpoint/pytorch_model.bin/p1026.model.layers.78.input_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c059e62110f6ffd08cc1877f1bd873cefaab8b3fc7fa614870c87f8708c92ac
3
+ size 33669
checkpoint/pytorch_model.bin/p1027.model.layers.78.post_attention_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9390d4913d8d9a7f6eb4a52ed6066bd99d4c8a0cde28db624947c39b5092cea0
3
+ size 33696
checkpoint/pytorch_model.bin/p1028.model.layers.79.self_attn.q_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc4203e66420cf3af9ec409d3dbc4275bbe51efe5ae5b13e4113acefa612753a
3
+ size 268436360
checkpoint/pytorch_model.bin/p1029.model.layers.79.self_attn.q_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f65142b19918cafb31ff7999188e0cae5ce6f3bd3bc10798a96354e473499a39
3
+ size 33672
checkpoint/pytorch_model.bin/p103.model.layers.7.input_layernorm.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a96748c23d540fce2a2b825e4f1fd377fa3b91fa6ef77f30bb30b51a9bba53b9
3
+ size 33663
checkpoint/pytorch_model.bin/p1030.model.layers.79.self_attn.k_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1587ba215131195fe4829a95b91dd3405e1f707b58d1bf44300641c74fc011a9
3
+ size 268436360
checkpoint/pytorch_model.bin/p1031.model.layers.79.self_attn.k_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01cd37eb94c86a33b304edec230116649740f6d1929067245455c542c8b49f81
3
+ size 33672
checkpoint/pytorch_model.bin/p1032.model.layers.79.self_attn.v_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b7044897d53e08ed91cd4d64145ec7bfe45b2c206bc8323e453d5fd11efbf29
3
+ size 268436360
checkpoint/pytorch_model.bin/p1033.model.layers.79.self_attn.v_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8cc4dc8b3d5697a55e5e178c13feffc6151dff98c662cd41ed9271a95d1dc1a
3
+ size 33672
checkpoint/pytorch_model.bin/p1034.model.layers.79.self_attn.o_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6183c14a67e5c09c0d57611b6d04f1e61e0cc92025b372dfe88aafb931abfe99
3
+ size 268436360
checkpoint/pytorch_model.bin/p1035.model.layers.79.self_attn.o_proj.bias ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4943793ccbafe4b0dfe082b6b7b9aff21973d7934a2fb5810230b9eaa1466665
3
+ size 33672
checkpoint/pytorch_model.bin/p1036.model.layers.79.mlp.gate_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e01b3446baf8c407683141d85e48bcd12b7de1dc862af39e9d34460d560bc683
3
+ size 805307263
checkpoint/pytorch_model.bin/p1037.model.layers.79.mlp.up_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7877e35babcb3a57f8a0fe8270cd2dc8bc019b68667d95d8505a7ae2d30146e6
3
+ size 805307257
checkpoint/pytorch_model.bin/p1038.model.layers.79.mlp.down_proj.weight ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cc4dd463df45331a9159d4d22229d520e41bc96400e164a87da3adad62fbf09
3
+ size 805307263