diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1f1cefb00851a8245d2be7dabd4c16e82559017 --- /dev/null +++ b/config.json @@ -0,0 +1,29 @@ +{ + "_name_or_path": "E:\\lab\\workspace\\PyRosa-70B-A\\", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 1, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 4096, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "pad_token_id": 0, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 10000.0, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.38.2", + "use_cache": true, + "vocab_size": 32000 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..4e758132677422f04ff1bcc081ebb88034b53528 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,7 @@ +{ + "_from_model_config": true, + "bos_token_id": 1, + "eos_token_id": 2, + "pad_token_id": 0, + "transformers_version": "4.38.2" +} diff --git a/model-00001-of-00081.safetensors b/model-00001-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c66d254b228bf3fa06365695e7baaac637ecc9bc --- /dev/null +++ b/model-00001-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d06e59d50a8c72dbed1e46c9d5cc98c822374889bf630d5fea7422aa9ea341a +size 1765802808 diff --git a/model-00002-of-00081.safetensors b/model-00002-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9ada06aa8a973a3251486f93938d6b6c2035e76a --- /dev/null +++ b/model-00002-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:481a2323ccf7bf51573cccadb827f5bc41ea25eb715b074a3c7913f8c88cf8d8 +size 1711309848 diff --git a/model-00003-of-00081.safetensors b/model-00003-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..990ad34a5c6db369bf19aae014554a7cc8d01407 --- /dev/null +++ b/model-00003-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49daabd7f120758418d684c35f781872027b15db542d5c1ea7b9c64bdf66193d +size 1711309848 diff --git a/model-00004-of-00081.safetensors b/model-00004-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2b93830b8a7795fe254c73c7be65b35b3fc1f3c5 --- /dev/null +++ b/model-00004-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1978debb3e92c9e818c863ac9d3d278ce1ae28625eaa21116ca4efff6c433b21 +size 1711309848 diff --git a/model-00005-of-00081.safetensors b/model-00005-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4696a8ec9edeb75540e82d46faddc8c93e8a2513 --- /dev/null +++ b/model-00005-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5bad4e9e60bf288c4f14775ddc0580a301ef848f5658fb3379c8ac670fe4a009 +size 1711309848 diff --git a/model-00006-of-00081.safetensors b/model-00006-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..79d66935ee41418552a9bfbb1c3ee575815e9e88 --- /dev/null +++ b/model-00006-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c29ede6773ba99c82fe51ce087a9e02959889224159283f059f04dd84ab2162c +size 1711309848 diff --git a/model-00007-of-00081.safetensors b/model-00007-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2b3af7879915c85c0d6b4269c3f98988dedeb7bf --- /dev/null +++ b/model-00007-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d39bc653fd5035111e7c470c6c7a22f6f6ab76b17058303fd6526b72ed5c564 +size 1711309848 diff --git a/model-00008-of-00081.safetensors b/model-00008-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8022e7eb80ed421c379b7100dbe1bd10d83ec94a --- /dev/null +++ b/model-00008-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03ad1cacea51fd9ad7ffcf4608513b69577baa6abb7be87beeddc01e46b45409 +size 1711309848 diff --git a/model-00009-of-00081.safetensors b/model-00009-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3d176d6cfdb949c10eafd3aa70b6402eb9cc341b --- /dev/null +++ b/model-00009-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad5de7a77f86ecaa9a88b8885ec728d7fec87374bc2b67949a3fdcba8aeaefe8 +size 1711309848 diff --git a/model-00010-of-00081.safetensors b/model-00010-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5e8e63baaedb44dd8944ec62fcb02b0b231cfc4c --- /dev/null +++ b/model-00010-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec2af9d7ac7b4a8897162818ffb5e84effda64a7b32035060a1462fc97ea2035 +size 1711309848 diff --git a/model-00011-of-00081.safetensors b/model-00011-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a5abccddadc11be2064accadec1f209d1dad5258 --- /dev/null +++ b/model-00011-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f6bd55e6258eb1823e907b6c5f601cb16f48ed3ea8d9b5c074dd7b1f4b45fdf +size 1711309864 diff --git a/model-00012-of-00081.safetensors b/model-00012-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9740cc60758839fe458808b39e2f0e5567d4673b --- /dev/null +++ b/model-00012-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:499ebd85be63d010278b6ff4750c90ef11387f7c6661ed3b42eddf8f922e2972 +size 1711309864 diff --git a/model-00013-of-00081.safetensors b/model-00013-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..543bda7582fd6fd394421b30e7d2b856373c5d93 --- /dev/null +++ b/model-00013-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c79aa87b910586ccdf6f05113bb873fb47bf450ab8db98b0fa85cb94fca6ef09 +size 1711309864 diff --git a/model-00014-of-00081.safetensors b/model-00014-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..38fdd9d84e732be61f9c4e9c3c7ac3d7aa8cc8a7 --- /dev/null +++ b/model-00014-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c130dd73b3e1183594485ebc71ab88c2189205521d0d986577923bd46b5d0d6e +size 1711309864 diff --git a/model-00015-of-00081.safetensors b/model-00015-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c881bb6795e8f0fbd3e8c10cb63d2c74f3fc3525 --- /dev/null +++ b/model-00015-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e19b0c83027c757f13095caf1ccdcab7af69154035e0b498821fe9a2b33138c +size 1711309864 diff --git a/model-00016-of-00081.safetensors b/model-00016-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..81ef19a10b593ad57285493839e23ec914f980f3 --- /dev/null +++ b/model-00016-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cc28c1cca2d962a22d66d23681b7af239945ea6c8c6d953567a8c0ac4de493c +size 1711309864 diff --git a/model-00017-of-00081.safetensors b/model-00017-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7be15934e751c5b3e1cda17c938cd66b72a92e2b --- /dev/null +++ b/model-00017-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9635bc56d4e008e06d8296643fd4bdf3dee4027047217cbfb4f5ffd1f511fcf3 +size 1711309864 diff --git a/model-00018-of-00081.safetensors b/model-00018-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a1d3146c6b30c5eeed21b6d53f3a85c1a8f4375b --- /dev/null +++ b/model-00018-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39a0c96a70a8b06e0cc0100c047e84daf00df25f0b06589b56d781cb77109ac5 +size 1711309864 diff --git a/model-00019-of-00081.safetensors b/model-00019-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..54400711ab515b1d5b37a643c9ec7c8073b35443 --- /dev/null +++ b/model-00019-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58282926fdea574151a57e319c2579f5b2b29755e5dfc797af6f49b53a14a72c +size 1711309864 diff --git a/model-00020-of-00081.safetensors b/model-00020-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5a593621b79b1060e7926f3716933ee70882da3c --- /dev/null +++ b/model-00020-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a5f3a715109aaa706173e5021c21268b35306cb6f82ae6df2bee268f83dc49c +size 1711309864 diff --git a/model-00021-of-00081.safetensors b/model-00021-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d81625c0fd2398baddc36a9e845e15e7412e9174 --- /dev/null +++ b/model-00021-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b48fafe4d4416b7451c6a5e3fabde4eb0589228746bb5566772e40703aedcae +size 1711309864 diff --git a/model-00022-of-00081.safetensors b/model-00022-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..02150832c3d73dbb09fc58b0e7817e9015eb27c1 --- /dev/null +++ b/model-00022-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50113c57c0be28d174bf91008c01ab8f3987659854505c17c404166c40d36785 +size 1711309864 diff --git a/model-00023-of-00081.safetensors b/model-00023-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..85eb060c43cb062c91a205b795321605d6ceb1c1 --- /dev/null +++ b/model-00023-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1024d8bb0d8f553f27b3e728ace883c44c036872b5832a0666b916ff033b0e6 +size 1711309864 diff --git a/model-00024-of-00081.safetensors b/model-00024-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9baf411e663ed13b4ad61b268e84011462cb03ba --- /dev/null +++ b/model-00024-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5bf1aa6268f9f2c8e875fc57b9a5e0f7bdbffc6722043515ea2c142f56e8354c +size 1711309864 diff --git a/model-00025-of-00081.safetensors b/model-00025-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fc647a5c3eb3c6c9834ccae13b3c4604148f75c8 --- /dev/null +++ b/model-00025-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:029f7f7360b6c60b50652ffe1d72d3cb92543d2f4a8ae795c65fc2b7394cf427 +size 1711309864 diff --git a/model-00026-of-00081.safetensors b/model-00026-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b199b061bad5981fa951edc933f38ebf6f379e5d --- /dev/null +++ b/model-00026-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0404e63cd1fd1094236c5310f6fbfa8b12311ac57e1d1217f56f35d9c54c6b60 +size 1711309864 diff --git a/model-00027-of-00081.safetensors b/model-00027-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..71107a6382451d1c4002c0162547a88ddb3981b2 --- /dev/null +++ b/model-00027-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d90df847b5ab3e6bcb93a547181e40d37dfce3722b8eebc34097135ffaecc207 +size 1711309864 diff --git a/model-00028-of-00081.safetensors b/model-00028-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6fcb1d4c1222fe494ed4109888763070b9913c29 --- /dev/null +++ b/model-00028-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c99e56888c377bdda9834390728eae41ffc85f9db472ad5b55a9362c7133505 +size 1711309864 diff --git a/model-00029-of-00081.safetensors b/model-00029-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3a16ceb44fc7870c873ca1e85d7970b2baf8b134 --- /dev/null +++ b/model-00029-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2a276b443470b3107831ce3910f43f3f1a47c9533ce1cd4e4b9e99defcc4f92 +size 1711309864 diff --git a/model-00030-of-00081.safetensors b/model-00030-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..222ecc4ab899a618b4fe062d6e5c18fe1fc4db8f --- /dev/null +++ b/model-00030-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf473c6caf556cc22aa504119e89ac9b0bd237b5705d105ab66ada446320235b +size 1711309864 diff --git a/model-00031-of-00081.safetensors b/model-00031-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cc238f98524bbd72e9437aeca227b7677ecb8c25 --- /dev/null +++ b/model-00031-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17009de797fc500339aed4a377d9d6c829fea19d28ae775744b1bbf0a7455440 +size 1711309864 diff --git a/model-00032-of-00081.safetensors b/model-00032-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..83dca5717fa02ae604b9492b0be592521881a174 --- /dev/null +++ b/model-00032-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c2a487803094682df05479660f4317ee46a9e3c77af3d384f6be1d555d91e5b +size 1711309864 diff --git a/model-00033-of-00081.safetensors b/model-00033-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d6d34d7405a2281b5559962f66e25a8dfa36226f --- /dev/null +++ b/model-00033-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b2dce255ec6d4288c6ebe22fd07492ebe1cf83657b3398f42d4ccebd5701c1b +size 1711309864 diff --git a/model-00034-of-00081.safetensors b/model-00034-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b9398b9bec8272fc0b6e61bc293c8ba2f690d344 --- /dev/null +++ b/model-00034-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52e9d34643515c13fdfed4cc3b1f6ec2f0ff8bc11ed9c02d3c86c29cb6923bad +size 1711309864 diff --git a/model-00035-of-00081.safetensors b/model-00035-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a1a4b1bbfb4fb88b734da83164d8f2ab91c7d96b --- /dev/null +++ b/model-00035-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6273310caed72bf2b5e88c0ebefb8e12b255b02be20c6fe1c014cb64c5d7f650 +size 1711309864 diff --git a/model-00036-of-00081.safetensors b/model-00036-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a361163190e840cdd8e728561b9bfbca224aa708 --- /dev/null +++ b/model-00036-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13c7ffa45063dcab7d5ed621bc581955c0922ff745faf22bb9b1be236e5d26a8 +size 1711309864 diff --git a/model-00037-of-00081.safetensors b/model-00037-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ee2289192b5592e2c5f079b9eba502cc1c67495b --- /dev/null +++ b/model-00037-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:edb49547aca72c3d801caf3f32948c5564f3b17026fe782a0c9c53a8de470188 +size 1711309864 diff --git a/model-00038-of-00081.safetensors b/model-00038-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3d7cb85044a4000671c2cace5fca30e0e1667f96 --- /dev/null +++ b/model-00038-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b5fcce8e7f02e0d82eba25e9726d6aac0aa9756a39e81dd1a3e1c06edbbb02a +size 1711309864 diff --git a/model-00039-of-00081.safetensors b/model-00039-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c4e58ef98bd837258f2d2ad88a2e7ac51530fa1b --- /dev/null +++ b/model-00039-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c22531fa44bbf6a64f2b6a6f5883f4cc7e8ff12d33ec63b41ab59f04ac5800e +size 1711309864 diff --git a/model-00040-of-00081.safetensors b/model-00040-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..06c831eb39694675d67531761d16d238dd077e91 --- /dev/null +++ b/model-00040-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f925850ff815864f36b1535bcf54e0a537a48b0adc312c08c2f701458f5df208 +size 1711309864 diff --git a/model-00041-of-00081.safetensors b/model-00041-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..70e964a54fb56a10a0b4c5acd77e61c6a9d1b3b5 --- /dev/null +++ b/model-00041-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7a1aea4b914f5e20ea25e3b7d63bef7c263d301785ab6e4e87fc7063b32c372 +size 1711309864 diff --git a/model-00042-of-00081.safetensors b/model-00042-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7dfaa905a461d69e6518b9c0a7c1685777497c42 --- /dev/null +++ b/model-00042-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f939d522e9e3d3345a5cfa05d68bdbf324b94e194b3b7338321cf43a1d441c4 +size 1711309864 diff --git a/model-00043-of-00081.safetensors b/model-00043-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7bf8ceed553fcbd3e508fbd9e4b5ecdc5f69dd69 --- /dev/null +++ b/model-00043-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c13ee420e300da2033c6ed0c662476eafe4f75862e243f3b4e4e6de146ab46e +size 1711309864 diff --git a/model-00044-of-00081.safetensors b/model-00044-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0198abb179779d4c869ab814fc70fade93609fd2 --- /dev/null +++ b/model-00044-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05ecec1054b2bd72722a23d5902e6ed11c4e0cb92f8b012a1adcc916561d86e0 +size 1711309864 diff --git a/model-00045-of-00081.safetensors b/model-00045-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..25c1998e874e605aaa12e06146e7656f5bad4c3a --- /dev/null +++ b/model-00045-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c25d7872419984fdc3153dce7f119f6cb4888fccfa913c86fbf915e3611f1eb6 +size 1711309864 diff --git a/model-00046-of-00081.safetensors b/model-00046-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7e6c8a0e68b7034bee50415cfee5ed323079a2c5 --- /dev/null +++ b/model-00046-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c5c35640203b7793d7447ed79e4bd0cf49d5b3109a09268f103d980db56ddf2 +size 1711309864 diff --git a/model-00047-of-00081.safetensors b/model-00047-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..537c151927c0e0d4029f7de9a079cb0ecfb43f48 --- /dev/null +++ b/model-00047-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fba0f87915049400d55724d671a7c5c94a39ffb9fb5b74d57acacd20b2a6318b +size 1711309864 diff --git a/model-00048-of-00081.safetensors b/model-00048-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cf0c34b54f8323b5eebb5ee56ad4ec0a8089e5b9 --- /dev/null +++ b/model-00048-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e73e7b178019d143633b835c41a8285f443e0c886c37d65256f27cf2ac1f507 +size 1711309864 diff --git a/model-00049-of-00081.safetensors b/model-00049-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3105999910beab85a98c33fd2e9756af190fdc17 --- /dev/null +++ b/model-00049-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb93e601f49d288ed3d0e0f2e38189c8e9031e4b956bb825dfe79d37559391bc +size 1711309864 diff --git a/model-00050-of-00081.safetensors b/model-00050-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d2211973e52453216f236b5edaa9baab99406fb2 --- /dev/null +++ b/model-00050-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4c47f834e42e293c53fa35da286f59299ee82998708066e9a0d4b6c7319236a +size 1711309864 diff --git a/model-00051-of-00081.safetensors b/model-00051-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e00f19cdc8fb7facbad7999182dc085d680c1c71 --- /dev/null +++ b/model-00051-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82ef5fedd4714e8cceaec9d7c873fee7147ed75c467bb024e948081283170323 +size 1711309864 diff --git a/model-00052-of-00081.safetensors b/model-00052-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..261c656ce19292bd0c660449361aa72b11cbe2ce --- /dev/null +++ b/model-00052-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6b1242a432d64f68b1a2dc3942874d0394cf5d248223b22dcd6807dc7351094 +size 1711309864 diff --git a/model-00053-of-00081.safetensors b/model-00053-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..021911a17955ff5a6ca0e2a71952253868973448 --- /dev/null +++ b/model-00053-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c53e0359e932aa2f4ef9897682cb76a6adc56c09f7e74aa2f79b0ccc478596f +size 1711309864 diff --git a/model-00054-of-00081.safetensors b/model-00054-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e36677fc837698e8b8c0cebc33f85e148d37067a --- /dev/null +++ b/model-00054-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba8040ce3c82aa779c47320bb3a2e94bc6d233c148782c931168fa3717df66e9 +size 1711309864 diff --git a/model-00055-of-00081.safetensors b/model-00055-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b63e4626ffb6a4fba995fdfc55f81a4079fe8a89 --- /dev/null +++ b/model-00055-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5de83cfd65fd673c494b93c9c6bc922383567a18829e528169af29ebaf608b5 +size 1711309864 diff --git a/model-00056-of-00081.safetensors b/model-00056-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..69ea1f46f85db82771491009311c5afc4780e057 --- /dev/null +++ b/model-00056-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3cb11d874fc406523e4827816ed3133555c285862fa70f106669c14e138114b2 +size 1711309864 diff --git a/model-00057-of-00081.safetensors b/model-00057-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c0ddd0fecca099f9f27394e60be6f23e138788f4 --- /dev/null +++ b/model-00057-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9f832e485360ff75871b509ccced4b827b69c562b1f42d80d8a0a440ca81ce82 +size 1711309864 diff --git a/model-00058-of-00081.safetensors b/model-00058-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ab046fb633f6eaf32344a3fbb965b54f0f93ac94 --- /dev/null +++ b/model-00058-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19ddcfeba62d04e187451d02b0e54441a0e72a805f7d949c9818b61c3fe1e079 +size 1711309864 diff --git a/model-00059-of-00081.safetensors b/model-00059-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..82dbaaaeed232945e80cae1edc23937cf91c768e --- /dev/null +++ b/model-00059-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e8c941055df549e55fe59f6e81b6d2b71bc389123d420fd8cc920218a4ad2e9 +size 1711309864 diff --git a/model-00060-of-00081.safetensors b/model-00060-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3705f05e2ddc4072e588670a3ce7bb9be1570118 --- /dev/null +++ b/model-00060-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92746b791118973a6777c153ea2d5a53686cdcb75ce5788ac03fa7f20a7bf42d +size 1711309864 diff --git a/model-00061-of-00081.safetensors b/model-00061-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a51dad17fa52736c0507786f8022972df34485b2 --- /dev/null +++ b/model-00061-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:697f32859f837126e9a2b498f715ebb2ab8c188bfa1ce2e2c73d011d3107d84e +size 1711309864 diff --git a/model-00062-of-00081.safetensors b/model-00062-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e5c0efd1265e9a8c2145ff7327d54ddca2250303 --- /dev/null +++ b/model-00062-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50bbd16d0ff06204cb52e94aadcfee69b8780349f5b81768a1fc21ef9a7893ad +size 1711309864 diff --git a/model-00063-of-00081.safetensors b/model-00063-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..850dd3b20f4b882c5640a31450e743c42cd6a151 --- /dev/null +++ b/model-00063-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44f6acd313d3cb350daaf46a6c50241f74bf17a2cd709cf66323c14972fe4825 +size 1711309864 diff --git a/model-00064-of-00081.safetensors b/model-00064-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8a77537b556627018229ef15a31242a09f12b4a3 --- /dev/null +++ b/model-00064-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71416462b87a83553175ca7ab1ce68d089ea54f7b9ade60703fbfb66bbf368ea +size 1711309864 diff --git a/model-00065-of-00081.safetensors b/model-00065-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0ce44fe25b684f0e1d07afed0025bd98127bc044 --- /dev/null +++ b/model-00065-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8118c90022f9bc1ef891ce98667913ec3dae1b1f81d6f7725966b994cc1d51f +size 1711309864 diff --git a/model-00066-of-00081.safetensors b/model-00066-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..665bd6576eedb75f389f8d6132efc38f0d09d8b7 --- /dev/null +++ b/model-00066-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1bc3c1944ff7dda15436522901680f854e845cb7a4e72f5745eb0fd9c13b05a +size 1711309864 diff --git a/model-00067-of-00081.safetensors b/model-00067-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2f628a86e43882bbb069eeb18cb4716b684db543 --- /dev/null +++ b/model-00067-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bab9b4cbaa945e0c519b54e56be4525e094d5835a0399f197ada79503a995397 +size 1711309864 diff --git a/model-00068-of-00081.safetensors b/model-00068-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ab0abf40c49ef80b311cd43db9f06e606d363daa --- /dev/null +++ b/model-00068-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93e3ed8b763bafa7cc3bc89ff6404fb3cee6b433c76fddf24626081f12bf7ee5 +size 1711309864 diff --git a/model-00069-of-00081.safetensors b/model-00069-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9af2fa87a033f72a8ba3743b68ae6bb7f656accb --- /dev/null +++ b/model-00069-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcf853bcd0f7719fb56db7b9349542d61062fce53c3afdff2409efcdf5ed86db +size 1711309864 diff --git a/model-00070-of-00081.safetensors b/model-00070-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7464bd1b8e0a1972f9c945a747a84d69e7695eb1 --- /dev/null +++ b/model-00070-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f69ab45c17877c5c50ed90af1ba7331792b23db7a7385b0600cb2110fce872b +size 1711309864 diff --git a/model-00071-of-00081.safetensors b/model-00071-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6cc8b156cbc8c69d3f42642575860fe5900ee0be --- /dev/null +++ b/model-00071-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e4a7f761fdfe7d710a0342dd1c8a34ae72595213a68239e76c287547bd6d4c0 +size 1711309864 diff --git a/model-00072-of-00081.safetensors b/model-00072-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7f65a4ab3b8adffcc88bf93ad12f8949e4a57a9b --- /dev/null +++ b/model-00072-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:245dae021312c80a2e05dcf33ca517b20ae493c5be9fe4f440b2df01d4bd11a1 +size 1711309864 diff --git a/model-00073-of-00081.safetensors b/model-00073-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..90bd407ed360b4aa9c0cf7493d5c69676fa7f7f7 --- /dev/null +++ b/model-00073-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf4c232217d2d1c0907459ca4323d54fcdb61c39fd7ae063ed049bb50ff9f361 +size 1711309864 diff --git a/model-00074-of-00081.safetensors b/model-00074-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3208266de9d85a9f62bf4cf8cd9f2e778817cb26 --- /dev/null +++ b/model-00074-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d540af0d32aa2b0ea78d4b3012e732d12859ea6795cdd5657b414a8253b36794 +size 1711309864 diff --git a/model-00075-of-00081.safetensors b/model-00075-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2a65340f238184e7b149133782a50d997fea252b --- /dev/null +++ b/model-00075-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33bbbae13a2db6226c2c20f403f0133a64ae4fc3d9d791da04fda43dbe31063f +size 1711309864 diff --git a/model-00076-of-00081.safetensors b/model-00076-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8f0d23bcc82ff18e8ae625064451f0137806f444 --- /dev/null +++ b/model-00076-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c57c3679fc983836aae99f0b0d6c2e9b64ebba2ae7c75df321ac78ddc6eeafe0 +size 1711309864 diff --git a/model-00077-of-00081.safetensors b/model-00077-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3f5aa164099220b41ee0ddf7386b4642246e8928 --- /dev/null +++ b/model-00077-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a6827e7daf34945b9ce69e3419fc8b2960a6099ef7ed9cacaeb581790b59cfe +size 1711309864 diff --git a/model-00078-of-00081.safetensors b/model-00078-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..567a1f86705a4a8d27963f29df6f5849caf2c8d4 --- /dev/null +++ b/model-00078-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3629f5c62e2633868c1c19dbc3c663442ffb132866626fee4b838129e0806496 +size 1711309864 diff --git a/model-00079-of-00081.safetensors b/model-00079-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d47b9dda7f1f6c34f4a65346b1450948a1a1ac61 --- /dev/null +++ b/model-00079-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf899820e4f48affbbe8f54d120c66a3789de1bf8133c314a69e7e3cd6a68307 +size 1711309864 diff --git a/model-00080-of-00081.safetensors b/model-00080-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1bfdb39bd09907042664c4c70cb32dac3ce5b99c --- /dev/null +++ b/model-00080-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d37b392916f8320b4e95b7a73680558b4aba4e191273c17a6d0880570344b2a4 +size 1711309864 diff --git a/model-00081-of-00081.safetensors b/model-00081-of-00081.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f2b3d1de897fa52b81970f86af4f9491c22607d3 --- /dev/null +++ b/model-00081-of-00081.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d01e1e4010eeba8520ffae35d080e3b0de2ba2cf82504d0bfcead53b52fe0e32 +size 994099752 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..20a7efed1fd8708cdf88784b67dfe3de57209c93 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,730 @@ +{ + "metadata": { + "total_size": 137953296384 + }, + "weight_map": { + "lm_head.weight": "model-00081-of-00081.safetensors", + "model.embed_tokens.weight": "model-00001-of-00081.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00081.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00002-of-00081.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00081.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00081.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00081.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00081.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00081.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00081.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00081.safetensors", + "model.layers.1.input_layernorm.weight": "model-00003-of-00081.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00003-of-00081.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00081.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00081.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00081.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00081.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00081.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00081.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00081.safetensors", + "model.layers.10.input_layernorm.weight": "model-00012-of-00081.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00012-of-00081.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00011-of-00081.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00011-of-00081.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00012-of-00081.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00011-of-00081.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00011-of-00081.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00011-of-00081.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00011-of-00081.safetensors", + "model.layers.11.input_layernorm.weight": "model-00013-of-00081.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00013-of-00081.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00012-of-00081.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00012-of-00081.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00013-of-00081.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00012-of-00081.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00012-of-00081.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00012-of-00081.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00012-of-00081.safetensors", + "model.layers.12.input_layernorm.weight": "model-00014-of-00081.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00014-of-00081.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00013-of-00081.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00013-of-00081.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00014-of-00081.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00013-of-00081.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00013-of-00081.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00013-of-00081.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00013-of-00081.safetensors", + "model.layers.13.input_layernorm.weight": "model-00015-of-00081.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00015-of-00081.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00014-of-00081.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00014-of-00081.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00015-of-00081.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00014-of-00081.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00014-of-00081.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00014-of-00081.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00014-of-00081.safetensors", + "model.layers.14.input_layernorm.weight": "model-00016-of-00081.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00016-of-00081.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00015-of-00081.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00015-of-00081.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00016-of-00081.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00015-of-00081.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00015-of-00081.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00015-of-00081.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00015-of-00081.safetensors", + "model.layers.15.input_layernorm.weight": "model-00017-of-00081.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00017-of-00081.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00016-of-00081.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00016-of-00081.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00017-of-00081.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00016-of-00081.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00016-of-00081.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00016-of-00081.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00016-of-00081.safetensors", + "model.layers.16.input_layernorm.weight": "model-00018-of-00081.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00018-of-00081.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00017-of-00081.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00017-of-00081.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00018-of-00081.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00017-of-00081.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00017-of-00081.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00017-of-00081.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00017-of-00081.safetensors", + "model.layers.17.input_layernorm.weight": "model-00019-of-00081.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00019-of-00081.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00018-of-00081.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00018-of-00081.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00019-of-00081.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00018-of-00081.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00018-of-00081.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00018-of-00081.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00018-of-00081.safetensors", + "model.layers.18.input_layernorm.weight": "model-00020-of-00081.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00020-of-00081.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00019-of-00081.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00019-of-00081.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00020-of-00081.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00019-of-00081.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00019-of-00081.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00019-of-00081.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00019-of-00081.safetensors", + "model.layers.19.input_layernorm.weight": "model-00021-of-00081.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00021-of-00081.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00020-of-00081.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00020-of-00081.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00021-of-00081.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00020-of-00081.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00020-of-00081.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00020-of-00081.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00020-of-00081.safetensors", + "model.layers.2.input_layernorm.weight": "model-00004-of-00081.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00004-of-00081.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00003-of-00081.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00081.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00004-of-00081.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00081.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00081.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00081.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00081.safetensors", + "model.layers.20.input_layernorm.weight": "model-00022-of-00081.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00022-of-00081.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00021-of-00081.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00021-of-00081.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00022-of-00081.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00021-of-00081.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00021-of-00081.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00021-of-00081.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00021-of-00081.safetensors", + "model.layers.21.input_layernorm.weight": "model-00023-of-00081.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00023-of-00081.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00022-of-00081.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00022-of-00081.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00023-of-00081.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00022-of-00081.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00022-of-00081.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00022-of-00081.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00022-of-00081.safetensors", + "model.layers.22.input_layernorm.weight": "model-00024-of-00081.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00024-of-00081.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00023-of-00081.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00023-of-00081.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00024-of-00081.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00023-of-00081.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00023-of-00081.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00023-of-00081.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00023-of-00081.safetensors", + "model.layers.23.input_layernorm.weight": "model-00025-of-00081.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00025-of-00081.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00024-of-00081.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00024-of-00081.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00025-of-00081.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00024-of-00081.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00024-of-00081.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00024-of-00081.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00024-of-00081.safetensors", + "model.layers.24.input_layernorm.weight": "model-00026-of-00081.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00026-of-00081.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00025-of-00081.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00025-of-00081.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00026-of-00081.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00025-of-00081.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00025-of-00081.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00025-of-00081.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00025-of-00081.safetensors", + "model.layers.25.input_layernorm.weight": "model-00027-of-00081.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00027-of-00081.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00026-of-00081.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00026-of-00081.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00027-of-00081.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00026-of-00081.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00026-of-00081.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00026-of-00081.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00026-of-00081.safetensors", + "model.layers.26.input_layernorm.weight": "model-00028-of-00081.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00028-of-00081.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00027-of-00081.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00027-of-00081.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00028-of-00081.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00027-of-00081.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00027-of-00081.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00027-of-00081.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00027-of-00081.safetensors", + "model.layers.27.input_layernorm.weight": "model-00029-of-00081.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00029-of-00081.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00028-of-00081.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00028-of-00081.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00029-of-00081.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00028-of-00081.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00028-of-00081.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00028-of-00081.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00028-of-00081.safetensors", + "model.layers.28.input_layernorm.weight": "model-00030-of-00081.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00030-of-00081.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00029-of-00081.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00029-of-00081.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00030-of-00081.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00029-of-00081.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00029-of-00081.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00029-of-00081.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00029-of-00081.safetensors", + "model.layers.29.input_layernorm.weight": "model-00031-of-00081.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00031-of-00081.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00030-of-00081.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00030-of-00081.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00031-of-00081.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00030-of-00081.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00030-of-00081.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00030-of-00081.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00030-of-00081.safetensors", + "model.layers.3.input_layernorm.weight": "model-00005-of-00081.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00005-of-00081.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00081.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00004-of-00081.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00005-of-00081.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00081.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00081.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00081.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00081.safetensors", + "model.layers.30.input_layernorm.weight": "model-00032-of-00081.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00032-of-00081.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00031-of-00081.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00031-of-00081.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00032-of-00081.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00031-of-00081.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00031-of-00081.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00031-of-00081.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00031-of-00081.safetensors", + "model.layers.31.input_layernorm.weight": "model-00033-of-00081.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00033-of-00081.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00032-of-00081.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00032-of-00081.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00033-of-00081.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00032-of-00081.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00032-of-00081.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00032-of-00081.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00032-of-00081.safetensors", + "model.layers.32.input_layernorm.weight": "model-00034-of-00081.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00034-of-00081.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00033-of-00081.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00033-of-00081.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00034-of-00081.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00033-of-00081.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00033-of-00081.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00033-of-00081.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00033-of-00081.safetensors", + "model.layers.33.input_layernorm.weight": "model-00035-of-00081.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00035-of-00081.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00034-of-00081.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00034-of-00081.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00035-of-00081.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00034-of-00081.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00034-of-00081.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00034-of-00081.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00034-of-00081.safetensors", + "model.layers.34.input_layernorm.weight": "model-00036-of-00081.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00036-of-00081.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00035-of-00081.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00035-of-00081.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00036-of-00081.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00035-of-00081.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00035-of-00081.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00035-of-00081.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00035-of-00081.safetensors", + "model.layers.35.input_layernorm.weight": "model-00037-of-00081.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00037-of-00081.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00036-of-00081.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00036-of-00081.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00037-of-00081.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00036-of-00081.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00036-of-00081.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00036-of-00081.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00036-of-00081.safetensors", + "model.layers.36.input_layernorm.weight": "model-00038-of-00081.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00038-of-00081.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00037-of-00081.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00037-of-00081.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00038-of-00081.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00037-of-00081.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00037-of-00081.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00037-of-00081.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00037-of-00081.safetensors", + "model.layers.37.input_layernorm.weight": "model-00039-of-00081.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00039-of-00081.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00038-of-00081.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00038-of-00081.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00039-of-00081.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00038-of-00081.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00038-of-00081.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00038-of-00081.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00038-of-00081.safetensors", + "model.layers.38.input_layernorm.weight": "model-00040-of-00081.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00040-of-00081.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00039-of-00081.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00039-of-00081.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00040-of-00081.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00039-of-00081.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00039-of-00081.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00039-of-00081.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00039-of-00081.safetensors", + "model.layers.39.input_layernorm.weight": "model-00041-of-00081.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00041-of-00081.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00040-of-00081.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00040-of-00081.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00041-of-00081.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00040-of-00081.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00040-of-00081.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00040-of-00081.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00040-of-00081.safetensors", + "model.layers.4.input_layernorm.weight": "model-00006-of-00081.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00006-of-00081.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00005-of-00081.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00005-of-00081.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00006-of-00081.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00081.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00081.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00081.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00081.safetensors", + "model.layers.40.input_layernorm.weight": "model-00042-of-00081.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00042-of-00081.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00041-of-00081.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00041-of-00081.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00042-of-00081.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00041-of-00081.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00041-of-00081.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00041-of-00081.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00041-of-00081.safetensors", + "model.layers.41.input_layernorm.weight": "model-00043-of-00081.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00043-of-00081.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00042-of-00081.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00042-of-00081.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00043-of-00081.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00042-of-00081.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00042-of-00081.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00042-of-00081.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00042-of-00081.safetensors", + "model.layers.42.input_layernorm.weight": "model-00044-of-00081.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00044-of-00081.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00043-of-00081.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00043-of-00081.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00044-of-00081.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00043-of-00081.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00043-of-00081.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00043-of-00081.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00043-of-00081.safetensors", + "model.layers.43.input_layernorm.weight": "model-00045-of-00081.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00045-of-00081.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00044-of-00081.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00044-of-00081.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00045-of-00081.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00044-of-00081.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00044-of-00081.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00044-of-00081.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00044-of-00081.safetensors", + "model.layers.44.input_layernorm.weight": "model-00046-of-00081.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00046-of-00081.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00045-of-00081.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00045-of-00081.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00046-of-00081.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00045-of-00081.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00045-of-00081.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00045-of-00081.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00045-of-00081.safetensors", + "model.layers.45.input_layernorm.weight": "model-00047-of-00081.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00047-of-00081.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00046-of-00081.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00046-of-00081.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00047-of-00081.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00046-of-00081.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00046-of-00081.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00046-of-00081.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00046-of-00081.safetensors", + "model.layers.46.input_layernorm.weight": "model-00048-of-00081.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00048-of-00081.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00047-of-00081.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00047-of-00081.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00048-of-00081.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00047-of-00081.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00047-of-00081.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00047-of-00081.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00047-of-00081.safetensors", + "model.layers.47.input_layernorm.weight": "model-00049-of-00081.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00049-of-00081.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00048-of-00081.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00048-of-00081.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00049-of-00081.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00048-of-00081.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00048-of-00081.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00048-of-00081.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00048-of-00081.safetensors", + "model.layers.48.input_layernorm.weight": "model-00050-of-00081.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00050-of-00081.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00049-of-00081.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00049-of-00081.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00050-of-00081.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00049-of-00081.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00049-of-00081.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00049-of-00081.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00049-of-00081.safetensors", + "model.layers.49.input_layernorm.weight": "model-00051-of-00081.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00051-of-00081.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00050-of-00081.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00050-of-00081.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00051-of-00081.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00050-of-00081.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00050-of-00081.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00050-of-00081.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00050-of-00081.safetensors", + "model.layers.5.input_layernorm.weight": "model-00007-of-00081.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00007-of-00081.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00006-of-00081.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00006-of-00081.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00007-of-00081.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00006-of-00081.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00006-of-00081.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00006-of-00081.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00006-of-00081.safetensors", + "model.layers.50.input_layernorm.weight": "model-00052-of-00081.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00052-of-00081.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00051-of-00081.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00051-of-00081.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00052-of-00081.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00051-of-00081.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00051-of-00081.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00051-of-00081.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00051-of-00081.safetensors", + "model.layers.51.input_layernorm.weight": "model-00053-of-00081.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00053-of-00081.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00052-of-00081.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00052-of-00081.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00053-of-00081.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00052-of-00081.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00052-of-00081.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00052-of-00081.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00052-of-00081.safetensors", + "model.layers.52.input_layernorm.weight": "model-00054-of-00081.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00054-of-00081.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00053-of-00081.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00053-of-00081.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00054-of-00081.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00053-of-00081.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00053-of-00081.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00053-of-00081.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00053-of-00081.safetensors", + "model.layers.53.input_layernorm.weight": "model-00055-of-00081.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00055-of-00081.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00054-of-00081.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00054-of-00081.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00055-of-00081.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00054-of-00081.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00054-of-00081.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00054-of-00081.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00054-of-00081.safetensors", + "model.layers.54.input_layernorm.weight": "model-00056-of-00081.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00056-of-00081.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00055-of-00081.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00055-of-00081.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00056-of-00081.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00055-of-00081.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00055-of-00081.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00055-of-00081.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00055-of-00081.safetensors", + "model.layers.55.input_layernorm.weight": "model-00057-of-00081.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00057-of-00081.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00056-of-00081.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00056-of-00081.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00057-of-00081.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00056-of-00081.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00056-of-00081.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00056-of-00081.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00056-of-00081.safetensors", + "model.layers.56.input_layernorm.weight": "model-00058-of-00081.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00058-of-00081.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00057-of-00081.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00057-of-00081.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00058-of-00081.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00057-of-00081.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00057-of-00081.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00057-of-00081.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00057-of-00081.safetensors", + "model.layers.57.input_layernorm.weight": "model-00059-of-00081.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00059-of-00081.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00058-of-00081.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00058-of-00081.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00059-of-00081.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00058-of-00081.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00058-of-00081.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00058-of-00081.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00058-of-00081.safetensors", + "model.layers.58.input_layernorm.weight": "model-00060-of-00081.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00060-of-00081.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00059-of-00081.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00059-of-00081.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00060-of-00081.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00059-of-00081.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00059-of-00081.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00059-of-00081.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00059-of-00081.safetensors", + "model.layers.59.input_layernorm.weight": "model-00061-of-00081.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00061-of-00081.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00060-of-00081.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00060-of-00081.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00061-of-00081.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00060-of-00081.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00060-of-00081.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00060-of-00081.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00060-of-00081.safetensors", + "model.layers.6.input_layernorm.weight": "model-00008-of-00081.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00008-of-00081.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00007-of-00081.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00007-of-00081.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00008-of-00081.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00007-of-00081.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00007-of-00081.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00007-of-00081.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00007-of-00081.safetensors", + "model.layers.60.input_layernorm.weight": "model-00062-of-00081.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00062-of-00081.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00061-of-00081.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00061-of-00081.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00062-of-00081.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00061-of-00081.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00061-of-00081.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00061-of-00081.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00061-of-00081.safetensors", + "model.layers.61.input_layernorm.weight": "model-00063-of-00081.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00063-of-00081.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00062-of-00081.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00062-of-00081.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00063-of-00081.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00062-of-00081.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00062-of-00081.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00062-of-00081.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00062-of-00081.safetensors", + "model.layers.62.input_layernorm.weight": "model-00064-of-00081.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00064-of-00081.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00063-of-00081.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00063-of-00081.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00064-of-00081.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00063-of-00081.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00063-of-00081.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00063-of-00081.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00063-of-00081.safetensors", + "model.layers.63.input_layernorm.weight": "model-00065-of-00081.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00065-of-00081.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00064-of-00081.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00064-of-00081.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00065-of-00081.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00064-of-00081.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00064-of-00081.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00064-of-00081.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00064-of-00081.safetensors", + "model.layers.64.input_layernorm.weight": "model-00066-of-00081.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00066-of-00081.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00065-of-00081.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00065-of-00081.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00066-of-00081.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00065-of-00081.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00065-of-00081.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00065-of-00081.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00065-of-00081.safetensors", + "model.layers.65.input_layernorm.weight": "model-00067-of-00081.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00067-of-00081.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00066-of-00081.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00066-of-00081.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00067-of-00081.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00066-of-00081.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00066-of-00081.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00066-of-00081.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00066-of-00081.safetensors", + "model.layers.66.input_layernorm.weight": "model-00068-of-00081.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00068-of-00081.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00067-of-00081.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00067-of-00081.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00068-of-00081.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00067-of-00081.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00067-of-00081.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00067-of-00081.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00067-of-00081.safetensors", + "model.layers.67.input_layernorm.weight": "model-00069-of-00081.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00069-of-00081.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00068-of-00081.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00068-of-00081.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00069-of-00081.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00068-of-00081.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00068-of-00081.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00068-of-00081.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00068-of-00081.safetensors", + "model.layers.68.input_layernorm.weight": "model-00070-of-00081.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00070-of-00081.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00069-of-00081.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00069-of-00081.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00070-of-00081.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00069-of-00081.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00069-of-00081.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00069-of-00081.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00069-of-00081.safetensors", + "model.layers.69.input_layernorm.weight": "model-00071-of-00081.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00071-of-00081.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00070-of-00081.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00070-of-00081.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00071-of-00081.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00070-of-00081.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00070-of-00081.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00070-of-00081.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00070-of-00081.safetensors", + "model.layers.7.input_layernorm.weight": "model-00009-of-00081.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00009-of-00081.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00008-of-00081.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00008-of-00081.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00009-of-00081.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00008-of-00081.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00008-of-00081.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00008-of-00081.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00008-of-00081.safetensors", + "model.layers.70.input_layernorm.weight": "model-00072-of-00081.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00072-of-00081.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00071-of-00081.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00071-of-00081.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00072-of-00081.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00071-of-00081.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00071-of-00081.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00071-of-00081.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00071-of-00081.safetensors", + "model.layers.71.input_layernorm.weight": "model-00073-of-00081.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00073-of-00081.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00072-of-00081.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00072-of-00081.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00073-of-00081.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00072-of-00081.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00072-of-00081.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00072-of-00081.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00072-of-00081.safetensors", + "model.layers.72.input_layernorm.weight": "model-00074-of-00081.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00074-of-00081.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00073-of-00081.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00073-of-00081.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00074-of-00081.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00073-of-00081.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00073-of-00081.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00073-of-00081.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00073-of-00081.safetensors", + "model.layers.73.input_layernorm.weight": "model-00075-of-00081.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00075-of-00081.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00074-of-00081.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00074-of-00081.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00075-of-00081.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00074-of-00081.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00074-of-00081.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00074-of-00081.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00074-of-00081.safetensors", + "model.layers.74.input_layernorm.weight": "model-00076-of-00081.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00076-of-00081.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00075-of-00081.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00075-of-00081.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00076-of-00081.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00075-of-00081.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00075-of-00081.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00075-of-00081.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00075-of-00081.safetensors", + "model.layers.75.input_layernorm.weight": "model-00077-of-00081.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00077-of-00081.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00076-of-00081.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00076-of-00081.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00077-of-00081.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00076-of-00081.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00076-of-00081.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00076-of-00081.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00076-of-00081.safetensors", + "model.layers.76.input_layernorm.weight": "model-00078-of-00081.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00078-of-00081.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00077-of-00081.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00077-of-00081.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00078-of-00081.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00077-of-00081.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00077-of-00081.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00077-of-00081.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00077-of-00081.safetensors", + "model.layers.77.input_layernorm.weight": "model-00079-of-00081.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00079-of-00081.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00078-of-00081.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00078-of-00081.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00079-of-00081.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00078-of-00081.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00078-of-00081.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00078-of-00081.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00078-of-00081.safetensors", + "model.layers.78.input_layernorm.weight": "model-00080-of-00081.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00080-of-00081.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00079-of-00081.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00079-of-00081.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00080-of-00081.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00079-of-00081.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00079-of-00081.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00079-of-00081.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00079-of-00081.safetensors", + "model.layers.79.input_layernorm.weight": "model-00081-of-00081.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00081-of-00081.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00080-of-00081.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00080-of-00081.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00081-of-00081.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00080-of-00081.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00080-of-00081.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00080-of-00081.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00080-of-00081.safetensors", + "model.layers.8.input_layernorm.weight": "model-00010-of-00081.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00010-of-00081.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00009-of-00081.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00009-of-00081.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00010-of-00081.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00009-of-00081.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00009-of-00081.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00009-of-00081.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00009-of-00081.safetensors", + "model.layers.9.input_layernorm.weight": "model-00011-of-00081.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00011-of-00081.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00010-of-00081.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00010-of-00081.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00011-of-00081.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00010-of-00081.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00010-of-00081.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00010-of-00081.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00010-of-00081.safetensors", + "model.norm.weight": "model-00081-of-00081.safetensors" + } +} diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..5819ea25d7b1b6340063a0629c2143c44b0452da --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,30 @@ +{ + "bos_token": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false + } +} diff --git a/tokenizer.model b/tokenizer.model new file mode 100644 index 0000000000000000000000000000000000000000..6c00c742ce03c627d6cd5b795984876fa49fa899 --- /dev/null +++ b/tokenizer.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347 +size 499723 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..c924a7970ffad36e88ef3161ebecebea932c1f3c --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,41 @@ +{ + "add_bos_token": true, + "add_eos_token": false, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": true, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "", + "clean_up_tokenization_spaces": false, + "eos_token": "", + "legacy": false, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "spaces_between_special_tokens": false, + "tokenizer_class": "LlamaTokenizer", + "unk_token": "", + "use_default_system_prompt": false +}