diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..c64ddcea32d66a7ca52fc77a4e835259173f91ad --- /dev/null +++ b/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "meta-llama/Meta-Llama-3-70b", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "eos_token_id": 128001, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 8192, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.40.0", + "use_cache": true, + "vocab_size": 128256 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..cc3991ac16d0bb2556f6c3c42b2801acb8381e17 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,9 @@ +{ + "bos_token_id": 128000, + "do_sample": true, + "eos_token_id": 128001, + "max_length": 4096, + "temperature": 0.6, + "top_p": 0.9, + "transformers_version": "4.40.0" +} diff --git a/model-00001-of-00082.safetensors b/model-00001-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8f4e3e127c4b9d46ecc006a224e6aaa4e8950a17 --- /dev/null +++ b/model-00001-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a44f6170b5c45de46d30e867ded1c181d446f8ef57420f4dce782fbc651bc9c +size 2101346440 diff --git a/model-00002-of-00082.safetensors b/model-00002-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..880a09718019f51cac0fde4391b753fb4bede4b5 --- /dev/null +++ b/model-00002-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e359ca37250544b461667bfa8a0a856350c4f15b0e1803ee00a589f7c4c0dc83 +size 1879082360 diff --git a/model-00003-of-00082.safetensors b/model-00003-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bb7c7550db750f321eecf61b9ca6d5c8a7ce742f --- /dev/null +++ b/model-00003-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dad0c53d913ec4cfe3080665d0824973c72ab77ea03f217704782b8eec6584da +size 1845527696 diff --git a/model-00004-of-00082.safetensors b/model-00004-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..61f46966be527d58cf3780580ea83a463efe99b0 --- /dev/null +++ b/model-00004-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:effa4395d3b3245bd15dac953f4545407a491cbf9ac53a521fa5b15f2391e887 +size 1711309856 diff --git a/model-00005-of-00082.safetensors b/model-00005-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b0053232c56940495327cba956b69fe02f520cf2 --- /dev/null +++ b/model-00005-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a7e1c3840e7b8fbd798fdbcf3c0b5e030c2abc258394ac57ec5b1da84a37aea +size 1711309856 diff --git a/model-00006-of-00082.safetensors b/model-00006-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..87bf6f20e91f8cb29723c9f9cbaf6a7a0185cf3c --- /dev/null +++ b/model-00006-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26c11e07f4eea041319c91cdd2f6dfcea48074143fd65a512891d4a965e2e693 +size 1711309856 diff --git a/model-00007-of-00082.safetensors b/model-00007-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f6e1952d4f4ec4e94b7597582e4575c98b777e54 --- /dev/null +++ b/model-00007-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c693057343a67e1d66b27256a80c94ea22a266caf9d4f159f6dd6d3b0bf4010e +size 1711309856 diff --git a/model-00008-of-00082.safetensors b/model-00008-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d9c3f970e16f8f5fc0b77440a65ed38604af7915 --- /dev/null +++ b/model-00008-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b103d5fca4c64a8c616beb9a63891198a245aa103fd01c0d7b4974098a3d5284 +size 1711309856 diff --git a/model-00009-of-00082.safetensors b/model-00009-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..42bba8fe7b60236e70ad549eaa523c7dc8ff15d7 --- /dev/null +++ b/model-00009-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df260515742dd0ae093198610bfcc079a17f04cfc914792b6d15b52c38f73532 +size 1711309856 diff --git a/model-00010-of-00082.safetensors b/model-00010-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..db6b01f60f76bf374ab634909b907197616c9ec5 --- /dev/null +++ b/model-00010-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:014206f0598974978b64f3bc46df052298e185dc8140e429282826c064dd357f +size 1711309856 diff --git a/model-00011-of-00082.safetensors b/model-00011-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b94c9faddeb387ff28970faeb75c8263d5b21a2e --- /dev/null +++ b/model-00011-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:154883612759e700da66d49fbf59b62e362a691949f9edbddea88659d2748a9a +size 1711309856 diff --git a/model-00012-of-00082.safetensors b/model-00012-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f8b6f4958a6853f7a0e9cfd753c0aa6e76109575 --- /dev/null +++ b/model-00012-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19a4f08b65ccb750bd2f14573bd77a215e1ea69668b222116e07b450752c245f +size 1711309864 diff --git a/model-00013-of-00082.safetensors b/model-00013-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1e900591cf3b884f3f40ea674bd0c71d86765c28 --- /dev/null +++ b/model-00013-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49d8b3746b41e787b4ef55b2b0a9d6b57c67d8b0ca12ce56d45596556f9b13a6 +size 1711309864 diff --git a/model-00014-of-00082.safetensors b/model-00014-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6203a2b6e24c6f7fa4f6a9283189b769aab24922 --- /dev/null +++ b/model-00014-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:114888ec230a1690c3903d89b2905a93e457ceb5ae00eebda2af1b8d2654a5f6 +size 1711309864 diff --git a/model-00015-of-00082.safetensors b/model-00015-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a4d8ccdbadf57c55318377a78bee4e899dcd7120 --- /dev/null +++ b/model-00015-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7120a7559c8225bc2103a5e548c50fc901ceaf486715f9f9a62e2867d8ba4ad5 +size 1711309864 diff --git a/model-00016-of-00082.safetensors b/model-00016-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d04666458beee062d1febb000ac77fa8d5a1cada --- /dev/null +++ b/model-00016-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11166c314736d9b38e65e54e71dc1443f783da12163f25e8ce62a3348fb56b9c +size 1711309864 diff --git a/model-00017-of-00082.safetensors b/model-00017-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..66574d8eb05f83befc5c9e56f2fa89b8c615e4c6 --- /dev/null +++ b/model-00017-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88435dfa5cea978eb5cd38cc455da10cf2a612aec3ca082fd28a50a55c0a0c23 +size 1711309864 diff --git a/model-00018-of-00082.safetensors b/model-00018-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..31237408c5aa272300d47cad4307943969328de4 --- /dev/null +++ b/model-00018-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79f0309e33ed80297909fc61029664e488ecfd5490827daa825ea598383efe59 +size 1711309864 diff --git a/model-00019-of-00082.safetensors b/model-00019-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b1848f9bc9800e7ed9b41dd26c0676d4f44551ee --- /dev/null +++ b/model-00019-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fd7e5e738be8dfcef4de6e4c9cdd09f5ce0ed0c2c6c205099dc69c0670653b5 +size 1711309864 diff --git a/model-00020-of-00082.safetensors b/model-00020-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6cbde2bce710e5cb6c0551a40ef2ee4bcfb95668 --- /dev/null +++ b/model-00020-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a227b63b67f850ab8ba10581b672ff001642d3f67b32146c6bff4f95ba82549 +size 1711309864 diff --git a/model-00021-of-00082.safetensors b/model-00021-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1b191590ae307d0b65d563649759ea973e8bf288 --- /dev/null +++ b/model-00021-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:148fd39fc5dade47b47c02eee7c97d7595493f57d471bfe4bf8d84fb4f820119 +size 1711309864 diff --git a/model-00022-of-00082.safetensors b/model-00022-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dd5a2a7bb2a0ddfc5444fa08f8bb140b2279a108 --- /dev/null +++ b/model-00022-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9104c4e57ac66cbf8c39cf2564df4dc2b26ab8c7cc1fee47d7867d49aee0183 +size 1711309864 diff --git a/model-00023-of-00082.safetensors b/model-00023-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ab19be9b5dc2a01c4d9f8d4a4fcbbe485fc9e476 --- /dev/null +++ b/model-00023-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2ea85a09ef3e6cf4896c8ba2fbd89ba79c933216685a1ef511ed24e80157c07 +size 1711309864 diff --git a/model-00024-of-00082.safetensors b/model-00024-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..11b922a7180a02d27c0d317c78461d5e9c683b16 --- /dev/null +++ b/model-00024-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04b13812f03ca490aeb65b3d078eaa7d785bb3b5563eb0b1c00ac2f03b7a4ed3 +size 1711309864 diff --git a/model-00025-of-00082.safetensors b/model-00025-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3603f844f2f6965d4de4cc9004f94841f4979d7a --- /dev/null +++ b/model-00025-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:431171b14b395fc96d4cdb2284ef0826b0127b3b0997a051d744f2ea51aa8593 +size 1711309864 diff --git a/model-00026-of-00082.safetensors b/model-00026-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6023f1ed02d990b002bf53b513fc9d3d0cc302f8 --- /dev/null +++ b/model-00026-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fe34e6ad3713d67fd1e6791da65ef218f66f14cdf60bb80ba0123240b454a31 +size 1711309864 diff --git a/model-00027-of-00082.safetensors b/model-00027-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7f097642d97308d72db761e4c62cb9e6601aef49 --- /dev/null +++ b/model-00027-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:088a02150f60f121fecb325141bd2b6e63bc40ed4b53b4447f9427b9d827a028 +size 1711309864 diff --git a/model-00028-of-00082.safetensors b/model-00028-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3ed4ed07c5b5685db22ce367874654be592baa24 --- /dev/null +++ b/model-00028-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7c82e69badd2fb080c8467fd255c999ee2abcca9746d708ccccc89f1cbd5033 +size 1711309864 diff --git a/model-00029-of-00082.safetensors b/model-00029-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c833064ee1000ccfe1961e4e5462acd0e76d9051 --- /dev/null +++ b/model-00029-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85280d80488d9ea7b7416eb41076b604d44a3c88a1d905299be46f3da128e9f6 +size 1711309864 diff --git a/model-00030-of-00082.safetensors b/model-00030-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7332a32de15c67adb8549c063beefa52db981aa8 --- /dev/null +++ b/model-00030-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92e723771b6f1b11bdfd383e24c8badbf34a0285311f228a8875a9444cf9fd00 +size 1711309864 diff --git a/model-00031-of-00082.safetensors b/model-00031-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..56e65192e003b0d95e0b0067d519f5f8c93be0ab --- /dev/null +++ b/model-00031-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27fe81bc050251d907cd031361c9bb46f1854d752c75cb3bc3aa13d14ba6ccc6 +size 1711309864 diff --git a/model-00032-of-00082.safetensors b/model-00032-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dcb157a174e3e8ca489c9425fd98af4d791a4a7c --- /dev/null +++ b/model-00032-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fed9a187a99afeab6295f26695096a8a47aa27fc551a83ea4ec29d52540aa7e1 +size 1711309864 diff --git a/model-00033-of-00082.safetensors b/model-00033-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..edc8dcb2a2102dfea34e68b1fba5e1fe38c04bec --- /dev/null +++ b/model-00033-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aaa84e89c9c655619e699f903937f925b10a8c5de9c6d2d5b124d9c4b7ee8f95 +size 1711309864 diff --git a/model-00034-of-00082.safetensors b/model-00034-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b1d6b32976c030c455a5889c5d05682026370607 --- /dev/null +++ b/model-00034-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7bff946fd66d2ad9f7b7a3d01b7e8c5ecadf74b4e522493a503ce6140f5e510a +size 1711309864 diff --git a/model-00035-of-00082.safetensors b/model-00035-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e87b08504c317d0abe254574ca787201a6d5711b --- /dev/null +++ b/model-00035-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e017c2d00e430b85fe74dbdf74ea24e3fe063aad412d95b3d1e3fe4937b6014 +size 1711309864 diff --git a/model-00036-of-00082.safetensors b/model-00036-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ef9c4298dae3d1ba232a2bcc543a094202c3f2d3 --- /dev/null +++ b/model-00036-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c475be1d36743e8351e146ab35325d2269496e86979da36b36209455e8f8558 +size 1711309864 diff --git a/model-00037-of-00082.safetensors b/model-00037-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f0c0952dd92fe2e8e5232c57120511addd47a2c5 --- /dev/null +++ b/model-00037-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db4dec09550551c030090c3846968e01fe52bd0398a53888040d466b0435c0f0 +size 1711309864 diff --git a/model-00038-of-00082.safetensors b/model-00038-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d779383e6f2c2ade7d3f48dc0ee4182e29767e48 --- /dev/null +++ b/model-00038-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c45b8c530950f7ef9202bc0a6e242e9d43e964474822b810076da7ec6f4ca985 +size 1711309864 diff --git a/model-00039-of-00082.safetensors b/model-00039-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5ccdd8adae009df703c5bc5896f1fa190a64aba5 --- /dev/null +++ b/model-00039-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e9864752da89b448b65ef4acd9df98543b93abef624e52813ba57fc042b09f7 +size 1711309864 diff --git a/model-00040-of-00082.safetensors b/model-00040-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..23f9ae6ff1170dc4610bcc824f8711921a1f221c --- /dev/null +++ b/model-00040-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4370c9a852980afabceefee6addb1312142b13287387262da21f2c06cd701da4 +size 1711309864 diff --git a/model-00041-of-00082.safetensors b/model-00041-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b4a3d519dee9bd451042f4d619600804e9f80283 --- /dev/null +++ b/model-00041-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a95bd4915fdd112e30a671b41642a5056b25aaaa1a85e858e0147d3a1a8ae7c +size 1711309864 diff --git a/model-00042-of-00082.safetensors b/model-00042-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0f533618fa26debb9e3afad0aab70dd21ef4cf0e --- /dev/null +++ b/model-00042-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bf567b03795b877a2a1ed46e1e55542727ef5cdc435502ba0fd78c556e5d00f +size 1711309864 diff --git a/model-00043-of-00082.safetensors b/model-00043-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..692fd80f431c454f3d59296954531f4d7a484c6d --- /dev/null +++ b/model-00043-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:315541925d04bb54ec09debf47138e361fc2c6c56eba9da0c3e1c4cc1d853360 +size 1711309864 diff --git a/model-00044-of-00082.safetensors b/model-00044-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c9613395dd29f0bd092e4f90998878c6aa5ec1e3 --- /dev/null +++ b/model-00044-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7475810df9607338dac79f89d7cf0e3d47c7f1e8b2f450a250cc0f8d8a84a45 +size 1711309864 diff --git a/model-00045-of-00082.safetensors b/model-00045-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..976a9f7906c12bff66eaa5cd86078f23fee714c0 --- /dev/null +++ b/model-00045-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8924034c8a4260d6c18694ecfd0d99e5dd1c3b112717cc3a7188b4b936eb70f +size 1711309864 diff --git a/model-00046-of-00082.safetensors b/model-00046-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d8850c09455e0797751b2a0f291464fb7474484e --- /dev/null +++ b/model-00046-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2064f532bbc3b44641d69669d69d37a4066d741cbd743227acd1f46c64f58806 +size 1711309864 diff --git a/model-00047-of-00082.safetensors b/model-00047-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..15fb34d9c2df577c16e854da8ba0dec3dffc1b67 --- /dev/null +++ b/model-00047-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc5c717e5c20029b591412681e1903f80900e9c11823e19f5da54fcaa2f5f61d +size 1711309864 diff --git a/model-00048-of-00082.safetensors b/model-00048-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b718627217971087fcea51b11f9ceb00ce539b3c --- /dev/null +++ b/model-00048-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4c9cc30930218167f2045fb03730f79ade6feabd6532ff6b526462b56d4008a +size 1711309864 diff --git a/model-00049-of-00082.safetensors b/model-00049-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..430328a54c9df9855ee2ccaa7c5e532504e03034 --- /dev/null +++ b/model-00049-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9323b479817dd70b61ee7ff67e1d3fb5fb8d468360a40f256fea3d000566b4a +size 1711309864 diff --git a/model-00050-of-00082.safetensors b/model-00050-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..34323f2c442ec49505c0b22125ede9c85c0924f8 --- /dev/null +++ b/model-00050-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30ec6b332c1e9b8e38d3f71bffe815226000fc8381b51a22769695a3880d1998 +size 1711309864 diff --git a/model-00051-of-00082.safetensors b/model-00051-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..21c3eaa5a50ee30604968da1920a7055b088cb2f --- /dev/null +++ b/model-00051-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81c3bb42bdfc538165329304ce2cacaf833d1bf0287eb567587e7cdf6ed35c9d +size 1711309864 diff --git a/model-00052-of-00082.safetensors b/model-00052-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bb6fd3d6eae93cd61f0baa63bf731e05ae7859d4 --- /dev/null +++ b/model-00052-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e63e375d309ca5c54054f340460b08a3ea0713379e7a120f571e7372fc2c5915 +size 1711309864 diff --git a/model-00053-of-00082.safetensors b/model-00053-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f78867251bbd624cc04d1d2ffb4d2210705a7279 --- /dev/null +++ b/model-00053-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67e77b06207f706870be1447ba8e099a32f4b194fd3938edee22c19ecc2df24a +size 1711309864 diff --git a/model-00054-of-00082.safetensors b/model-00054-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..086919f0d06302bc203200eba641afe3f12e0aa6 --- /dev/null +++ b/model-00054-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f2c818ef3589f8d4234a468bd474aafd637c1f0c3b1c0bddc5e19005857ce92 +size 1711309864 diff --git a/model-00055-of-00082.safetensors b/model-00055-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4f0842f522557de2a3888ece1b1b0dfcae528e52 --- /dev/null +++ b/model-00055-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:488836da13711014f8a1d2084d91317e33473b2de87d9018a13f83fce3d4a7a8 +size 1711309864 diff --git a/model-00056-of-00082.safetensors b/model-00056-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..440e2cf64a67c5f2383f07b6b1a87c0023135f7b --- /dev/null +++ b/model-00056-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ef08ba384fb4a8e447a8b2e5316b38406d9cff9044c1242937dda4d2429c851 +size 1711309864 diff --git a/model-00057-of-00082.safetensors b/model-00057-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bf9549e2ee7e1f17512add5ac43df47d2c4c2a0d --- /dev/null +++ b/model-00057-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9da3707728058e91f76d1bef75611068516e42c67c7e09766f93b8b6e55a1610 +size 1711309864 diff --git a/model-00058-of-00082.safetensors b/model-00058-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8333a922bb61589501232495ea862d91034879c7 --- /dev/null +++ b/model-00058-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cee9f0b4e9549186972faa50d87a9ad2b033eace93dbe2c8538978e6958e34f5 +size 1711309864 diff --git a/model-00059-of-00082.safetensors b/model-00059-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ef3f19c37bcc5f2a33105a0abbe8b5e1fee96d5e --- /dev/null +++ b/model-00059-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4f3fc9c339f8e8ba36f7765f67a513e8eb84c9465736a1bff2fb70c854ba558 +size 1711309864 diff --git a/model-00060-of-00082.safetensors b/model-00060-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9494533f70189d5df7b1cd88026839744b755ce2 --- /dev/null +++ b/model-00060-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f683e9174c52c9059fc3210effe4110083fe81a8a7e97fba3f15f2cc747e121 +size 1711309864 diff --git a/model-00061-of-00082.safetensors b/model-00061-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fe2dafbd88883c3dc53fead6e1210fe4704b67a7 --- /dev/null +++ b/model-00061-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03b2a97689bd7af3ed0a7248428f4cda8046a37b82b44e51a80848aadb94f78d +size 1711309864 diff --git a/model-00062-of-00082.safetensors b/model-00062-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..713a89cc24dda00571f7dfd61875497b5c6e48fc --- /dev/null +++ b/model-00062-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdc3e70b4f33332241886001f346b23c323b9605e42a416765c4bcbdcea6bdad +size 1711309864 diff --git a/model-00063-of-00082.safetensors b/model-00063-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2c3566d8a83bc99ffd9e071b4845810b6205c265 --- /dev/null +++ b/model-00063-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb51b302c79b8d2fb3fd6c8e4e26405ec50820cf2c421edf48ae91e2452b7418 +size 1711309864 diff --git a/model-00064-of-00082.safetensors b/model-00064-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bde41d9dd4bb63c0c4e426dd12307df759af311a --- /dev/null +++ b/model-00064-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1cd135a4bb029d836b0e1710818d195135688229c693fdb82ae959d611098cd3 +size 1711309864 diff --git a/model-00065-of-00082.safetensors b/model-00065-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b5a2f6bfbe4655ea4b1fb984d4f6cb7946cfe7ed --- /dev/null +++ b/model-00065-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e0e4b6fb72ea94c3f27d07146ef05274d9f46763f5a3337428c4bd95a722af9 +size 1711309864 diff --git a/model-00066-of-00082.safetensors b/model-00066-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..da912b1658d5adf32f3fea169b3f052877d645da --- /dev/null +++ b/model-00066-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5377fc7765185bd6150126c5866cf7a77da3f094380724f26608acb282c7b0b +size 1711309864 diff --git a/model-00067-of-00082.safetensors b/model-00067-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..abd38357ec6286b8d822da0036267bb9ad11a276 --- /dev/null +++ b/model-00067-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ff96082624c29d9870209faaec71dbf8450fe85635d41dfc08ca153870b34e7 +size 1711309864 diff --git a/model-00068-of-00082.safetensors b/model-00068-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..59a87ba7f40674c16ec567e22674cfb63492292f --- /dev/null +++ b/model-00068-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9effa5fbd3e3d5037651ff0f8fc93d1d3253090d2cf2f3874f1ac4a6b50da39b +size 1711309864 diff --git a/model-00069-of-00082.safetensors b/model-00069-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e0751b860db397738109114ca4c4424948e7d0df --- /dev/null +++ b/model-00069-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3531c4521b6a902aabf00c0e51dd11c64a8eed844de1c9a0713be51d3543945f +size 1711309864 diff --git a/model-00070-of-00082.safetensors b/model-00070-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6dd654c36d86cd6a9ba5c7ac35ba526be1e77abb --- /dev/null +++ b/model-00070-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8494c71a603abba80d51cbaa86fc8e1428343432cf6a064686e5fb7d96627dec +size 1711309864 diff --git a/model-00071-of-00082.safetensors b/model-00071-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d7878f424febea8c4ea3ea7510e7edf349508a49 --- /dev/null +++ b/model-00071-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7b36fdfbeb0de7402554c457785547f5b8e0b9b559cc15166b9963b0888f66c +size 1711309864 diff --git a/model-00072-of-00082.safetensors b/model-00072-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..09b58c28fb4acdeaad0419174541ac2d35cbd7fb --- /dev/null +++ b/model-00072-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abaaeb19f3b4f2a167e05f3a85c324dd9b2d25453c08837efcc6011c0dd18ef1 +size 1711309864 diff --git a/model-00073-of-00082.safetensors b/model-00073-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..db28bbd6d819ab7efbd8befbf82aaab1f6a596d2 --- /dev/null +++ b/model-00073-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9a845e44f3f7771ea32c8f3af0bbaf604e931e08a535c0a5bcd641749fcf0b2 +size 1711309864 diff --git a/model-00074-of-00082.safetensors b/model-00074-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..62dc880056f61e1797e655a300e3614df364a812 --- /dev/null +++ b/model-00074-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cbfc929bdbae128e1b0ae65a84aadf473c1417c9a5fbce29881ff0cdf3bfce3e +size 1711309864 diff --git a/model-00075-of-00082.safetensors b/model-00075-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e98db169e5ebd93080c24dcb09ffd37307d18853 --- /dev/null +++ b/model-00075-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:acdf35b3255a773266fdf925adad9c750c31d064399fb0f9fea96d13ca9fcbf8 +size 1711309864 diff --git a/model-00076-of-00082.safetensors b/model-00076-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..24137400b43b58b66b1284f803f27cce7bd5e1f7 --- /dev/null +++ b/model-00076-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0ad6be7f3dead53765753eeed94d4f54540b1aada4e76030c51516bdb9ab171 +size 1711309864 diff --git a/model-00077-of-00082.safetensors b/model-00077-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8fca2c10b1bf13eeaf55e5438ca9d3e393f7e1df --- /dev/null +++ b/model-00077-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c220881ec423b5b26bcff891d3b77184e4d98f6a5537e0976669d8962ab96c28 +size 1711309864 diff --git a/model-00078-of-00082.safetensors b/model-00078-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b0cc7218560fc0b20adfa5381244ee1f9c2e32b6 --- /dev/null +++ b/model-00078-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94ebc626b8973725ca697f5f797a80428fc8494568df4fd3a578e9571bc60d38 +size 1711309864 diff --git a/model-00079-of-00082.safetensors b/model-00079-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b9149f5ab05acbc759975ce7922f71df3b5eecf8 --- /dev/null +++ b/model-00079-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dfced4ee9524a9b4cc0cc81a47ffcdf7fed61446128cef4d55da189edca9a1e4 +size 1711309864 diff --git a/model-00080-of-00082.safetensors b/model-00080-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6874b027c57e478ded4728da5d175bc05090af69 --- /dev/null +++ b/model-00080-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e09579d5877f752c08e7677201d35b0ca0c2f4764124630adac921add30af911 +size 1711309864 diff --git a/model-00081-of-00082.safetensors b/model-00081-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..17845407028116d6416eb792a564a6accf956906 --- /dev/null +++ b/model-00081-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08e01d4ef1ce280ab6d4aab51a2d03332e459b15c0cf1c9eba33d44ab40d7443 +size 1409335984 diff --git a/model-00082-of-00082.safetensors b/model-00082-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..47bbd5d406e55fa86114c9704d4223676de38034 --- /dev/null +++ b/model-00082-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8bef21c543cda4b20d25fced088d0092bf76815468b73b9d95d16190c2dfb5a +size 2101346432 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..ae9e1c3033b9c33d45108c1ef1a946dbf34669b9 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,730 @@ +{ + "metadata": { + "total_size": 141107412992 + }, + "weight_map": { + "lm_head.weight": "model-00082-of-00082.safetensors", + "model.embed_tokens.weight": "model-00001-of-00082.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00082.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.1.input_layernorm.weight": "model-00003-of-00082.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00082.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.10.input_layernorm.weight": "model-00012-of-00082.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00012-of-00082.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.11.input_layernorm.weight": "model-00013-of-00082.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00013-of-00082.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.12.input_layernorm.weight": "model-00014-of-00082.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00014-of-00082.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.13.input_layernorm.weight": "model-00015-of-00082.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00015-of-00082.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.14.input_layernorm.weight": "model-00016-of-00082.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00016-of-00082.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.15.input_layernorm.weight": "model-00017-of-00082.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00017-of-00082.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.16.input_layernorm.weight": "model-00018-of-00082.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00018-of-00082.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.17.input_layernorm.weight": "model-00019-of-00082.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00019-of-00082.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.18.input_layernorm.weight": "model-00020-of-00082.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00020-of-00082.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.19.input_layernorm.weight": "model-00021-of-00082.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00021-of-00082.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.2.input_layernorm.weight": "model-00004-of-00082.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00004-of-00082.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.20.input_layernorm.weight": "model-00022-of-00082.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00022-of-00082.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.21.input_layernorm.weight": "model-00023-of-00082.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00023-of-00082.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.22.input_layernorm.weight": "model-00024-of-00082.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00024-of-00082.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.23.input_layernorm.weight": "model-00025-of-00082.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00025-of-00082.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.24.input_layernorm.weight": "model-00026-of-00082.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00026-of-00082.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.25.input_layernorm.weight": "model-00027-of-00082.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00027-of-00082.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.26.input_layernorm.weight": "model-00028-of-00082.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00028-of-00082.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.27.input_layernorm.weight": "model-00029-of-00082.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00029-of-00082.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.28.input_layernorm.weight": "model-00030-of-00082.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00030-of-00082.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.29.input_layernorm.weight": "model-00031-of-00082.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00031-of-00082.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.3.input_layernorm.weight": "model-00005-of-00082.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00005-of-00082.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.30.input_layernorm.weight": "model-00032-of-00082.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00032-of-00082.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.31.input_layernorm.weight": "model-00033-of-00082.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00033-of-00082.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.32.input_layernorm.weight": "model-00034-of-00082.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00034-of-00082.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.33.input_layernorm.weight": "model-00035-of-00082.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00035-of-00082.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.34.input_layernorm.weight": "model-00036-of-00082.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00036-of-00082.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.35.input_layernorm.weight": "model-00037-of-00082.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00037-of-00082.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.36.input_layernorm.weight": "model-00038-of-00082.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00038-of-00082.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.37.input_layernorm.weight": "model-00039-of-00082.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00039-of-00082.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.38.input_layernorm.weight": "model-00040-of-00082.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00040-of-00082.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.39.input_layernorm.weight": "model-00041-of-00082.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00041-of-00082.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.4.input_layernorm.weight": "model-00006-of-00082.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00006-of-00082.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.40.input_layernorm.weight": "model-00042-of-00082.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00042-of-00082.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.41.input_layernorm.weight": "model-00043-of-00082.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00043-of-00082.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.42.input_layernorm.weight": "model-00044-of-00082.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00044-of-00082.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.43.input_layernorm.weight": "model-00045-of-00082.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00045-of-00082.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.44.input_layernorm.weight": "model-00046-of-00082.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00046-of-00082.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.45.input_layernorm.weight": "model-00047-of-00082.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00047-of-00082.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.46.input_layernorm.weight": "model-00048-of-00082.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00048-of-00082.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.47.input_layernorm.weight": "model-00049-of-00082.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00049-of-00082.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.48.input_layernorm.weight": "model-00050-of-00082.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00050-of-00082.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.49.input_layernorm.weight": "model-00051-of-00082.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00051-of-00082.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.5.input_layernorm.weight": "model-00007-of-00082.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00007-of-00082.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.50.input_layernorm.weight": "model-00052-of-00082.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00052-of-00082.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.51.input_layernorm.weight": "model-00053-of-00082.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00053-of-00082.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.52.input_layernorm.weight": "model-00054-of-00082.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00054-of-00082.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.53.input_layernorm.weight": "model-00055-of-00082.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00055-of-00082.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.54.input_layernorm.weight": "model-00056-of-00082.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00056-of-00082.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.55.input_layernorm.weight": "model-00057-of-00082.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00057-of-00082.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.56.input_layernorm.weight": "model-00058-of-00082.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00058-of-00082.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.57.input_layernorm.weight": "model-00059-of-00082.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00059-of-00082.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.58.input_layernorm.weight": "model-00060-of-00082.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00060-of-00082.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.59.input_layernorm.weight": "model-00061-of-00082.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00061-of-00082.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.6.input_layernorm.weight": "model-00008-of-00082.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00008-of-00082.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.60.input_layernorm.weight": "model-00062-of-00082.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00062-of-00082.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.61.input_layernorm.weight": "model-00063-of-00082.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00063-of-00082.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.62.input_layernorm.weight": "model-00064-of-00082.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00064-of-00082.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.63.input_layernorm.weight": "model-00065-of-00082.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00065-of-00082.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.64.input_layernorm.weight": "model-00066-of-00082.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00066-of-00082.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.65.input_layernorm.weight": "model-00067-of-00082.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00067-of-00082.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.66.input_layernorm.weight": "model-00068-of-00082.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00068-of-00082.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.67.input_layernorm.weight": "model-00069-of-00082.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00069-of-00082.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.68.input_layernorm.weight": "model-00070-of-00082.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00070-of-00082.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.69.input_layernorm.weight": "model-00071-of-00082.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00071-of-00082.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.7.input_layernorm.weight": "model-00009-of-00082.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00009-of-00082.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.70.input_layernorm.weight": "model-00072-of-00082.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00072-of-00082.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.71.input_layernorm.weight": "model-00073-of-00082.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00073-of-00082.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.72.input_layernorm.weight": "model-00074-of-00082.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00074-of-00082.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.73.input_layernorm.weight": "model-00075-of-00082.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00075-of-00082.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.74.input_layernorm.weight": "model-00076-of-00082.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00076-of-00082.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.75.input_layernorm.weight": "model-00077-of-00082.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00077-of-00082.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.76.input_layernorm.weight": "model-00078-of-00082.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00078-of-00082.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.77.input_layernorm.weight": "model-00079-of-00082.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00079-of-00082.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.78.input_layernorm.weight": "model-00080-of-00082.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00080-of-00082.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.79.input_layernorm.weight": "model-00081-of-00082.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00081-of-00082.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00081-of-00082.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00081-of-00082.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00081-of-00082.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.8.input_layernorm.weight": "model-00010-of-00082.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00010-of-00082.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.9.input_layernorm.weight": "model-00011-of-00082.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00011-of-00082.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00010-of-00082.safetensors", + "model.norm.weight": "model-00081-of-00082.safetensors" + } +}