diff --git a/README.md b/README.md index d7035fe2cabc33aa0f87a8864e3e57f45f6d1c1c..05e324c13c4fcf9c82634efd61c3839e78cd07a4 100644 --- a/README.md +++ b/README.md @@ -22,22 +22,19 @@ Shining Valiant is a chat model built on the Llama 2 architecture, finetuned on - Uses the llama-2-70b-chat model, with safetensors - Finetuned on multiple runs across private and public data - Data focused on knowledge, enthusiasm, and structured reasoning + - **Our new release features greatly expanded personality capability**, bringing a more immersive chat experience ## Version -The current version is **1.3!** +The current version is **1.4!** We've greatly expanded our personality dataset and fixed some bugs to deliver our strongest real-chat experience so far. -Version 1.4 is preparing for release! We've expanded our personality dataset and fixed some bugs to deliver our strongest real-chat experience so far. - -(We're also exploring new models and architectures!) +(We're also exploring **new models and architectures**, to deliver helpful open source capabilities for users and creators!) Previous versions remain available in the repository. New models will be released for everyone once our team's training and validation process is complete. ## Evaluation -| Model | Avg | ARC | HS | MMLU | TQA | WG | GSM | -|-----------------------|--------|-------|-------|--------|-------|-------|-------| -| **Shining Valiant 1.3** | 73.78 | 71.33 | 90.96 | 71.21 | 70.29 | 84.21 | 54.66 | +Version 1.4 is awaiting results from the Open LLM leaderboard. ## Prompting Guide Shining Valiant uses the same prompt format as Llama 2 Chat - feel free to use your existing prompts and scripts! @@ -50,13 +47,11 @@ A few examples of different formats: 3. [INST] << SYS >>You are an intelligent, helpful AI assistant.<< /SYS >>Deep dive about a country with interesting history: [/INST] ## The Model -Shining Valiant is built on top of Sunset Boulevard, which uses Llama 2's 70b parameter architecture and features upgraded general capability. - -From there, we've created Shining Valiant through multiple finetuning runs on different compositions of our private dataset. +Shining Valiant is built on top of Spell Blade, which uses Llama 2's 70b parameter architecture and features upgraded general and chat capability. -Our private data focuses primarily on applying Shining Valiant's personality: she's friendly, enthusiastic, insightful, knowledgeable, and loves to learn! +Our private data focuses primarily on applying Shining Valiant's personality: she's friendly, enthusiastic, insightful, knowledgeable, and loves to learn! -We are actively working on expanding and improving the Shining Valiant dataset for use in future releases of this model and others. +With this release, the personality component of our Shining Valiant dataset has been greatly improved. We're excited to use it in future releases of this model and others. diff --git a/config.json b/config.json index 61dca1b5c1d8f5dbcc61a337f70495ed1721466c..3dee0b76ce21c41e9a9515cd53ff5eec6ef5d0d7 100644 --- a/config.json +++ b/config.json @@ -4,6 +4,7 @@ "LlamaForCausalLM" ], "attention_bias": false, + "attention_dropout": 0.0, "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", @@ -22,7 +23,7 @@ "rope_theta": 10000.0, "tie_word_embeddings": false, "torch_dtype": "float32", - "transformers_version": "4.35.0", + "transformers_version": "4.36.2", "use_cache": false, "vocab_size": 32000 } diff --git a/generation_config.json b/generation_config.json index 972c9af96157ad4fbfc642babe102daf22e9d4e0..5786877b25c8295c74590b8a5aace96539c44dd2 100644 --- a/generation_config.json +++ b/generation_config.json @@ -3,6 +3,6 @@ "bos_token_id": 1, "eos_token_id": 2, "pad_token_id": 0, - "transformers_version": "4.35.0", + "transformers_version": "4.36.2", "use_cache": false } diff --git a/model-00001-of-00061.safetensors b/model-00001-of-00061.safetensors index 27bf0febefbc748eae84eb33fb4bb2833e397a9c..c959630fb436e73e6d71f36c4d911496b2a0817e 100644 --- a/model-00001-of-00061.safetensors +++ b/model-00001-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:3dc4ed54e736c5d13b2ced1c4c5d96d36854f823c2a80316c08674ac4dc12905 -size 4829678080 +oid sha256:fadced9ed406cba213082d7af2d6e8a2b0fb74bf8c2aec923a6d00c078e15edd +size 4806739440 diff --git a/model-00002-of-00061.safetensors b/model-00002-of-00061.safetensors index 7aa534e8420f9a3b35d171eb362a1ff2739935cc..732042d99dee93cd8e1e101f21f34b842b321934 100644 --- a/model-00002-of-00061.safetensors +++ b/model-00002-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:46b3af357e97b16ff83a04c91b76957b58b7b4c398017c3fe23b7852a5fa973c -size 4642047760 +oid sha256:d71e9ba35d69f49616f91e6b257160b0504e1d71709cfa8aeb993258b3d32353 +size 4630578440 diff --git a/model-00003-of-00061.safetensors b/model-00003-of-00061.safetensors index 7dfddaadd184815e24dc841b83d7583f68b57a84..97c4f99b6eeede9531249358fb337f9cdb2f065f 100644 --- a/model-00003-of-00061.safetensors +++ b/model-00003-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:7c30362d287354bbfb8a03a1ac1f00f01d113388cb2ad5654dd1b5861820450b -size 4373612184 +oid sha256:8fbdcd01f86fed6f95527e3235231b7ea5f61b54d0a4fa654eb5dc1331f6b23b +size 4362142864 diff --git a/model-00004-of-00061.safetensors b/model-00004-of-00061.safetensors index 0175a1f997d883a721cb2d25905b6d2616c2d6c4..7d93a28816d69fdb50ca4f9db6a12ae7ec9393ea 100644 --- a/model-00004-of-00061.safetensors +++ b/model-00004-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:7f6ec8050a5cec810bf98716a50c48587ea240f1a8ae1037587c4a069c0c5346 -size 4989127504 +oid sha256:f30c7fc97d6f6e93873a744bddb05507d8f473ff02ed345dcf91491f1172aac3 +size 4966188864 diff --git a/model-00005-of-00061.safetensors b/model-00005-of-00061.safetensors index b2754ce80e0506532d16e09ea9d7b4e7386f2b7b..12bc22ea8d807dc8271cfd46f4d6f8f5aaac912e 100644 --- a/model-00005-of-00061.safetensors +++ b/model-00005-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:f798e0c23342f96a09bbac844334d8542dece23627168c01819aefd8db1dbae4 -size 4373612184 +oid sha256:84d2299a7ab80828e5515b27a965274246b4f25fceeae0cfa5b9425fc671e67a +size 4362142864 diff --git a/model-00006-of-00061.safetensors b/model-00006-of-00061.safetensors index a4c97584f42de111848e03d838bec33a3e1b435a..c9e420d2c5a934a506523fb34c65250adb2de31b 100644 --- a/model-00006-of-00061.safetensors +++ b/model-00006-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:77d2a7b76a95c7fda85e63b262e3d00cb0178306aaff700cea4603dd55f7977a -size 4373612184 +oid sha256:57e6bb3e7cd211ce5e38c6ce444c83a0cbd195c48c7c0fdd9fc9a5aaf037d154 +size 4362142864 diff --git a/model-00007-of-00061.safetensors b/model-00007-of-00061.safetensors index 53011d63dc16f956baeccd8584538d60da4073b2..923f0402508f15c347132c309c9bf268dd43a958 100644 --- a/model-00007-of-00061.safetensors +++ b/model-00007-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:f2b621b2600196832c8e85a4bf68d111d03159b0c08e9731dfd4f45df4ab8c10 -size 4989127504 +oid sha256:2b626c0de24768faa49651ea3f86198ab5b18d860e5998cc5a2162b791d225b7 +size 4966188864 diff --git a/model-00008-of-00061.safetensors b/model-00008-of-00061.safetensors index bc08a8827586155fce7c209332e6e775d2322f2c..e90a2c833d87a064297bee8db7d9a027189d1c04 100644 --- a/model-00008-of-00061.safetensors +++ b/model-00008-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:bc3bce3873e52561fdd0d3d3594d9be7e02256d60ace42be717aba51b896e096 -size 4373612200 +oid sha256:1c0655747bb6df21938ed8ddc9f0989c0246b16a628a065aa20b7465f6516f5b +size 4362142880 diff --git a/model-00009-of-00061.safetensors b/model-00009-of-00061.safetensors index 397b0ab4052859baa6c4e82fd58799bb74d5ba2c..f177f8645d8b60868e70ec3842da8ef879f7c7cb 100644 --- a/model-00009-of-00061.safetensors +++ b/model-00009-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:d1f7c5ef4e3f7979037b25a5c6aa8e0824ec80793d7be37a4586bbab81036c40 -size 4373612200 +oid sha256:4f69090e812eddfc4a05fab5fa08e03dc0a00a401dc5f859351bf0d16db4c017 +size 4362142872 diff --git a/model-00010-of-00061.safetensors b/model-00010-of-00061.safetensors index dd68cd01cc2baefb048923b5566ef31e991d17b0..a801ed410190193c7f2039e1236a283d16085844 100644 --- a/model-00010-of-00061.safetensors +++ b/model-00010-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:eaf35b1f54fea6f42dad09ae0c359a7cd8c3a6e937ed0677536f5cde9a04f322 -size 4989127528 +oid sha256:8135da12b9270d2a2b9fef41f3f7a2cb68bdcf29fb784d70a9e0571b9208decc +size 4966188880 diff --git a/model-00011-of-00061.safetensors b/model-00011-of-00061.safetensors index c405a1fbd83e8d5482139815efea04fa3fc8738e..ca8f5c78463c5ba4613110180f2bbb6467816502 100644 --- a/model-00011-of-00061.safetensors +++ b/model-00011-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:5010617f685d5c77ac291540a3e746b585264f54dfd84e715e572eed2dbd76d6 -size 4373612200 +oid sha256:86b4c94e4d81015e884baf82eb688cd5f70a9d2b0b782c0fcbb4aff7ffd6895c +size 4362142872 diff --git a/model-00012-of-00061.safetensors b/model-00012-of-00061.safetensors index ca11d39d6614082c63db03d30523658cc42b0e37..6e52f0403781585ebed7c275e7debd308e1f9038 100644 --- a/model-00012-of-00061.safetensors +++ b/model-00012-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:c5247cb589001678aa6fcabe321eaff2b99ae98c56e940296b8ee900d7dbc0a7 -size 4373612200 +oid sha256:772c447cf406df2b55c8dd06b68e90b958cb5d1e04013574295e028f94666b2b +size 4362142872 diff --git a/model-00013-of-00061.safetensors b/model-00013-of-00061.safetensors index 552ea748c20d97e8f67389c34f7e1e2b852d4b1a..03a32a6ce55a2d00b889b7592a0019acb4bc99a3 100644 --- a/model-00013-of-00061.safetensors +++ b/model-00013-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:53b7f85259c4576142f2a8700fbc12809758340b59f16a649d8ace3093fc3d9f -size 4989127528 +oid sha256:5b8889cd0c49ed15074a8fac2f4fb29109c7a30c2c374d2d37c3f7378c16af0e +size 4966188880 diff --git a/model-00014-of-00061.safetensors b/model-00014-of-00061.safetensors index 421305f621134a2e289c4c4afe6c955e709711af..fdfab39ef0d19728131db33a17ad7d9b20850627 100644 --- a/model-00014-of-00061.safetensors +++ b/model-00014-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:94d96ea5b1bd5ea70fefdb276b5071e2822dc3b56db3211ec8397d701d8230b4 -size 4373612200 +oid sha256:6a2d51a87a86e76d213b57a1bf6e11269eb5d09498d2d8a52acd499ec2e5316d +size 4362142872 diff --git a/model-00015-of-00061.safetensors b/model-00015-of-00061.safetensors index 2edcbe9641c6b951410bb73e4567bea29e4a9c4c..a2b339e93e24b8a05f0a8cc22e0ba1fcf8c188f7 100644 --- a/model-00015-of-00061.safetensors +++ b/model-00015-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:894d89e5b9b97d924101027848e58c0ac9f79b4436eff5dd9de45d37d4152cdc -size 4373612200 +oid sha256:af0e0010590fa927b51b0b78662c74e8753633e1ef2f7e496de8bbf87b2ff94e +size 4362142872 diff --git a/model-00016-of-00061.safetensors b/model-00016-of-00061.safetensors index 9fdcf4be9ff1a89008717d5e992cb6d7b16cc9b5..819fda72b3ff973433d1ef7717b5fa5c44d34f0a 100644 --- a/model-00016-of-00061.safetensors +++ b/model-00016-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:46d0b2a79ab34a91fa8b8f279ff88cf24d84a96fce7fe8cd98f8357701205d5f -size 4989127528 +oid sha256:49d1e9e62b881b2c5ddb41255ef1976aeeb8fa24ab84c0195edc6f4e76412e5c +size 4966188880 diff --git a/model-00017-of-00061.safetensors b/model-00017-of-00061.safetensors index fe80876611320bc1f5306d4e62ac3a6cbc304d9c..49be5d070392b9708e504e70194a407140342fb1 100644 --- a/model-00017-of-00061.safetensors +++ b/model-00017-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:c3bf163933cbe771e0836a82aa38473da7a1803fb059e4cb04e69504d29862ad -size 4373612200 +oid sha256:fa629921ef4fd76258e1bf0c4998f9790dec7b6f7b1b14b561cee769be12a9d3 +size 4362142872 diff --git a/model-00018-of-00061.safetensors b/model-00018-of-00061.safetensors index d0bcfae60eb0c093eef5e9d36ab11cecf10ba536..7fe33a41288e870c7ee0c682b84338ed7ba85a82 100644 --- a/model-00018-of-00061.safetensors +++ b/model-00018-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:d9d9390a0c9dd22dc16b1e1ed30a0f8595d8f12bf52a498b2b29a6c62b6467e1 -size 4373612200 +oid sha256:b639228d922c57aa9c032ca6567ebb18bf3f919d1d9b242d466e748839f683f4 +size 4362142872 diff --git a/model-00019-of-00061.safetensors b/model-00019-of-00061.safetensors index 914171cbdc0e4950de9c439c82514901ed420742..16ae478b249efb25b9c07a73d0c1fa7352377350 100644 --- a/model-00019-of-00061.safetensors +++ b/model-00019-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:b4529d07c0e6dcdf8a77a2627d117ee922abeee618069569451849ecc7aa339d -size 4989127528 +oid sha256:3d88d8cbb0974a6503eb9a039f2ca8fd78b5c345e275b05e1919f3092ee598fc +size 4966188880 diff --git a/model-00020-of-00061.safetensors b/model-00020-of-00061.safetensors index 7603cbe0d5149ea3bc3fa144e4e26a582963e64a..0c662c1b80558b2af6754f5e03ecbbdca8b9bd82 100644 --- a/model-00020-of-00061.safetensors +++ b/model-00020-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:3bd585996c4d5af5587b7a807bdf2fd2170981f041db89e8b2923469ec940d9d -size 4373612200 +oid sha256:3d360eb276ac5f13e5178a26615bb83bc12b9414e20993b18faa55d7bb8e45ab +size 4362142872 diff --git a/model-00021-of-00061.safetensors b/model-00021-of-00061.safetensors index 39cfd69e53624e181a11ffd465c4f37d1b1e52fa..04e9200cc7ed363661155969f9cff7586c7e0f5d 100644 --- a/model-00021-of-00061.safetensors +++ b/model-00021-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:30854cb22ca8777646ba273e0591f4cf1690e4078c9298077c50cf0e2dde583c -size 4373612200 +oid sha256:ec018b78d09ecc777ac4b82c90ee00a4bf8f3f480c08ae07956a5247cb4d17b8 +size 4362142872 diff --git a/model-00022-of-00061.safetensors b/model-00022-of-00061.safetensors index 88108163b201cd44db1ee274b0b869bfdfcf4020..332b3ba44e4c22e25fd71ffd01aed41958cf9d7a 100644 --- a/model-00022-of-00061.safetensors +++ b/model-00022-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:651b0e0d6d0726aa720f4ce68c67bd4b47a74b22d370590510ddb6348c69ab5a -size 4989127528 +oid sha256:38815ffe4fbf483f818a087b337ed9870511cb578b739cdfad7677277516865e +size 4966188880 diff --git a/model-00023-of-00061.safetensors b/model-00023-of-00061.safetensors index 995cb0e735007a720cdcadef9f66e114a1fe373b..5649b43ee0e69a6a98abd90d5e5fd3d1a9b442bc 100644 --- a/model-00023-of-00061.safetensors +++ b/model-00023-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:5e4fd8b76d520900d420b1dd619c17d6dbca9db12c3f9dfc543811c3613ae08b -size 4373612200 +oid sha256:0816e06830527980788d2429c1ce8a48ae4b7b98d25dd3f19581ce4b261687fd +size 4362142872 diff --git a/model-00024-of-00061.safetensors b/model-00024-of-00061.safetensors index bb64f46de9ac6e92f7eeca3ade47f95a0c4b7d54..c5bdadc6b324c5489911493812ed712b29e7a5d0 100644 --- a/model-00024-of-00061.safetensors +++ b/model-00024-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:bf67fc00883bbe333f29a5fe93a481f9ac8c548705a0b96b79f1b2072af2e391 -size 4373612200 +oid sha256:c28a941e302b583181efe48462bd98954df63f889431568d47f18a33137cfa41 +size 4362142872 diff --git a/model-00025-of-00061.safetensors b/model-00025-of-00061.safetensors index b48e6cb3f1a8d109d7088d052c89cbda7347ed05..37f54d3b5f06da19641aa972239ed15c82d56896 100644 --- a/model-00025-of-00061.safetensors +++ b/model-00025-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:383ad969a694badc469b5d313ccffb384c25e5a9280a69772aca961968ced11a -size 4989127528 +oid sha256:83c6fcb6318508758cc958afc13a07c7032771d52a5e71b3d2858d0e5b01f70b +size 4966188880 diff --git a/model-00026-of-00061.safetensors b/model-00026-of-00061.safetensors index c6e781b9a3740c89103c96f05fcc0659c0f00574..318cdc9c0192844702d3a15f0882b6528f64dd1d 100644 --- a/model-00026-of-00061.safetensors +++ b/model-00026-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:a305dad26ba99db15c0d788f0703fb8d8652289b9a1d83f906eabc04fa2a1965 -size 4373612200 +oid sha256:59b32cf8cb9910fe40767b59304e6c699145190cf01e8c67ec84e9899c28daf4 +size 4362142872 diff --git a/model-00027-of-00061.safetensors b/model-00027-of-00061.safetensors index ebc325bc6862a0d89542b3c198724472042049a3..9ff1b71fbc0b45385e25edb7d459ae80545d17f4 100644 --- a/model-00027-of-00061.safetensors +++ b/model-00027-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:a78b51174496d7caf102675247fd4cc7d25c6b7c08bb10eea6e913b812ec42fe -size 4373612200 +oid sha256:806a61770de81c984aa132a819dc577390c589b3773c170f99cab88ef836d040 +size 4362142872 diff --git a/model-00028-of-00061.safetensors b/model-00028-of-00061.safetensors index 75db7d3c699bb2bbe5a7f6873cde91dae8cf35fe..172a3a900ef23f6b7ae4137f7d3d293d63b5a124 100644 --- a/model-00028-of-00061.safetensors +++ b/model-00028-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:80c35fafdad91444fd97f5b7c18b0eb704bf1a8d1e60508ec8fa4d365ad7202d -size 4989127528 +oid sha256:7e875e621a124a6f3b6744466131b782f48b9541005d425ee366dbee28bf564a +size 4966188880 diff --git a/model-00029-of-00061.safetensors b/model-00029-of-00061.safetensors index 42764f0af40b011edebf42ef112993c6d163252c..0dfda31ded638591be7482735cdffa6fed989ed6 100644 --- a/model-00029-of-00061.safetensors +++ b/model-00029-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:8c0633b845bccf0924cdf6f723b42b89cb3a48eda645e1d57452300037911524 -size 4373612200 +oid sha256:d356bdd0447b67f6e9f699d1b4dc1c778c9382a6650668d0d6e884c7e514b4fa +size 4362142872 diff --git a/model-00030-of-00061.safetensors b/model-00030-of-00061.safetensors index 2f433bf7682682fe8f89fe5fb5686149af7f9c8c..7f40f92d5efafb673d0da43a91bad4ab2df8a219 100644 --- a/model-00030-of-00061.safetensors +++ b/model-00030-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:e946e96b50482a20eb7df6d2584a7bc6f23471e4d7890111fd01b92812d1aa1f -size 4373612200 +oid sha256:cdb537ac83889a26d44c8e74077f3340ae08e370dbd43d08bd0df22182e4eeb9 +size 4362142872 diff --git a/model-00031-of-00061.safetensors b/model-00031-of-00061.safetensors index 7e3f69b4d66c67c2080fb0af2606e01b66a8a1fb..5ae3853b7119a3fcdd0b903a71bb37307f5b6606 100644 --- a/model-00031-of-00061.safetensors +++ b/model-00031-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:5543aa4273bf925573330575ec83dd17483a6c5bbf5864cc9429ca6e61729fa0 -size 4989127528 +oid sha256:dac8057a579dabe02fb1e2949af46bac00ce13338b0828e0972de008b3e8beda +size 4966188880 diff --git a/model-00032-of-00061.safetensors b/model-00032-of-00061.safetensors index 325a1b15b4e5d77f7901aa2f2f186c681c4f84b3..cac63a0d6d2cd4d0e2a01aa6d4f926c88f90eca3 100644 --- a/model-00032-of-00061.safetensors +++ b/model-00032-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:c2a1adee9e1983d93ddd08d8fc5b0d2cc2a701f441afe7a018f88414d6b181e7 -size 4373612200 +oid sha256:23ded11e0d3af59cd60617d690e94bba4aafdae9d5ee9c89b0bc1cefafff7989 +size 4362142872 diff --git a/model-00033-of-00061.safetensors b/model-00033-of-00061.safetensors index 2e704a79a1558ccb2f645de01d0d6a3333783986..83d61ecc9a184f2be6bb3438f13d78eee46fcd0c 100644 --- a/model-00033-of-00061.safetensors +++ b/model-00033-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:9432a07eea4f92853667c83f60f662b36f6e0fd1adfd92e9e6c7d52f87648d52 -size 4373612200 +oid sha256:02778744f534af84e670537891075189c185664a301eb17d36c84aa47c099f7f +size 4362142872 diff --git a/model-00034-of-00061.safetensors b/model-00034-of-00061.safetensors index f5ab6125ca9b805ce2d4e7e6b91ed10aa73a67d9..279032919a3169b901a93825e136f9a3af1f66f9 100644 --- a/model-00034-of-00061.safetensors +++ b/model-00034-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:0b9cc6274a106b89b659f7952497e3958354650e640dc8051376abbf3e3e293b -size 4989127528 +oid sha256:4ffaf5278c9e94b2274b011823064147e670f97dc6a5ec5252029149ce2dddf2 +size 4966188880 diff --git a/model-00035-of-00061.safetensors b/model-00035-of-00061.safetensors index 8545b2f5e09122e88b5a3e2f162066b4d29f5e48..772f5c5355644c9d6bce5460326829e57fead7e0 100644 --- a/model-00035-of-00061.safetensors +++ b/model-00035-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:4f17095616c39cc562f354d348f7a2fa9ca851213831e1b78c0f958ec4e4ab0c -size 4373612200 +oid sha256:ac765112af36774ae928c2ecd3c2e2aaebaf055616d1f781ec05c53a02c59be2 +size 4362142872 diff --git a/model-00036-of-00061.safetensors b/model-00036-of-00061.safetensors index eeeb1abfc8c8cd55f8579be0e72a4ed8ed1a705a..5efc1bfc7326935ac8245f986f029fb7246c65cb 100644 --- a/model-00036-of-00061.safetensors +++ b/model-00036-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:7a006ebb9c9a48e49d021626ff12eb1a81bd7eb41298d6e2633372624e6764ae -size 4373612200 +oid sha256:e3f9f6452355bd26a699161ca58d6f1a47ee9217d1014dde36c102d9ee0c5a41 +size 4362142872 diff --git a/model-00037-of-00061.safetensors b/model-00037-of-00061.safetensors index 30c83e554007e7ed1f782c2d6c57e81a7a03769b..bc74fc9053bd488427b891d5511fa0e0f8dc3b62 100644 --- a/model-00037-of-00061.safetensors +++ b/model-00037-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:3f0cba34fb68521ca64535aa1b4926c8dc42def86990d88c5ff9ec2f361ef403 -size 4989127528 +oid sha256:136869e51158690c204cdbf521967de8a47e8ef0a3ee8eaafaffafda04d18b35 +size 4966188880 diff --git a/model-00038-of-00061.safetensors b/model-00038-of-00061.safetensors index 13eca929d487b195a2f3536fea3b765e2d7c1fc7..2e9323fd6eb40aad2b8fc05684ee4b7bf0696bbe 100644 --- a/model-00038-of-00061.safetensors +++ b/model-00038-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:cc3bcfc5feeacb1fd29e55c4e06ff376a41706b1274dfb34fb6ef388ed36da5e -size 4373612200 +oid sha256:ef7d7256e4340154be3dfb35a55695a881848f8cd69417590ec65ae142f7cfe4 +size 4362142872 diff --git a/model-00039-of-00061.safetensors b/model-00039-of-00061.safetensors index daf3fa2854e1d84c1d2b351476dee913012d4406..a5e3cadce3a0096171eba37db134de9138cc9c25 100644 --- a/model-00039-of-00061.safetensors +++ b/model-00039-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:52dd5434217e8956e78be5bb6f8762809a565812675e8c2a72077fccaaeadf33 -size 4373612200 +oid sha256:7632607ac273f894499af7094d45b0e0d23bca7a1314764df942ba0425fa5817 +size 4362142872 diff --git a/model-00040-of-00061.safetensors b/model-00040-of-00061.safetensors index 19be479b5c41da1610355f768b9a3cccd58ecfe0..4ffde7534bff4fd3432ee78e2db205968ad277cf 100644 --- a/model-00040-of-00061.safetensors +++ b/model-00040-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:81a3f6dfe3126fa87cb4911adcf1f78bf879d529a328c53a1cde2e3d4dea241e -size 4989127528 +oid sha256:52154328f66a3b22ab3198377dda9dbdf3ce9f513fbad6cd9fe54de5ad0c6a92 +size 4966188880 diff --git a/model-00041-of-00061.safetensors b/model-00041-of-00061.safetensors index 7cdcee7f1a974f2b89d9a6e738ffe2bb1708c1c7..c073c63869246ad1ca712ea4b51b70eb4cbe6b1d 100644 --- a/model-00041-of-00061.safetensors +++ b/model-00041-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:546a8fab65a3d70e826ef3255138a0fd5d9bff51ef6c146c01fcb0227a4d2153 -size 4373612200 +oid sha256:ae9f95288175eae3c476a94fc10f311e8e134b3b8a18f9f6ba504d019b119873 +size 4362142872 diff --git a/model-00042-of-00061.safetensors b/model-00042-of-00061.safetensors index 261fce9d8e002a80f38a79cc9f9d694495f597d3..0c2118ea8157b463873622039f4a5dc94c1ed507 100644 --- a/model-00042-of-00061.safetensors +++ b/model-00042-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:fdb73cb8ebdef799f5d331b44e6246a68b38908f50d9dc4ef3a99a98f1ff99ea -size 4373612200 +oid sha256:d058d1aafdda3154e8fe41ddaba2248d62b27d8d4c8e8f72ff446942b869d5cc +size 4362142872 diff --git a/model-00043-of-00061.safetensors b/model-00043-of-00061.safetensors index 57ec7bae2c2ded9c51cafdd8cd2a6f1288767638..d6501364a59e0896c4ce86c261e965f0320c4675 100644 --- a/model-00043-of-00061.safetensors +++ b/model-00043-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:5b4944c19cbf313aacfa5b8d70a47cd9c052b34adc15be12cf80cb8838b8ca6e -size 4989127528 +oid sha256:e5369d570a43e13cd5886e10bf9e8d3510da2d92312c30fc82823f4f4202ff26 +size 4966188880 diff --git a/model-00044-of-00061.safetensors b/model-00044-of-00061.safetensors index 2a18c4ea8ca4d17cf55f1b766f87eefe10a6b99f..fb219cc335ad7ca9a69b509d2dc7f1e13f3ddd01 100644 --- a/model-00044-of-00061.safetensors +++ b/model-00044-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:1a7799868fac8c86eea07b50046b46545f1631a47411bfec132a04d646a5a0bd -size 4373612200 +oid sha256:1a0fe48c143a2349891d131155257b4196a43caa357fc79f0b545c99c79d2111 +size 4362142872 diff --git a/model-00045-of-00061.safetensors b/model-00045-of-00061.safetensors index 03683a845d16b1646f73fe1a6668b074d3200612..591fc6104448c2ce844d48543ea3194b93aa8bc3 100644 --- a/model-00045-of-00061.safetensors +++ b/model-00045-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:f90c6d73620da6c3adf1312fb9ee60cf86b87e0f8c6ce87e81b8ba46d3fbda70 -size 4373612200 +oid sha256:565d7442071ab90499e8274faf365033441669a418bf87f8f19385c85f39fe31 +size 4362142872 diff --git a/model-00046-of-00061.safetensors b/model-00046-of-00061.safetensors index f66d1b3bbdaae38410ed8bd9f4fb530c2658c2f5..56f20380a81e6ad6ee6c6ae248e9cf9ca8b8f024 100644 --- a/model-00046-of-00061.safetensors +++ b/model-00046-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:4d9a4db2a1fec6341c31afabd7098d2c609a9d8cc7e461f25ad8642d39d48ba0 -size 4989127528 +oid sha256:d65f504379bb803eca8d3403c42dadb01aa9ce4ed8303530a4beaaf66f048192 +size 4966188880 diff --git a/model-00047-of-00061.safetensors b/model-00047-of-00061.safetensors index 920d342c249aeaa7fa1ad578d3b5b15f83f5fac2..0d9dcc325231e50439c50f86ffede8e8fe24a096 100644 --- a/model-00047-of-00061.safetensors +++ b/model-00047-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:fde1a6af79a036be96b7e7c385900bdc295d0f8e20e5d3814aaa907c162152f5 -size 4373612200 +oid sha256:2869706ccff30fab1df0c13d1a22694effe155ae2d5ed60c6839ec55f772c772 +size 4362142872 diff --git a/model-00048-of-00061.safetensors b/model-00048-of-00061.safetensors index 6ee9ca2af5d6ae912eed6206272ffe6fd06e0ed8..da514573d0d62f6cd17859e6148f5651a1e5ec3c 100644 --- a/model-00048-of-00061.safetensors +++ b/model-00048-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:ce67f6a4d4b9fe331baedc3133abef8a260a60fe37ed8413a379cdd0dd2ae575 -size 4373612200 +oid sha256:efb893a5d02baeac6f1ed1dde6702a625a57deba7414d0fcecccc9a28b032f6c +size 4362142872 diff --git a/model-00049-of-00061.safetensors b/model-00049-of-00061.safetensors index 72a97e0f0c626c8d52cda3096dfd23fd9151d209..0ae52c152fdfbe73535a6337eeff0cd4ebb9e4c8 100644 --- a/model-00049-of-00061.safetensors +++ b/model-00049-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:bfc59ca81c4a8f85dd723149149a3c585bada9c0da7996a36b8ee81fca162e12 -size 4989127528 +oid sha256:93e12af3baeb84343a1e9b8b23e21fb3affd540daafe1f3acc2e763ca1cdb267 +size 4966188880 diff --git a/model-00050-of-00061.safetensors b/model-00050-of-00061.safetensors index 6568bd784169eb8477fa0f8da6e71d3a8ed9886e..dc3a6eb82b8f25609554d1ee48c601b11d88333c 100644 --- a/model-00050-of-00061.safetensors +++ b/model-00050-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:eefff4df3f049d652b9bc9c24a5112107067c2e07a21b891b496ea38d2b7c689 -size 4373612200 +oid sha256:098dea268a48cadc54b76c98eb5c4bbc4d3074b5aa56f71165fc3626cd75f85b +size 4362142872 diff --git a/model-00051-of-00061.safetensors b/model-00051-of-00061.safetensors index c523205186ca88bbcf2560881979836954e2d44c..f8a128554314ba024f1d522945d73178b7e61445 100644 --- a/model-00051-of-00061.safetensors +++ b/model-00051-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:012b4bc81e1f439f67920ca9109ccabe10cdab0182c3e687bd389ec6663b1fcd -size 4373612200 +oid sha256:c398c756012d8a47b78b06ec09c973bbcbfadbe68df99a4b618692fec9832351 +size 4362142872 diff --git a/model-00052-of-00061.safetensors b/model-00052-of-00061.safetensors index 47e4bbb4a243352cf63d5075168a6246af925c25..822a40546e154aa249ae05d69bc447ec434f8ed7 100644 --- a/model-00052-of-00061.safetensors +++ b/model-00052-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:5be8a68f8243469be9feeedfce196d133b28d5ee239af2e227c22aeb018e487b -size 4989127528 +oid sha256:1821ecbe328f5187ad540f0ad14681bd213c766af18453da6a2dd3d32ea9c4f1 +size 4966188880 diff --git a/model-00053-of-00061.safetensors b/model-00053-of-00061.safetensors index fca1fe02748a0cac85d8e95512c0cf32aa6dabd0..53397e59b05b6a85f19f742a48dd386fb5712170 100644 --- a/model-00053-of-00061.safetensors +++ b/model-00053-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:9637c9df48a125275b77e534fb5d44ed4587ad484b58224b1d4218035bde7e22 -size 4373612200 +oid sha256:56aaf14146f5381b567a34a1d80b6a0192333745999d61718ae37628b7043cfc +size 4362142872 diff --git a/model-00054-of-00061.safetensors b/model-00054-of-00061.safetensors index f09c4a8b118c473eac1c1aea18d562285c263470..f537d4576c344257f8846eeac16429d39e069051 100644 --- a/model-00054-of-00061.safetensors +++ b/model-00054-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:cc5774a350a4022293ec74dea3c132433aa2ac621fa3b1528506034a4567e7de -size 4373612200 +oid sha256:ff7004c81da2797bd2d53be185e929de0f33ce4d810ec2af2ecf142d8f73dee0 +size 4362142872 diff --git a/model-00055-of-00061.safetensors b/model-00055-of-00061.safetensors index c920f493ad271a24a98cd0a01905c7ff4f9e5e4b..d17ab14f41bba01b2e32dfe7de4f0d3ee19d30b5 100644 --- a/model-00055-of-00061.safetensors +++ b/model-00055-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:a3fa07dea54956d3e871a662f62c0681f35fe347a0794dd9ab48af98e2d93f1f -size 4989127528 +oid sha256:6a156aa489404a215f1d1cefcf456d828d6966e797093def1e5c3659838536a4 +size 4966188880 diff --git a/model-00056-of-00061.safetensors b/model-00056-of-00061.safetensors index bb0e7679fc55709b2dae071fd2fadfca9eb4e806..9eaf77a500bb3efec14f22e515f4428fb5bb935f 100644 --- a/model-00056-of-00061.safetensors +++ b/model-00056-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:878b0eeed0949288ed3e6a52b41e572b324c55ae89709e8a863f605448aad70c -size 4373612200 +oid sha256:305aa8c6af889bb3aa6c9e61d072d1a2b135f03a8b50825520e38c61d30a090a +size 4362142872 diff --git a/model-00057-of-00061.safetensors b/model-00057-of-00061.safetensors index 7c0f690a12f02b6084e38d239578544c9acc8d9e..4d107fa92b7ad11061ca7d817d420bc3a85c6823 100644 --- a/model-00057-of-00061.safetensors +++ b/model-00057-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:7a1bab1b9611953a58483129712df3d7ffb17d6171f23df3b2ccc4fee67c96b6 -size 4373612200 +oid sha256:174bbf1f2b706b3a05736865c40a8b95549c720806f9d5462c768b0f58fd81a9 +size 4362142872 diff --git a/model-00058-of-00061.safetensors b/model-00058-of-00061.safetensors index 28b7d53315dfce763acee29f40f5891be05835f8..9889043eba08998704ce2097ed7b60eb8de5fef3 100644 --- a/model-00058-of-00061.safetensors +++ b/model-00058-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:f0bdfefa9c2d6b65477995975c1a3e22860152d63b6f51412be6b5b27f664391 -size 4989127528 +oid sha256:1ddb7f5edaffa08c30aff90bcaacea0f29011f03ce27810afaea5c41baf3522e +size 4966188880 diff --git a/model-00059-of-00061.safetensors b/model-00059-of-00061.safetensors index c3bab595346d3d5513d10b296b4c35c6488c8003..c9aa257e9ee0337e0687f08bc1f731dbb0e68e34 100644 --- a/model-00059-of-00061.safetensors +++ b/model-00059-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:d3e95bfa154015e1b7b6d8d3d0b2849afb131ae960a9eb7e3ca907da47ead523 -size 4373612200 +oid sha256:e45d5c9f25fcce5953006452e8d8ad35afc8dcd77d9b3459c15b789b989e2b26 +size 4362142872 diff --git a/model-00060-of-00061.safetensors b/model-00060-of-00061.safetensors index 28388d212ba8778213311c0dba096f5720c630a0..3b2308046c2e87d9f7e95dd1b3ffed57ca3e875e 100644 --- a/model-00060-of-00061.safetensors +++ b/model-00060-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:e5f9c3c6844cdb028e95f33f6871a7d7355eca6a12668174070fa211ee565a0c -size 4373612200 +oid sha256:9da33a677ced328c5d5ca2fd6d195ab32f25c1d3212382c0cd6cbb37080ae0d7 +size 4362142872 diff --git a/model-00061-of-00061.safetensors b/model-00061-of-00061.safetensors index eed51e430e97073d5a3526ee1619767e1084813d..8c5364bc3f0b5e75b7762bbb3948a07776169aa4 100644 --- a/model-00061-of-00061.safetensors +++ b/model-00061-of-00061.safetensors @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:c25429c3da71e1db5ad2b6d7cd64f22675a117f9d9967881bba38219ea7d5068 +oid sha256:9c258787badeecfd94315dbef93055566eff4434d4d0572a37345b65f1b61452 size 1988198960 diff --git a/model.safetensors.index.json b/model.safetensors.index.json index efccb6e9b0e614705696e3583f0201b6fc7c1f8a..4156aa9dfeed60097cda66ec60af33ad1607de57 100644 --- a/model.safetensors.index.json +++ b/model.safetensors.index.json @@ -1,6 +1,6 @@ { "metadata": { - "total_size": 276824096768 + "total_size": 275906592768 }, "weight_map": { "lm_head.weight": "model-00061-of-00061.safetensors", @@ -12,11 +12,7 @@ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00061.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00061.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00061.safetensors", - "model.layers.0.self_attn.q_proj.lora_A.default.weight": "model-00001-of-00061.safetensors", - "model.layers.0.self_attn.q_proj.lora_B.default.weight": "model-00001-of-00061.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00061.safetensors", - "model.layers.0.self_attn.v_proj.lora_A.default.weight": "model-00001-of-00061.safetensors", - "model.layers.0.self_attn.v_proj.lora_B.default.weight": "model-00001-of-00061.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00061.safetensors", "model.layers.1.input_layernorm.weight": "model-00002-of-00061.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00002-of-00061.safetensors", @@ -25,11 +21,7 @@ "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00061.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00061.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00061.safetensors", - "model.layers.1.self_attn.q_proj.lora_A.default.weight": "model-00001-of-00061.safetensors", - "model.layers.1.self_attn.q_proj.lora_B.default.weight": "model-00001-of-00061.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00061.safetensors", - "model.layers.1.self_attn.v_proj.lora_A.default.weight": "model-00001-of-00061.safetensors", - "model.layers.1.self_attn.v_proj.lora_B.default.weight": "model-00001-of-00061.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00061.safetensors", "model.layers.10.input_layernorm.weight": "model-00009-of-00061.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00009-of-00061.safetensors", @@ -38,11 +30,7 @@ "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00061.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00061.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00061.safetensors", - "model.layers.10.self_attn.q_proj.lora_A.default.weight": "model-00008-of-00061.safetensors", - "model.layers.10.self_attn.q_proj.lora_B.default.weight": "model-00008-of-00061.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00061.safetensors", - "model.layers.10.self_attn.v_proj.lora_A.default.weight": "model-00008-of-00061.safetensors", - "model.layers.10.self_attn.v_proj.lora_B.default.weight": "model-00008-of-00061.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00061.safetensors", "model.layers.11.input_layernorm.weight": "model-00010-of-00061.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00010-of-00061.safetensors", @@ -51,11 +39,7 @@ "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00061.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00061.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00061.safetensors", - "model.layers.11.self_attn.q_proj.lora_A.default.weight": "model-00009-of-00061.safetensors", - "model.layers.11.self_attn.q_proj.lora_B.default.weight": "model-00009-of-00061.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00061.safetensors", - "model.layers.11.self_attn.v_proj.lora_A.default.weight": "model-00009-of-00061.safetensors", - "model.layers.11.self_attn.v_proj.lora_B.default.weight": "model-00009-of-00061.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00061.safetensors", "model.layers.12.input_layernorm.weight": "model-00010-of-00061.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00010-of-00061.safetensors", @@ -64,11 +48,7 @@ "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00061.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00061.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00061.safetensors", - "model.layers.12.self_attn.q_proj.lora_A.default.weight": "model-00010-of-00061.safetensors", - "model.layers.12.self_attn.q_proj.lora_B.default.weight": "model-00010-of-00061.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00061.safetensors", - "model.layers.12.self_attn.v_proj.lora_A.default.weight": "model-00010-of-00061.safetensors", - "model.layers.12.self_attn.v_proj.lora_B.default.weight": "model-00010-of-00061.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00061.safetensors", "model.layers.13.input_layernorm.weight": "model-00011-of-00061.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00011-of-00061.safetensors", @@ -77,11 +57,7 @@ "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00061.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00061.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00061.safetensors", - "model.layers.13.self_attn.q_proj.lora_A.default.weight": "model-00010-of-00061.safetensors", - "model.layers.13.self_attn.q_proj.lora_B.default.weight": "model-00010-of-00061.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00061.safetensors", - "model.layers.13.self_attn.v_proj.lora_A.default.weight": "model-00010-of-00061.safetensors", - "model.layers.13.self_attn.v_proj.lora_B.default.weight": "model-00010-of-00061.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00061.safetensors", "model.layers.14.input_layernorm.weight": "model-00012-of-00061.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00012-of-00061.safetensors", @@ -90,11 +66,7 @@ "model.layers.14.post_attention_layernorm.weight": "model-00012-of-00061.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00061.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00061.safetensors", - "model.layers.14.self_attn.q_proj.lora_A.default.weight": "model-00011-of-00061.safetensors", - "model.layers.14.self_attn.q_proj.lora_B.default.weight": "model-00011-of-00061.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00061.safetensors", - "model.layers.14.self_attn.v_proj.lora_A.default.weight": "model-00011-of-00061.safetensors", - "model.layers.14.self_attn.v_proj.lora_B.default.weight": "model-00011-of-00061.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00061.safetensors", "model.layers.15.input_layernorm.weight": "model-00013-of-00061.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00013-of-00061.safetensors", @@ -103,11 +75,7 @@ "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00061.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00012-of-00061.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00012-of-00061.safetensors", - "model.layers.15.self_attn.q_proj.lora_A.default.weight": "model-00012-of-00061.safetensors", - "model.layers.15.self_attn.q_proj.lora_B.default.weight": "model-00012-of-00061.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00012-of-00061.safetensors", - "model.layers.15.self_attn.v_proj.lora_A.default.weight": "model-00012-of-00061.safetensors", - "model.layers.15.self_attn.v_proj.lora_B.default.weight": "model-00012-of-00061.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00012-of-00061.safetensors", "model.layers.16.input_layernorm.weight": "model-00013-of-00061.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00013-of-00061.safetensors", @@ -116,11 +84,7 @@ "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00061.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00061.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00061.safetensors", - "model.layers.16.self_attn.q_proj.lora_A.default.weight": "model-00013-of-00061.safetensors", - "model.layers.16.self_attn.q_proj.lora_B.default.weight": "model-00013-of-00061.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00061.safetensors", - "model.layers.16.self_attn.v_proj.lora_A.default.weight": "model-00013-of-00061.safetensors", - "model.layers.16.self_attn.v_proj.lora_B.default.weight": "model-00013-of-00061.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00061.safetensors", "model.layers.17.input_layernorm.weight": "model-00014-of-00061.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00014-of-00061.safetensors", @@ -129,11 +93,7 @@ "model.layers.17.post_attention_layernorm.weight": "model-00014-of-00061.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00061.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00061.safetensors", - "model.layers.17.self_attn.q_proj.lora_A.default.weight": "model-00013-of-00061.safetensors", - "model.layers.17.self_attn.q_proj.lora_B.default.weight": "model-00013-of-00061.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00061.safetensors", - "model.layers.17.self_attn.v_proj.lora_A.default.weight": "model-00013-of-00061.safetensors", - "model.layers.17.self_attn.v_proj.lora_B.default.weight": "model-00013-of-00061.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00061.safetensors", "model.layers.18.input_layernorm.weight": "model-00015-of-00061.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00015-of-00061.safetensors", @@ -142,11 +102,7 @@ "model.layers.18.post_attention_layernorm.weight": "model-00015-of-00061.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00014-of-00061.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00014-of-00061.safetensors", - "model.layers.18.self_attn.q_proj.lora_A.default.weight": "model-00014-of-00061.safetensors", - "model.layers.18.self_attn.q_proj.lora_B.default.weight": "model-00014-of-00061.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00014-of-00061.safetensors", - "model.layers.18.self_attn.v_proj.lora_A.default.weight": "model-00014-of-00061.safetensors", - "model.layers.18.self_attn.v_proj.lora_B.default.weight": "model-00014-of-00061.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00014-of-00061.safetensors", "model.layers.19.input_layernorm.weight": "model-00016-of-00061.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00016-of-00061.safetensors", @@ -155,11 +111,7 @@ "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00061.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00015-of-00061.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00015-of-00061.safetensors", - "model.layers.19.self_attn.q_proj.lora_A.default.weight": "model-00015-of-00061.safetensors", - "model.layers.19.self_attn.q_proj.lora_B.default.weight": "model-00015-of-00061.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00015-of-00061.safetensors", - "model.layers.19.self_attn.v_proj.lora_A.default.weight": "model-00015-of-00061.safetensors", - "model.layers.19.self_attn.v_proj.lora_B.default.weight": "model-00015-of-00061.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00015-of-00061.safetensors", "model.layers.2.input_layernorm.weight": "model-00003-of-00061.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00003-of-00061.safetensors", @@ -168,11 +120,7 @@ "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00061.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00061.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00061.safetensors", - "model.layers.2.self_attn.q_proj.lora_A.default.weight": "model-00002-of-00061.safetensors", - "model.layers.2.self_attn.q_proj.lora_B.default.weight": "model-00002-of-00061.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00061.safetensors", - "model.layers.2.self_attn.v_proj.lora_A.default.weight": "model-00002-of-00061.safetensors", - "model.layers.2.self_attn.v_proj.lora_B.default.weight": "model-00002-of-00061.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00061.safetensors", "model.layers.20.input_layernorm.weight": "model-00016-of-00061.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00016-of-00061.safetensors", @@ -181,11 +129,7 @@ "model.layers.20.post_attention_layernorm.weight": "model-00016-of-00061.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00061.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00061.safetensors", - "model.layers.20.self_attn.q_proj.lora_A.default.weight": "model-00016-of-00061.safetensors", - "model.layers.20.self_attn.q_proj.lora_B.default.weight": "model-00016-of-00061.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00061.safetensors", - "model.layers.20.self_attn.v_proj.lora_A.default.weight": "model-00016-of-00061.safetensors", - "model.layers.20.self_attn.v_proj.lora_B.default.weight": "model-00016-of-00061.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00061.safetensors", "model.layers.21.input_layernorm.weight": "model-00017-of-00061.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00017-of-00061.safetensors", @@ -194,11 +138,7 @@ "model.layers.21.post_attention_layernorm.weight": "model-00017-of-00061.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00061.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00061.safetensors", - "model.layers.21.self_attn.q_proj.lora_A.default.weight": "model-00016-of-00061.safetensors", - "model.layers.21.self_attn.q_proj.lora_B.default.weight": "model-00016-of-00061.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00061.safetensors", - "model.layers.21.self_attn.v_proj.lora_A.default.weight": "model-00016-of-00061.safetensors", - "model.layers.21.self_attn.v_proj.lora_B.default.weight": "model-00016-of-00061.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00061.safetensors", "model.layers.22.input_layernorm.weight": "model-00018-of-00061.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00018-of-00061.safetensors", @@ -207,11 +147,7 @@ "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00061.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00061.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00061.safetensors", - "model.layers.22.self_attn.q_proj.lora_A.default.weight": "model-00017-of-00061.safetensors", - "model.layers.22.self_attn.q_proj.lora_B.default.weight": "model-00017-of-00061.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00061.safetensors", - "model.layers.22.self_attn.v_proj.lora_A.default.weight": "model-00017-of-00061.safetensors", - "model.layers.22.self_attn.v_proj.lora_B.default.weight": "model-00017-of-00061.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00061.safetensors", "model.layers.23.input_layernorm.weight": "model-00019-of-00061.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00019-of-00061.safetensors", @@ -220,11 +156,7 @@ "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00061.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00061.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00061.safetensors", - "model.layers.23.self_attn.q_proj.lora_A.default.weight": "model-00018-of-00061.safetensors", - "model.layers.23.self_attn.q_proj.lora_B.default.weight": "model-00018-of-00061.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00061.safetensors", - "model.layers.23.self_attn.v_proj.lora_A.default.weight": "model-00018-of-00061.safetensors", - "model.layers.23.self_attn.v_proj.lora_B.default.weight": "model-00018-of-00061.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00061.safetensors", "model.layers.24.input_layernorm.weight": "model-00019-of-00061.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00019-of-00061.safetensors", @@ -233,11 +165,7 @@ "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00061.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00061.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00061.safetensors", - "model.layers.24.self_attn.q_proj.lora_A.default.weight": "model-00019-of-00061.safetensors", - "model.layers.24.self_attn.q_proj.lora_B.default.weight": "model-00019-of-00061.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00061.safetensors", - "model.layers.24.self_attn.v_proj.lora_A.default.weight": "model-00019-of-00061.safetensors", - "model.layers.24.self_attn.v_proj.lora_B.default.weight": "model-00019-of-00061.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00061.safetensors", "model.layers.25.input_layernorm.weight": "model-00020-of-00061.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00020-of-00061.safetensors", @@ -246,11 +174,7 @@ "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00061.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00019-of-00061.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00061.safetensors", - "model.layers.25.self_attn.q_proj.lora_A.default.weight": "model-00019-of-00061.safetensors", - "model.layers.25.self_attn.q_proj.lora_B.default.weight": "model-00019-of-00061.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00019-of-00061.safetensors", - "model.layers.25.self_attn.v_proj.lora_A.default.weight": "model-00019-of-00061.safetensors", - "model.layers.25.self_attn.v_proj.lora_B.default.weight": "model-00019-of-00061.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00061.safetensors", "model.layers.26.input_layernorm.weight": "model-00021-of-00061.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00021-of-00061.safetensors", @@ -259,11 +183,7 @@ "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00061.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00020-of-00061.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00020-of-00061.safetensors", - "model.layers.26.self_attn.q_proj.lora_A.default.weight": "model-00020-of-00061.safetensors", - "model.layers.26.self_attn.q_proj.lora_B.default.weight": "model-00020-of-00061.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00020-of-00061.safetensors", - "model.layers.26.self_attn.v_proj.lora_A.default.weight": "model-00020-of-00061.safetensors", - "model.layers.26.self_attn.v_proj.lora_B.default.weight": "model-00020-of-00061.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00020-of-00061.safetensors", "model.layers.27.input_layernorm.weight": "model-00022-of-00061.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00022-of-00061.safetensors", @@ -272,11 +192,7 @@ "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00061.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00021-of-00061.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00021-of-00061.safetensors", - "model.layers.27.self_attn.q_proj.lora_A.default.weight": "model-00021-of-00061.safetensors", - "model.layers.27.self_attn.q_proj.lora_B.default.weight": "model-00021-of-00061.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00021-of-00061.safetensors", - "model.layers.27.self_attn.v_proj.lora_A.default.weight": "model-00021-of-00061.safetensors", - "model.layers.27.self_attn.v_proj.lora_B.default.weight": "model-00021-of-00061.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00021-of-00061.safetensors", "model.layers.28.input_layernorm.weight": "model-00022-of-00061.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00022-of-00061.safetensors", @@ -285,11 +201,7 @@ "model.layers.28.post_attention_layernorm.weight": "model-00022-of-00061.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00061.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00061.safetensors", - "model.layers.28.self_attn.q_proj.lora_A.default.weight": "model-00022-of-00061.safetensors", - "model.layers.28.self_attn.q_proj.lora_B.default.weight": "model-00022-of-00061.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00061.safetensors", - "model.layers.28.self_attn.v_proj.lora_A.default.weight": "model-00022-of-00061.safetensors", - "model.layers.28.self_attn.v_proj.lora_B.default.weight": "model-00022-of-00061.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00061.safetensors", "model.layers.29.input_layernorm.weight": "model-00023-of-00061.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00023-of-00061.safetensors", @@ -298,11 +210,7 @@ "model.layers.29.post_attention_layernorm.weight": "model-00023-of-00061.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00022-of-00061.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00022-of-00061.safetensors", - "model.layers.29.self_attn.q_proj.lora_A.default.weight": "model-00022-of-00061.safetensors", - "model.layers.29.self_attn.q_proj.lora_B.default.weight": "model-00022-of-00061.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00022-of-00061.safetensors", - "model.layers.29.self_attn.v_proj.lora_A.default.weight": "model-00022-of-00061.safetensors", - "model.layers.29.self_attn.v_proj.lora_B.default.weight": "model-00022-of-00061.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00022-of-00061.safetensors", "model.layers.3.input_layernorm.weight": "model-00004-of-00061.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00004-of-00061.safetensors", @@ -311,11 +219,7 @@ "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00061.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00061.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00061.safetensors", - "model.layers.3.self_attn.q_proj.lora_A.default.weight": "model-00003-of-00061.safetensors", - "model.layers.3.self_attn.q_proj.lora_B.default.weight": "model-00003-of-00061.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00061.safetensors", - "model.layers.3.self_attn.v_proj.lora_A.default.weight": "model-00003-of-00061.safetensors", - "model.layers.3.self_attn.v_proj.lora_B.default.weight": "model-00003-of-00061.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00061.safetensors", "model.layers.30.input_layernorm.weight": "model-00024-of-00061.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00024-of-00061.safetensors", @@ -324,11 +228,7 @@ "model.layers.30.post_attention_layernorm.weight": "model-00024-of-00061.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00023-of-00061.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00023-of-00061.safetensors", - "model.layers.30.self_attn.q_proj.lora_A.default.weight": "model-00023-of-00061.safetensors", - "model.layers.30.self_attn.q_proj.lora_B.default.weight": "model-00023-of-00061.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00023-of-00061.safetensors", - "model.layers.30.self_attn.v_proj.lora_A.default.weight": "model-00023-of-00061.safetensors", - "model.layers.30.self_attn.v_proj.lora_B.default.weight": "model-00023-of-00061.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00023-of-00061.safetensors", "model.layers.31.input_layernorm.weight": "model-00025-of-00061.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00025-of-00061.safetensors", @@ -337,11 +237,7 @@ "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00061.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00024-of-00061.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00024-of-00061.safetensors", - "model.layers.31.self_attn.q_proj.lora_A.default.weight": "model-00024-of-00061.safetensors", - "model.layers.31.self_attn.q_proj.lora_B.default.weight": "model-00024-of-00061.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00024-of-00061.safetensors", - "model.layers.31.self_attn.v_proj.lora_A.default.weight": "model-00024-of-00061.safetensors", - "model.layers.31.self_attn.v_proj.lora_B.default.weight": "model-00024-of-00061.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00024-of-00061.safetensors", "model.layers.32.input_layernorm.weight": "model-00025-of-00061.safetensors", "model.layers.32.mlp.down_proj.weight": "model-00025-of-00061.safetensors", @@ -350,11 +246,7 @@ "model.layers.32.post_attention_layernorm.weight": "model-00025-of-00061.safetensors", "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00061.safetensors", "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00061.safetensors", - "model.layers.32.self_attn.q_proj.lora_A.default.weight": "model-00025-of-00061.safetensors", - "model.layers.32.self_attn.q_proj.lora_B.default.weight": "model-00025-of-00061.safetensors", "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00061.safetensors", - "model.layers.32.self_attn.v_proj.lora_A.default.weight": "model-00025-of-00061.safetensors", - "model.layers.32.self_attn.v_proj.lora_B.default.weight": "model-00025-of-00061.safetensors", "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00061.safetensors", "model.layers.33.input_layernorm.weight": "model-00026-of-00061.safetensors", "model.layers.33.mlp.down_proj.weight": "model-00026-of-00061.safetensors", @@ -363,11 +255,7 @@ "model.layers.33.post_attention_layernorm.weight": "model-00026-of-00061.safetensors", "model.layers.33.self_attn.k_proj.weight": "model-00025-of-00061.safetensors", "model.layers.33.self_attn.o_proj.weight": "model-00025-of-00061.safetensors", - "model.layers.33.self_attn.q_proj.lora_A.default.weight": "model-00025-of-00061.safetensors", - "model.layers.33.self_attn.q_proj.lora_B.default.weight": "model-00025-of-00061.safetensors", "model.layers.33.self_attn.q_proj.weight": "model-00025-of-00061.safetensors", - "model.layers.33.self_attn.v_proj.lora_A.default.weight": "model-00025-of-00061.safetensors", - "model.layers.33.self_attn.v_proj.lora_B.default.weight": "model-00025-of-00061.safetensors", "model.layers.33.self_attn.v_proj.weight": "model-00025-of-00061.safetensors", "model.layers.34.input_layernorm.weight": "model-00027-of-00061.safetensors", "model.layers.34.mlp.down_proj.weight": "model-00027-of-00061.safetensors", @@ -376,11 +264,7 @@ "model.layers.34.post_attention_layernorm.weight": "model-00027-of-00061.safetensors", "model.layers.34.self_attn.k_proj.weight": "model-00026-of-00061.safetensors", "model.layers.34.self_attn.o_proj.weight": "model-00026-of-00061.safetensors", - "model.layers.34.self_attn.q_proj.lora_A.default.weight": "model-00026-of-00061.safetensors", - "model.layers.34.self_attn.q_proj.lora_B.default.weight": "model-00026-of-00061.safetensors", "model.layers.34.self_attn.q_proj.weight": "model-00026-of-00061.safetensors", - "model.layers.34.self_attn.v_proj.lora_A.default.weight": "model-00026-of-00061.safetensors", - "model.layers.34.self_attn.v_proj.lora_B.default.weight": "model-00026-of-00061.safetensors", "model.layers.34.self_attn.v_proj.weight": "model-00026-of-00061.safetensors", "model.layers.35.input_layernorm.weight": "model-00028-of-00061.safetensors", "model.layers.35.mlp.down_proj.weight": "model-00028-of-00061.safetensors", @@ -389,11 +273,7 @@ "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00061.safetensors", "model.layers.35.self_attn.k_proj.weight": "model-00027-of-00061.safetensors", "model.layers.35.self_attn.o_proj.weight": "model-00027-of-00061.safetensors", - "model.layers.35.self_attn.q_proj.lora_A.default.weight": "model-00027-of-00061.safetensors", - "model.layers.35.self_attn.q_proj.lora_B.default.weight": "model-00027-of-00061.safetensors", "model.layers.35.self_attn.q_proj.weight": "model-00027-of-00061.safetensors", - "model.layers.35.self_attn.v_proj.lora_A.default.weight": "model-00027-of-00061.safetensors", - "model.layers.35.self_attn.v_proj.lora_B.default.weight": "model-00027-of-00061.safetensors", "model.layers.35.self_attn.v_proj.weight": "model-00027-of-00061.safetensors", "model.layers.36.input_layernorm.weight": "model-00028-of-00061.safetensors", "model.layers.36.mlp.down_proj.weight": "model-00028-of-00061.safetensors", @@ -402,11 +282,7 @@ "model.layers.36.post_attention_layernorm.weight": "model-00028-of-00061.safetensors", "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00061.safetensors", "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00061.safetensors", - "model.layers.36.self_attn.q_proj.lora_A.default.weight": "model-00028-of-00061.safetensors", - "model.layers.36.self_attn.q_proj.lora_B.default.weight": "model-00028-of-00061.safetensors", "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00061.safetensors", - "model.layers.36.self_attn.v_proj.lora_A.default.weight": "model-00028-of-00061.safetensors", - "model.layers.36.self_attn.v_proj.lora_B.default.weight": "model-00028-of-00061.safetensors", "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00061.safetensors", "model.layers.37.input_layernorm.weight": "model-00029-of-00061.safetensors", "model.layers.37.mlp.down_proj.weight": "model-00029-of-00061.safetensors", @@ -415,11 +291,7 @@ "model.layers.37.post_attention_layernorm.weight": "model-00029-of-00061.safetensors", "model.layers.37.self_attn.k_proj.weight": "model-00028-of-00061.safetensors", "model.layers.37.self_attn.o_proj.weight": "model-00028-of-00061.safetensors", - "model.layers.37.self_attn.q_proj.lora_A.default.weight": "model-00028-of-00061.safetensors", - "model.layers.37.self_attn.q_proj.lora_B.default.weight": "model-00028-of-00061.safetensors", "model.layers.37.self_attn.q_proj.weight": "model-00028-of-00061.safetensors", - "model.layers.37.self_attn.v_proj.lora_A.default.weight": "model-00028-of-00061.safetensors", - "model.layers.37.self_attn.v_proj.lora_B.default.weight": "model-00028-of-00061.safetensors", "model.layers.37.self_attn.v_proj.weight": "model-00028-of-00061.safetensors", "model.layers.38.input_layernorm.weight": "model-00030-of-00061.safetensors", "model.layers.38.mlp.down_proj.weight": "model-00030-of-00061.safetensors", @@ -428,11 +300,7 @@ "model.layers.38.post_attention_layernorm.weight": "model-00030-of-00061.safetensors", "model.layers.38.self_attn.k_proj.weight": "model-00029-of-00061.safetensors", "model.layers.38.self_attn.o_proj.weight": "model-00029-of-00061.safetensors", - "model.layers.38.self_attn.q_proj.lora_A.default.weight": "model-00029-of-00061.safetensors", - "model.layers.38.self_attn.q_proj.lora_B.default.weight": "model-00029-of-00061.safetensors", "model.layers.38.self_attn.q_proj.weight": "model-00029-of-00061.safetensors", - "model.layers.38.self_attn.v_proj.lora_A.default.weight": "model-00029-of-00061.safetensors", - "model.layers.38.self_attn.v_proj.lora_B.default.weight": "model-00029-of-00061.safetensors", "model.layers.38.self_attn.v_proj.weight": "model-00029-of-00061.safetensors", "model.layers.39.input_layernorm.weight": "model-00031-of-00061.safetensors", "model.layers.39.mlp.down_proj.weight": "model-00031-of-00061.safetensors", @@ -441,11 +309,7 @@ "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00061.safetensors", "model.layers.39.self_attn.k_proj.weight": "model-00030-of-00061.safetensors", "model.layers.39.self_attn.o_proj.weight": "model-00030-of-00061.safetensors", - "model.layers.39.self_attn.q_proj.lora_A.default.weight": "model-00030-of-00061.safetensors", - "model.layers.39.self_attn.q_proj.lora_B.default.weight": "model-00030-of-00061.safetensors", "model.layers.39.self_attn.q_proj.weight": "model-00030-of-00061.safetensors", - "model.layers.39.self_attn.v_proj.lora_A.default.weight": "model-00030-of-00061.safetensors", - "model.layers.39.self_attn.v_proj.lora_B.default.weight": "model-00030-of-00061.safetensors", "model.layers.39.self_attn.v_proj.weight": "model-00030-of-00061.safetensors", "model.layers.4.input_layernorm.weight": "model-00004-of-00061.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00004-of-00061.safetensors", @@ -454,11 +318,7 @@ "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00061.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00061.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00061.safetensors", - "model.layers.4.self_attn.q_proj.lora_A.default.weight": "model-00004-of-00061.safetensors", - "model.layers.4.self_attn.q_proj.lora_B.default.weight": "model-00004-of-00061.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00061.safetensors", - "model.layers.4.self_attn.v_proj.lora_A.default.weight": "model-00004-of-00061.safetensors", - "model.layers.4.self_attn.v_proj.lora_B.default.weight": "model-00004-of-00061.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00061.safetensors", "model.layers.40.input_layernorm.weight": "model-00031-of-00061.safetensors", "model.layers.40.mlp.down_proj.weight": "model-00031-of-00061.safetensors", @@ -467,11 +327,7 @@ "model.layers.40.post_attention_layernorm.weight": "model-00031-of-00061.safetensors", "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00061.safetensors", "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00061.safetensors", - "model.layers.40.self_attn.q_proj.lora_A.default.weight": "model-00031-of-00061.safetensors", - "model.layers.40.self_attn.q_proj.lora_B.default.weight": "model-00031-of-00061.safetensors", "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00061.safetensors", - "model.layers.40.self_attn.v_proj.lora_A.default.weight": "model-00031-of-00061.safetensors", - "model.layers.40.self_attn.v_proj.lora_B.default.weight": "model-00031-of-00061.safetensors", "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00061.safetensors", "model.layers.41.input_layernorm.weight": "model-00032-of-00061.safetensors", "model.layers.41.mlp.down_proj.weight": "model-00032-of-00061.safetensors", @@ -480,11 +336,7 @@ "model.layers.41.post_attention_layernorm.weight": "model-00032-of-00061.safetensors", "model.layers.41.self_attn.k_proj.weight": "model-00031-of-00061.safetensors", "model.layers.41.self_attn.o_proj.weight": "model-00031-of-00061.safetensors", - "model.layers.41.self_attn.q_proj.lora_A.default.weight": "model-00031-of-00061.safetensors", - "model.layers.41.self_attn.q_proj.lora_B.default.weight": "model-00031-of-00061.safetensors", "model.layers.41.self_attn.q_proj.weight": "model-00031-of-00061.safetensors", - "model.layers.41.self_attn.v_proj.lora_A.default.weight": "model-00031-of-00061.safetensors", - "model.layers.41.self_attn.v_proj.lora_B.default.weight": "model-00031-of-00061.safetensors", "model.layers.41.self_attn.v_proj.weight": "model-00031-of-00061.safetensors", "model.layers.42.input_layernorm.weight": "model-00033-of-00061.safetensors", "model.layers.42.mlp.down_proj.weight": "model-00033-of-00061.safetensors", @@ -493,11 +345,7 @@ "model.layers.42.post_attention_layernorm.weight": "model-00033-of-00061.safetensors", "model.layers.42.self_attn.k_proj.weight": "model-00032-of-00061.safetensors", "model.layers.42.self_attn.o_proj.weight": "model-00032-of-00061.safetensors", - "model.layers.42.self_attn.q_proj.lora_A.default.weight": "model-00032-of-00061.safetensors", - "model.layers.42.self_attn.q_proj.lora_B.default.weight": "model-00032-of-00061.safetensors", "model.layers.42.self_attn.q_proj.weight": "model-00032-of-00061.safetensors", - "model.layers.42.self_attn.v_proj.lora_A.default.weight": "model-00032-of-00061.safetensors", - "model.layers.42.self_attn.v_proj.lora_B.default.weight": "model-00032-of-00061.safetensors", "model.layers.42.self_attn.v_proj.weight": "model-00032-of-00061.safetensors", "model.layers.43.input_layernorm.weight": "model-00034-of-00061.safetensors", "model.layers.43.mlp.down_proj.weight": "model-00034-of-00061.safetensors", @@ -506,11 +354,7 @@ "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00061.safetensors", "model.layers.43.self_attn.k_proj.weight": "model-00033-of-00061.safetensors", "model.layers.43.self_attn.o_proj.weight": "model-00033-of-00061.safetensors", - "model.layers.43.self_attn.q_proj.lora_A.default.weight": "model-00033-of-00061.safetensors", - "model.layers.43.self_attn.q_proj.lora_B.default.weight": "model-00033-of-00061.safetensors", "model.layers.43.self_attn.q_proj.weight": "model-00033-of-00061.safetensors", - "model.layers.43.self_attn.v_proj.lora_A.default.weight": "model-00033-of-00061.safetensors", - "model.layers.43.self_attn.v_proj.lora_B.default.weight": "model-00033-of-00061.safetensors", "model.layers.43.self_attn.v_proj.weight": "model-00033-of-00061.safetensors", "model.layers.44.input_layernorm.weight": "model-00034-of-00061.safetensors", "model.layers.44.mlp.down_proj.weight": "model-00034-of-00061.safetensors", @@ -519,11 +363,7 @@ "model.layers.44.post_attention_layernorm.weight": "model-00034-of-00061.safetensors", "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00061.safetensors", "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00061.safetensors", - "model.layers.44.self_attn.q_proj.lora_A.default.weight": "model-00034-of-00061.safetensors", - "model.layers.44.self_attn.q_proj.lora_B.default.weight": "model-00034-of-00061.safetensors", "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00061.safetensors", - "model.layers.44.self_attn.v_proj.lora_A.default.weight": "model-00034-of-00061.safetensors", - "model.layers.44.self_attn.v_proj.lora_B.default.weight": "model-00034-of-00061.safetensors", "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00061.safetensors", "model.layers.45.input_layernorm.weight": "model-00035-of-00061.safetensors", "model.layers.45.mlp.down_proj.weight": "model-00035-of-00061.safetensors", @@ -532,11 +372,7 @@ "model.layers.45.post_attention_layernorm.weight": "model-00035-of-00061.safetensors", "model.layers.45.self_attn.k_proj.weight": "model-00034-of-00061.safetensors", "model.layers.45.self_attn.o_proj.weight": "model-00034-of-00061.safetensors", - "model.layers.45.self_attn.q_proj.lora_A.default.weight": "model-00034-of-00061.safetensors", - "model.layers.45.self_attn.q_proj.lora_B.default.weight": "model-00034-of-00061.safetensors", "model.layers.45.self_attn.q_proj.weight": "model-00034-of-00061.safetensors", - "model.layers.45.self_attn.v_proj.lora_A.default.weight": "model-00034-of-00061.safetensors", - "model.layers.45.self_attn.v_proj.lora_B.default.weight": "model-00034-of-00061.safetensors", "model.layers.45.self_attn.v_proj.weight": "model-00034-of-00061.safetensors", "model.layers.46.input_layernorm.weight": "model-00036-of-00061.safetensors", "model.layers.46.mlp.down_proj.weight": "model-00036-of-00061.safetensors", @@ -545,11 +381,7 @@ "model.layers.46.post_attention_layernorm.weight": "model-00036-of-00061.safetensors", "model.layers.46.self_attn.k_proj.weight": "model-00035-of-00061.safetensors", "model.layers.46.self_attn.o_proj.weight": "model-00035-of-00061.safetensors", - "model.layers.46.self_attn.q_proj.lora_A.default.weight": "model-00035-of-00061.safetensors", - "model.layers.46.self_attn.q_proj.lora_B.default.weight": "model-00035-of-00061.safetensors", "model.layers.46.self_attn.q_proj.weight": "model-00035-of-00061.safetensors", - "model.layers.46.self_attn.v_proj.lora_A.default.weight": "model-00035-of-00061.safetensors", - "model.layers.46.self_attn.v_proj.lora_B.default.weight": "model-00035-of-00061.safetensors", "model.layers.46.self_attn.v_proj.weight": "model-00035-of-00061.safetensors", "model.layers.47.input_layernorm.weight": "model-00037-of-00061.safetensors", "model.layers.47.mlp.down_proj.weight": "model-00037-of-00061.safetensors", @@ -558,11 +390,7 @@ "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00061.safetensors", "model.layers.47.self_attn.k_proj.weight": "model-00036-of-00061.safetensors", "model.layers.47.self_attn.o_proj.weight": "model-00036-of-00061.safetensors", - "model.layers.47.self_attn.q_proj.lora_A.default.weight": "model-00036-of-00061.safetensors", - "model.layers.47.self_attn.q_proj.lora_B.default.weight": "model-00036-of-00061.safetensors", "model.layers.47.self_attn.q_proj.weight": "model-00036-of-00061.safetensors", - "model.layers.47.self_attn.v_proj.lora_A.default.weight": "model-00036-of-00061.safetensors", - "model.layers.47.self_attn.v_proj.lora_B.default.weight": "model-00036-of-00061.safetensors", "model.layers.47.self_attn.v_proj.weight": "model-00036-of-00061.safetensors", "model.layers.48.input_layernorm.weight": "model-00037-of-00061.safetensors", "model.layers.48.mlp.down_proj.weight": "model-00037-of-00061.safetensors", @@ -571,11 +399,7 @@ "model.layers.48.post_attention_layernorm.weight": "model-00037-of-00061.safetensors", "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00061.safetensors", "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00061.safetensors", - "model.layers.48.self_attn.q_proj.lora_A.default.weight": "model-00037-of-00061.safetensors", - "model.layers.48.self_attn.q_proj.lora_B.default.weight": "model-00037-of-00061.safetensors", "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00061.safetensors", - "model.layers.48.self_attn.v_proj.lora_A.default.weight": "model-00037-of-00061.safetensors", - "model.layers.48.self_attn.v_proj.lora_B.default.weight": "model-00037-of-00061.safetensors", "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00061.safetensors", "model.layers.49.input_layernorm.weight": "model-00038-of-00061.safetensors", "model.layers.49.mlp.down_proj.weight": "model-00038-of-00061.safetensors", @@ -584,11 +408,7 @@ "model.layers.49.post_attention_layernorm.weight": "model-00038-of-00061.safetensors", "model.layers.49.self_attn.k_proj.weight": "model-00037-of-00061.safetensors", "model.layers.49.self_attn.o_proj.weight": "model-00037-of-00061.safetensors", - "model.layers.49.self_attn.q_proj.lora_A.default.weight": "model-00037-of-00061.safetensors", - "model.layers.49.self_attn.q_proj.lora_B.default.weight": "model-00037-of-00061.safetensors", "model.layers.49.self_attn.q_proj.weight": "model-00037-of-00061.safetensors", - "model.layers.49.self_attn.v_proj.lora_A.default.weight": "model-00037-of-00061.safetensors", - "model.layers.49.self_attn.v_proj.lora_B.default.weight": "model-00037-of-00061.safetensors", "model.layers.49.self_attn.v_proj.weight": "model-00037-of-00061.safetensors", "model.layers.5.input_layernorm.weight": "model-00005-of-00061.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00005-of-00061.safetensors", @@ -597,11 +417,7 @@ "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00061.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00061.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00061.safetensors", - "model.layers.5.self_attn.q_proj.lora_A.default.weight": "model-00004-of-00061.safetensors", - "model.layers.5.self_attn.q_proj.lora_B.default.weight": "model-00004-of-00061.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00061.safetensors", - "model.layers.5.self_attn.v_proj.lora_A.default.weight": "model-00004-of-00061.safetensors", - "model.layers.5.self_attn.v_proj.lora_B.default.weight": "model-00004-of-00061.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00061.safetensors", "model.layers.50.input_layernorm.weight": "model-00039-of-00061.safetensors", "model.layers.50.mlp.down_proj.weight": "model-00039-of-00061.safetensors", @@ -610,11 +426,7 @@ "model.layers.50.post_attention_layernorm.weight": "model-00039-of-00061.safetensors", "model.layers.50.self_attn.k_proj.weight": "model-00038-of-00061.safetensors", "model.layers.50.self_attn.o_proj.weight": "model-00038-of-00061.safetensors", - "model.layers.50.self_attn.q_proj.lora_A.default.weight": "model-00038-of-00061.safetensors", - "model.layers.50.self_attn.q_proj.lora_B.default.weight": "model-00038-of-00061.safetensors", "model.layers.50.self_attn.q_proj.weight": "model-00038-of-00061.safetensors", - "model.layers.50.self_attn.v_proj.lora_A.default.weight": "model-00038-of-00061.safetensors", - "model.layers.50.self_attn.v_proj.lora_B.default.weight": "model-00038-of-00061.safetensors", "model.layers.50.self_attn.v_proj.weight": "model-00038-of-00061.safetensors", "model.layers.51.input_layernorm.weight": "model-00040-of-00061.safetensors", "model.layers.51.mlp.down_proj.weight": "model-00040-of-00061.safetensors", @@ -623,11 +435,7 @@ "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00061.safetensors", "model.layers.51.self_attn.k_proj.weight": "model-00039-of-00061.safetensors", "model.layers.51.self_attn.o_proj.weight": "model-00039-of-00061.safetensors", - "model.layers.51.self_attn.q_proj.lora_A.default.weight": "model-00039-of-00061.safetensors", - "model.layers.51.self_attn.q_proj.lora_B.default.weight": "model-00039-of-00061.safetensors", "model.layers.51.self_attn.q_proj.weight": "model-00039-of-00061.safetensors", - "model.layers.51.self_attn.v_proj.lora_A.default.weight": "model-00039-of-00061.safetensors", - "model.layers.51.self_attn.v_proj.lora_B.default.weight": "model-00039-of-00061.safetensors", "model.layers.51.self_attn.v_proj.weight": "model-00039-of-00061.safetensors", "model.layers.52.input_layernorm.weight": "model-00040-of-00061.safetensors", "model.layers.52.mlp.down_proj.weight": "model-00040-of-00061.safetensors", @@ -636,11 +444,7 @@ "model.layers.52.post_attention_layernorm.weight": "model-00040-of-00061.safetensors", "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00061.safetensors", "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00061.safetensors", - "model.layers.52.self_attn.q_proj.lora_A.default.weight": "model-00040-of-00061.safetensors", - "model.layers.52.self_attn.q_proj.lora_B.default.weight": "model-00040-of-00061.safetensors", "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00061.safetensors", - "model.layers.52.self_attn.v_proj.lora_A.default.weight": "model-00040-of-00061.safetensors", - "model.layers.52.self_attn.v_proj.lora_B.default.weight": "model-00040-of-00061.safetensors", "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00061.safetensors", "model.layers.53.input_layernorm.weight": "model-00041-of-00061.safetensors", "model.layers.53.mlp.down_proj.weight": "model-00041-of-00061.safetensors", @@ -649,11 +453,7 @@ "model.layers.53.post_attention_layernorm.weight": "model-00041-of-00061.safetensors", "model.layers.53.self_attn.k_proj.weight": "model-00040-of-00061.safetensors", "model.layers.53.self_attn.o_proj.weight": "model-00040-of-00061.safetensors", - "model.layers.53.self_attn.q_proj.lora_A.default.weight": "model-00040-of-00061.safetensors", - "model.layers.53.self_attn.q_proj.lora_B.default.weight": "model-00040-of-00061.safetensors", "model.layers.53.self_attn.q_proj.weight": "model-00040-of-00061.safetensors", - "model.layers.53.self_attn.v_proj.lora_A.default.weight": "model-00040-of-00061.safetensors", - "model.layers.53.self_attn.v_proj.lora_B.default.weight": "model-00040-of-00061.safetensors", "model.layers.53.self_attn.v_proj.weight": "model-00040-of-00061.safetensors", "model.layers.54.input_layernorm.weight": "model-00042-of-00061.safetensors", "model.layers.54.mlp.down_proj.weight": "model-00042-of-00061.safetensors", @@ -662,11 +462,7 @@ "model.layers.54.post_attention_layernorm.weight": "model-00042-of-00061.safetensors", "model.layers.54.self_attn.k_proj.weight": "model-00041-of-00061.safetensors", "model.layers.54.self_attn.o_proj.weight": "model-00041-of-00061.safetensors", - "model.layers.54.self_attn.q_proj.lora_A.default.weight": "model-00041-of-00061.safetensors", - "model.layers.54.self_attn.q_proj.lora_B.default.weight": "model-00041-of-00061.safetensors", "model.layers.54.self_attn.q_proj.weight": "model-00041-of-00061.safetensors", - "model.layers.54.self_attn.v_proj.lora_A.default.weight": "model-00041-of-00061.safetensors", - "model.layers.54.self_attn.v_proj.lora_B.default.weight": "model-00041-of-00061.safetensors", "model.layers.54.self_attn.v_proj.weight": "model-00041-of-00061.safetensors", "model.layers.55.input_layernorm.weight": "model-00043-of-00061.safetensors", "model.layers.55.mlp.down_proj.weight": "model-00043-of-00061.safetensors", @@ -675,11 +471,7 @@ "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00061.safetensors", "model.layers.55.self_attn.k_proj.weight": "model-00042-of-00061.safetensors", "model.layers.55.self_attn.o_proj.weight": "model-00042-of-00061.safetensors", - "model.layers.55.self_attn.q_proj.lora_A.default.weight": "model-00042-of-00061.safetensors", - "model.layers.55.self_attn.q_proj.lora_B.default.weight": "model-00042-of-00061.safetensors", "model.layers.55.self_attn.q_proj.weight": "model-00042-of-00061.safetensors", - "model.layers.55.self_attn.v_proj.lora_A.default.weight": "model-00042-of-00061.safetensors", - "model.layers.55.self_attn.v_proj.lora_B.default.weight": "model-00042-of-00061.safetensors", "model.layers.55.self_attn.v_proj.weight": "model-00042-of-00061.safetensors", "model.layers.56.input_layernorm.weight": "model-00043-of-00061.safetensors", "model.layers.56.mlp.down_proj.weight": "model-00043-of-00061.safetensors", @@ -688,11 +480,7 @@ "model.layers.56.post_attention_layernorm.weight": "model-00043-of-00061.safetensors", "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00061.safetensors", "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00061.safetensors", - "model.layers.56.self_attn.q_proj.lora_A.default.weight": "model-00043-of-00061.safetensors", - "model.layers.56.self_attn.q_proj.lora_B.default.weight": "model-00043-of-00061.safetensors", "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00061.safetensors", - "model.layers.56.self_attn.v_proj.lora_A.default.weight": "model-00043-of-00061.safetensors", - "model.layers.56.self_attn.v_proj.lora_B.default.weight": "model-00043-of-00061.safetensors", "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00061.safetensors", "model.layers.57.input_layernorm.weight": "model-00044-of-00061.safetensors", "model.layers.57.mlp.down_proj.weight": "model-00044-of-00061.safetensors", @@ -701,11 +489,7 @@ "model.layers.57.post_attention_layernorm.weight": "model-00044-of-00061.safetensors", "model.layers.57.self_attn.k_proj.weight": "model-00043-of-00061.safetensors", "model.layers.57.self_attn.o_proj.weight": "model-00043-of-00061.safetensors", - "model.layers.57.self_attn.q_proj.lora_A.default.weight": "model-00043-of-00061.safetensors", - "model.layers.57.self_attn.q_proj.lora_B.default.weight": "model-00043-of-00061.safetensors", "model.layers.57.self_attn.q_proj.weight": "model-00043-of-00061.safetensors", - "model.layers.57.self_attn.v_proj.lora_A.default.weight": "model-00043-of-00061.safetensors", - "model.layers.57.self_attn.v_proj.lora_B.default.weight": "model-00043-of-00061.safetensors", "model.layers.57.self_attn.v_proj.weight": "model-00043-of-00061.safetensors", "model.layers.58.input_layernorm.weight": "model-00045-of-00061.safetensors", "model.layers.58.mlp.down_proj.weight": "model-00045-of-00061.safetensors", @@ -714,11 +498,7 @@ "model.layers.58.post_attention_layernorm.weight": "model-00045-of-00061.safetensors", "model.layers.58.self_attn.k_proj.weight": "model-00044-of-00061.safetensors", "model.layers.58.self_attn.o_proj.weight": "model-00044-of-00061.safetensors", - "model.layers.58.self_attn.q_proj.lora_A.default.weight": "model-00044-of-00061.safetensors", - "model.layers.58.self_attn.q_proj.lora_B.default.weight": "model-00044-of-00061.safetensors", "model.layers.58.self_attn.q_proj.weight": "model-00044-of-00061.safetensors", - "model.layers.58.self_attn.v_proj.lora_A.default.weight": "model-00044-of-00061.safetensors", - "model.layers.58.self_attn.v_proj.lora_B.default.weight": "model-00044-of-00061.safetensors", "model.layers.58.self_attn.v_proj.weight": "model-00044-of-00061.safetensors", "model.layers.59.input_layernorm.weight": "model-00046-of-00061.safetensors", "model.layers.59.mlp.down_proj.weight": "model-00046-of-00061.safetensors", @@ -727,11 +507,7 @@ "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00061.safetensors", "model.layers.59.self_attn.k_proj.weight": "model-00045-of-00061.safetensors", "model.layers.59.self_attn.o_proj.weight": "model-00045-of-00061.safetensors", - "model.layers.59.self_attn.q_proj.lora_A.default.weight": "model-00045-of-00061.safetensors", - "model.layers.59.self_attn.q_proj.lora_B.default.weight": "model-00045-of-00061.safetensors", "model.layers.59.self_attn.q_proj.weight": "model-00045-of-00061.safetensors", - "model.layers.59.self_attn.v_proj.lora_A.default.weight": "model-00045-of-00061.safetensors", - "model.layers.59.self_attn.v_proj.lora_B.default.weight": "model-00045-of-00061.safetensors", "model.layers.59.self_attn.v_proj.weight": "model-00045-of-00061.safetensors", "model.layers.6.input_layernorm.weight": "model-00006-of-00061.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00006-of-00061.safetensors", @@ -740,11 +516,7 @@ "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00061.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00061.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00061.safetensors", - "model.layers.6.self_attn.q_proj.lora_A.default.weight": "model-00005-of-00061.safetensors", - "model.layers.6.self_attn.q_proj.lora_B.default.weight": "model-00005-of-00061.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00061.safetensors", - "model.layers.6.self_attn.v_proj.lora_A.default.weight": "model-00005-of-00061.safetensors", - "model.layers.6.self_attn.v_proj.lora_B.default.weight": "model-00005-of-00061.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00061.safetensors", "model.layers.60.input_layernorm.weight": "model-00046-of-00061.safetensors", "model.layers.60.mlp.down_proj.weight": "model-00046-of-00061.safetensors", @@ -753,11 +525,7 @@ "model.layers.60.post_attention_layernorm.weight": "model-00046-of-00061.safetensors", "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00061.safetensors", "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00061.safetensors", - "model.layers.60.self_attn.q_proj.lora_A.default.weight": "model-00046-of-00061.safetensors", - "model.layers.60.self_attn.q_proj.lora_B.default.weight": "model-00046-of-00061.safetensors", "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00061.safetensors", - "model.layers.60.self_attn.v_proj.lora_A.default.weight": "model-00046-of-00061.safetensors", - "model.layers.60.self_attn.v_proj.lora_B.default.weight": "model-00046-of-00061.safetensors", "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00061.safetensors", "model.layers.61.input_layernorm.weight": "model-00047-of-00061.safetensors", "model.layers.61.mlp.down_proj.weight": "model-00047-of-00061.safetensors", @@ -766,11 +534,7 @@ "model.layers.61.post_attention_layernorm.weight": "model-00047-of-00061.safetensors", "model.layers.61.self_attn.k_proj.weight": "model-00046-of-00061.safetensors", "model.layers.61.self_attn.o_proj.weight": "model-00046-of-00061.safetensors", - "model.layers.61.self_attn.q_proj.lora_A.default.weight": "model-00046-of-00061.safetensors", - "model.layers.61.self_attn.q_proj.lora_B.default.weight": "model-00046-of-00061.safetensors", "model.layers.61.self_attn.q_proj.weight": "model-00046-of-00061.safetensors", - "model.layers.61.self_attn.v_proj.lora_A.default.weight": "model-00046-of-00061.safetensors", - "model.layers.61.self_attn.v_proj.lora_B.default.weight": "model-00046-of-00061.safetensors", "model.layers.61.self_attn.v_proj.weight": "model-00046-of-00061.safetensors", "model.layers.62.input_layernorm.weight": "model-00048-of-00061.safetensors", "model.layers.62.mlp.down_proj.weight": "model-00048-of-00061.safetensors", @@ -779,11 +543,7 @@ "model.layers.62.post_attention_layernorm.weight": "model-00048-of-00061.safetensors", "model.layers.62.self_attn.k_proj.weight": "model-00047-of-00061.safetensors", "model.layers.62.self_attn.o_proj.weight": "model-00047-of-00061.safetensors", - "model.layers.62.self_attn.q_proj.lora_A.default.weight": "model-00047-of-00061.safetensors", - "model.layers.62.self_attn.q_proj.lora_B.default.weight": "model-00047-of-00061.safetensors", "model.layers.62.self_attn.q_proj.weight": "model-00047-of-00061.safetensors", - "model.layers.62.self_attn.v_proj.lora_A.default.weight": "model-00047-of-00061.safetensors", - "model.layers.62.self_attn.v_proj.lora_B.default.weight": "model-00047-of-00061.safetensors", "model.layers.62.self_attn.v_proj.weight": "model-00047-of-00061.safetensors", "model.layers.63.input_layernorm.weight": "model-00049-of-00061.safetensors", "model.layers.63.mlp.down_proj.weight": "model-00049-of-00061.safetensors", @@ -792,11 +552,7 @@ "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00061.safetensors", "model.layers.63.self_attn.k_proj.weight": "model-00048-of-00061.safetensors", "model.layers.63.self_attn.o_proj.weight": "model-00048-of-00061.safetensors", - "model.layers.63.self_attn.q_proj.lora_A.default.weight": "model-00048-of-00061.safetensors", - "model.layers.63.self_attn.q_proj.lora_B.default.weight": "model-00048-of-00061.safetensors", "model.layers.63.self_attn.q_proj.weight": "model-00048-of-00061.safetensors", - "model.layers.63.self_attn.v_proj.lora_A.default.weight": "model-00048-of-00061.safetensors", - "model.layers.63.self_attn.v_proj.lora_B.default.weight": "model-00048-of-00061.safetensors", "model.layers.63.self_attn.v_proj.weight": "model-00048-of-00061.safetensors", "model.layers.64.input_layernorm.weight": "model-00049-of-00061.safetensors", "model.layers.64.mlp.down_proj.weight": "model-00049-of-00061.safetensors", @@ -805,11 +561,7 @@ "model.layers.64.post_attention_layernorm.weight": "model-00049-of-00061.safetensors", "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00061.safetensors", "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00061.safetensors", - "model.layers.64.self_attn.q_proj.lora_A.default.weight": "model-00049-of-00061.safetensors", - "model.layers.64.self_attn.q_proj.lora_B.default.weight": "model-00049-of-00061.safetensors", "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00061.safetensors", - "model.layers.64.self_attn.v_proj.lora_A.default.weight": "model-00049-of-00061.safetensors", - "model.layers.64.self_attn.v_proj.lora_B.default.weight": "model-00049-of-00061.safetensors", "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00061.safetensors", "model.layers.65.input_layernorm.weight": "model-00050-of-00061.safetensors", "model.layers.65.mlp.down_proj.weight": "model-00050-of-00061.safetensors", @@ -818,11 +570,7 @@ "model.layers.65.post_attention_layernorm.weight": "model-00050-of-00061.safetensors", "model.layers.65.self_attn.k_proj.weight": "model-00049-of-00061.safetensors", "model.layers.65.self_attn.o_proj.weight": "model-00049-of-00061.safetensors", - "model.layers.65.self_attn.q_proj.lora_A.default.weight": "model-00049-of-00061.safetensors", - "model.layers.65.self_attn.q_proj.lora_B.default.weight": "model-00049-of-00061.safetensors", "model.layers.65.self_attn.q_proj.weight": "model-00049-of-00061.safetensors", - "model.layers.65.self_attn.v_proj.lora_A.default.weight": "model-00049-of-00061.safetensors", - "model.layers.65.self_attn.v_proj.lora_B.default.weight": "model-00049-of-00061.safetensors", "model.layers.65.self_attn.v_proj.weight": "model-00049-of-00061.safetensors", "model.layers.66.input_layernorm.weight": "model-00051-of-00061.safetensors", "model.layers.66.mlp.down_proj.weight": "model-00051-of-00061.safetensors", @@ -831,11 +579,7 @@ "model.layers.66.post_attention_layernorm.weight": "model-00051-of-00061.safetensors", "model.layers.66.self_attn.k_proj.weight": "model-00050-of-00061.safetensors", "model.layers.66.self_attn.o_proj.weight": "model-00050-of-00061.safetensors", - "model.layers.66.self_attn.q_proj.lora_A.default.weight": "model-00050-of-00061.safetensors", - "model.layers.66.self_attn.q_proj.lora_B.default.weight": "model-00050-of-00061.safetensors", "model.layers.66.self_attn.q_proj.weight": "model-00050-of-00061.safetensors", - "model.layers.66.self_attn.v_proj.lora_A.default.weight": "model-00050-of-00061.safetensors", - "model.layers.66.self_attn.v_proj.lora_B.default.weight": "model-00050-of-00061.safetensors", "model.layers.66.self_attn.v_proj.weight": "model-00050-of-00061.safetensors", "model.layers.67.input_layernorm.weight": "model-00052-of-00061.safetensors", "model.layers.67.mlp.down_proj.weight": "model-00052-of-00061.safetensors", @@ -844,11 +588,7 @@ "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00061.safetensors", "model.layers.67.self_attn.k_proj.weight": "model-00051-of-00061.safetensors", "model.layers.67.self_attn.o_proj.weight": "model-00051-of-00061.safetensors", - "model.layers.67.self_attn.q_proj.lora_A.default.weight": "model-00051-of-00061.safetensors", - "model.layers.67.self_attn.q_proj.lora_B.default.weight": "model-00051-of-00061.safetensors", "model.layers.67.self_attn.q_proj.weight": "model-00051-of-00061.safetensors", - "model.layers.67.self_attn.v_proj.lora_A.default.weight": "model-00051-of-00061.safetensors", - "model.layers.67.self_attn.v_proj.lora_B.default.weight": "model-00051-of-00061.safetensors", "model.layers.67.self_attn.v_proj.weight": "model-00051-of-00061.safetensors", "model.layers.68.input_layernorm.weight": "model-00052-of-00061.safetensors", "model.layers.68.mlp.down_proj.weight": "model-00052-of-00061.safetensors", @@ -857,11 +597,7 @@ "model.layers.68.post_attention_layernorm.weight": "model-00052-of-00061.safetensors", "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00061.safetensors", "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00061.safetensors", - "model.layers.68.self_attn.q_proj.lora_A.default.weight": "model-00052-of-00061.safetensors", - "model.layers.68.self_attn.q_proj.lora_B.default.weight": "model-00052-of-00061.safetensors", "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00061.safetensors", - "model.layers.68.self_attn.v_proj.lora_A.default.weight": "model-00052-of-00061.safetensors", - "model.layers.68.self_attn.v_proj.lora_B.default.weight": "model-00052-of-00061.safetensors", "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00061.safetensors", "model.layers.69.input_layernorm.weight": "model-00053-of-00061.safetensors", "model.layers.69.mlp.down_proj.weight": "model-00053-of-00061.safetensors", @@ -870,11 +606,7 @@ "model.layers.69.post_attention_layernorm.weight": "model-00053-of-00061.safetensors", "model.layers.69.self_attn.k_proj.weight": "model-00052-of-00061.safetensors", "model.layers.69.self_attn.o_proj.weight": "model-00052-of-00061.safetensors", - "model.layers.69.self_attn.q_proj.lora_A.default.weight": "model-00052-of-00061.safetensors", - "model.layers.69.self_attn.q_proj.lora_B.default.weight": "model-00052-of-00061.safetensors", "model.layers.69.self_attn.q_proj.weight": "model-00052-of-00061.safetensors", - "model.layers.69.self_attn.v_proj.lora_A.default.weight": "model-00052-of-00061.safetensors", - "model.layers.69.self_attn.v_proj.lora_B.default.weight": "model-00052-of-00061.safetensors", "model.layers.69.self_attn.v_proj.weight": "model-00052-of-00061.safetensors", "model.layers.7.input_layernorm.weight": "model-00007-of-00061.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00007-of-00061.safetensors", @@ -883,11 +615,7 @@ "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00061.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00061.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00061.safetensors", - "model.layers.7.self_attn.q_proj.lora_A.default.weight": "model-00006-of-00061.safetensors", - "model.layers.7.self_attn.q_proj.lora_B.default.weight": "model-00006-of-00061.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00061.safetensors", - "model.layers.7.self_attn.v_proj.lora_A.default.weight": "model-00006-of-00061.safetensors", - "model.layers.7.self_attn.v_proj.lora_B.default.weight": "model-00006-of-00061.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00061.safetensors", "model.layers.70.input_layernorm.weight": "model-00054-of-00061.safetensors", "model.layers.70.mlp.down_proj.weight": "model-00054-of-00061.safetensors", @@ -896,11 +624,7 @@ "model.layers.70.post_attention_layernorm.weight": "model-00054-of-00061.safetensors", "model.layers.70.self_attn.k_proj.weight": "model-00053-of-00061.safetensors", "model.layers.70.self_attn.o_proj.weight": "model-00053-of-00061.safetensors", - "model.layers.70.self_attn.q_proj.lora_A.default.weight": "model-00053-of-00061.safetensors", - "model.layers.70.self_attn.q_proj.lora_B.default.weight": "model-00053-of-00061.safetensors", "model.layers.70.self_attn.q_proj.weight": "model-00053-of-00061.safetensors", - "model.layers.70.self_attn.v_proj.lora_A.default.weight": "model-00053-of-00061.safetensors", - "model.layers.70.self_attn.v_proj.lora_B.default.weight": "model-00053-of-00061.safetensors", "model.layers.70.self_attn.v_proj.weight": "model-00053-of-00061.safetensors", "model.layers.71.input_layernorm.weight": "model-00055-of-00061.safetensors", "model.layers.71.mlp.down_proj.weight": "model-00055-of-00061.safetensors", @@ -909,11 +633,7 @@ "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00061.safetensors", "model.layers.71.self_attn.k_proj.weight": "model-00054-of-00061.safetensors", "model.layers.71.self_attn.o_proj.weight": "model-00054-of-00061.safetensors", - "model.layers.71.self_attn.q_proj.lora_A.default.weight": "model-00054-of-00061.safetensors", - "model.layers.71.self_attn.q_proj.lora_B.default.weight": "model-00054-of-00061.safetensors", "model.layers.71.self_attn.q_proj.weight": "model-00054-of-00061.safetensors", - "model.layers.71.self_attn.v_proj.lora_A.default.weight": "model-00054-of-00061.safetensors", - "model.layers.71.self_attn.v_proj.lora_B.default.weight": "model-00054-of-00061.safetensors", "model.layers.71.self_attn.v_proj.weight": "model-00054-of-00061.safetensors", "model.layers.72.input_layernorm.weight": "model-00055-of-00061.safetensors", "model.layers.72.mlp.down_proj.weight": "model-00055-of-00061.safetensors", @@ -922,11 +642,7 @@ "model.layers.72.post_attention_layernorm.weight": "model-00055-of-00061.safetensors", "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00061.safetensors", "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00061.safetensors", - "model.layers.72.self_attn.q_proj.lora_A.default.weight": "model-00055-of-00061.safetensors", - "model.layers.72.self_attn.q_proj.lora_B.default.weight": "model-00055-of-00061.safetensors", "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00061.safetensors", - "model.layers.72.self_attn.v_proj.lora_A.default.weight": "model-00055-of-00061.safetensors", - "model.layers.72.self_attn.v_proj.lora_B.default.weight": "model-00055-of-00061.safetensors", "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00061.safetensors", "model.layers.73.input_layernorm.weight": "model-00056-of-00061.safetensors", "model.layers.73.mlp.down_proj.weight": "model-00056-of-00061.safetensors", @@ -935,11 +651,7 @@ "model.layers.73.post_attention_layernorm.weight": "model-00056-of-00061.safetensors", "model.layers.73.self_attn.k_proj.weight": "model-00055-of-00061.safetensors", "model.layers.73.self_attn.o_proj.weight": "model-00055-of-00061.safetensors", - "model.layers.73.self_attn.q_proj.lora_A.default.weight": "model-00055-of-00061.safetensors", - "model.layers.73.self_attn.q_proj.lora_B.default.weight": "model-00055-of-00061.safetensors", "model.layers.73.self_attn.q_proj.weight": "model-00055-of-00061.safetensors", - "model.layers.73.self_attn.v_proj.lora_A.default.weight": "model-00055-of-00061.safetensors", - "model.layers.73.self_attn.v_proj.lora_B.default.weight": "model-00055-of-00061.safetensors", "model.layers.73.self_attn.v_proj.weight": "model-00055-of-00061.safetensors", "model.layers.74.input_layernorm.weight": "model-00057-of-00061.safetensors", "model.layers.74.mlp.down_proj.weight": "model-00057-of-00061.safetensors", @@ -948,11 +660,7 @@ "model.layers.74.post_attention_layernorm.weight": "model-00057-of-00061.safetensors", "model.layers.74.self_attn.k_proj.weight": "model-00056-of-00061.safetensors", "model.layers.74.self_attn.o_proj.weight": "model-00056-of-00061.safetensors", - "model.layers.74.self_attn.q_proj.lora_A.default.weight": "model-00056-of-00061.safetensors", - "model.layers.74.self_attn.q_proj.lora_B.default.weight": "model-00056-of-00061.safetensors", "model.layers.74.self_attn.q_proj.weight": "model-00056-of-00061.safetensors", - "model.layers.74.self_attn.v_proj.lora_A.default.weight": "model-00056-of-00061.safetensors", - "model.layers.74.self_attn.v_proj.lora_B.default.weight": "model-00056-of-00061.safetensors", "model.layers.74.self_attn.v_proj.weight": "model-00056-of-00061.safetensors", "model.layers.75.input_layernorm.weight": "model-00058-of-00061.safetensors", "model.layers.75.mlp.down_proj.weight": "model-00058-of-00061.safetensors", @@ -961,11 +669,7 @@ "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00061.safetensors", "model.layers.75.self_attn.k_proj.weight": "model-00057-of-00061.safetensors", "model.layers.75.self_attn.o_proj.weight": "model-00057-of-00061.safetensors", - "model.layers.75.self_attn.q_proj.lora_A.default.weight": "model-00057-of-00061.safetensors", - "model.layers.75.self_attn.q_proj.lora_B.default.weight": "model-00057-of-00061.safetensors", "model.layers.75.self_attn.q_proj.weight": "model-00057-of-00061.safetensors", - "model.layers.75.self_attn.v_proj.lora_A.default.weight": "model-00057-of-00061.safetensors", - "model.layers.75.self_attn.v_proj.lora_B.default.weight": "model-00057-of-00061.safetensors", "model.layers.75.self_attn.v_proj.weight": "model-00057-of-00061.safetensors", "model.layers.76.input_layernorm.weight": "model-00058-of-00061.safetensors", "model.layers.76.mlp.down_proj.weight": "model-00058-of-00061.safetensors", @@ -974,11 +678,7 @@ "model.layers.76.post_attention_layernorm.weight": "model-00058-of-00061.safetensors", "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00061.safetensors", "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00061.safetensors", - "model.layers.76.self_attn.q_proj.lora_A.default.weight": "model-00058-of-00061.safetensors", - "model.layers.76.self_attn.q_proj.lora_B.default.weight": "model-00058-of-00061.safetensors", "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00061.safetensors", - "model.layers.76.self_attn.v_proj.lora_A.default.weight": "model-00058-of-00061.safetensors", - "model.layers.76.self_attn.v_proj.lora_B.default.weight": "model-00058-of-00061.safetensors", "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00061.safetensors", "model.layers.77.input_layernorm.weight": "model-00059-of-00061.safetensors", "model.layers.77.mlp.down_proj.weight": "model-00059-of-00061.safetensors", @@ -987,11 +687,7 @@ "model.layers.77.post_attention_layernorm.weight": "model-00059-of-00061.safetensors", "model.layers.77.self_attn.k_proj.weight": "model-00058-of-00061.safetensors", "model.layers.77.self_attn.o_proj.weight": "model-00058-of-00061.safetensors", - "model.layers.77.self_attn.q_proj.lora_A.default.weight": "model-00058-of-00061.safetensors", - "model.layers.77.self_attn.q_proj.lora_B.default.weight": "model-00058-of-00061.safetensors", "model.layers.77.self_attn.q_proj.weight": "model-00058-of-00061.safetensors", - "model.layers.77.self_attn.v_proj.lora_A.default.weight": "model-00058-of-00061.safetensors", - "model.layers.77.self_attn.v_proj.lora_B.default.weight": "model-00058-of-00061.safetensors", "model.layers.77.self_attn.v_proj.weight": "model-00058-of-00061.safetensors", "model.layers.78.input_layernorm.weight": "model-00060-of-00061.safetensors", "model.layers.78.mlp.down_proj.weight": "model-00060-of-00061.safetensors", @@ -1000,11 +696,7 @@ "model.layers.78.post_attention_layernorm.weight": "model-00060-of-00061.safetensors", "model.layers.78.self_attn.k_proj.weight": "model-00059-of-00061.safetensors", "model.layers.78.self_attn.o_proj.weight": "model-00059-of-00061.safetensors", - "model.layers.78.self_attn.q_proj.lora_A.default.weight": "model-00059-of-00061.safetensors", - "model.layers.78.self_attn.q_proj.lora_B.default.weight": "model-00059-of-00061.safetensors", "model.layers.78.self_attn.q_proj.weight": "model-00059-of-00061.safetensors", - "model.layers.78.self_attn.v_proj.lora_A.default.weight": "model-00059-of-00061.safetensors", - "model.layers.78.self_attn.v_proj.lora_B.default.weight": "model-00059-of-00061.safetensors", "model.layers.78.self_attn.v_proj.weight": "model-00059-of-00061.safetensors", "model.layers.79.input_layernorm.weight": "model-00061-of-00061.safetensors", "model.layers.79.mlp.down_proj.weight": "model-00061-of-00061.safetensors", @@ -1013,11 +705,7 @@ "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00061.safetensors", "model.layers.79.self_attn.k_proj.weight": "model-00060-of-00061.safetensors", "model.layers.79.self_attn.o_proj.weight": "model-00060-of-00061.safetensors", - "model.layers.79.self_attn.q_proj.lora_A.default.weight": "model-00060-of-00061.safetensors", - "model.layers.79.self_attn.q_proj.lora_B.default.weight": "model-00060-of-00061.safetensors", "model.layers.79.self_attn.q_proj.weight": "model-00060-of-00061.safetensors", - "model.layers.79.self_attn.v_proj.lora_A.default.weight": "model-00060-of-00061.safetensors", - "model.layers.79.self_attn.v_proj.lora_B.default.weight": "model-00060-of-00061.safetensors", "model.layers.79.self_attn.v_proj.weight": "model-00060-of-00061.safetensors", "model.layers.8.input_layernorm.weight": "model-00007-of-00061.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00007-of-00061.safetensors", @@ -1026,11 +714,7 @@ "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00061.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00061.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00061.safetensors", - "model.layers.8.self_attn.q_proj.lora_A.default.weight": "model-00007-of-00061.safetensors", - "model.layers.8.self_attn.q_proj.lora_B.default.weight": "model-00007-of-00061.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00061.safetensors", - "model.layers.8.self_attn.v_proj.lora_A.default.weight": "model-00007-of-00061.safetensors", - "model.layers.8.self_attn.v_proj.lora_B.default.weight": "model-00007-of-00061.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00061.safetensors", "model.layers.9.input_layernorm.weight": "model-00008-of-00061.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00008-of-00061.safetensors", @@ -1039,11 +723,7 @@ "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00061.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00061.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00061.safetensors", - "model.layers.9.self_attn.q_proj.lora_A.default.weight": "model-00007-of-00061.safetensors", - "model.layers.9.self_attn.q_proj.lora_B.default.weight": "model-00007-of-00061.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00061.safetensors", - "model.layers.9.self_attn.v_proj.lora_A.default.weight": "model-00007-of-00061.safetensors", - "model.layers.9.self_attn.v_proj.lora_B.default.weight": "model-00007-of-00061.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00061.safetensors", "model.norm.weight": "model-00061-of-00061.safetensors" } diff --git a/special_tokens_map.json b/special_tokens_map.json index 8cd5f1eb30d4e97d74cbf915c36db116aea5eca7..72ecfeeb7e14d244c936169d2ed139eeae235ef1 100644 --- a/special_tokens_map.json +++ b/special_tokens_map.json @@ -1,11 +1,24 @@ { - "additional_special_tokens": [ - "", - "", - "" - ], - "bos_token": "", - "eos_token": "", + "bos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, "pad_token": "", - "unk_token": "" + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } } diff --git a/tokenizer_config.json b/tokenizer_config.json index 7e109ec5ff993eb919d1777f22b937921f769c3d..1fad71d2f9e4cd97bae83a238028f2973c5fb87a 100644 --- a/tokenizer_config.json +++ b/tokenizer_config.json @@ -1,4 +1,6 @@ { + "add_bos_token": true, + "add_eos_token": false, "added_tokens_decoder": { "0": { "content": "", @@ -25,24 +27,14 @@ "special": true } }, - "additional_special_tokens": [ - "", - "", - "" - ], "bos_token": "", "clean_up_tokenization_spaces": false, "eos_token": "", "legacy": false, - "max_length": 2048, "model_max_length": 1000000000000000019884624838656, "pad_token": "", - "padding_side": "right", "sp_model_kwargs": {}, - "stride": 0, "tokenizer_class": "LlamaTokenizer", - "truncation_side": "right", - "truncation_strategy": "longest_first", "unk_token": "", "use_default_system_prompt": true }