diff --git a/README.md b/README.md index 13ec9c9c3ac7b018da39c01ee6162244ff81f1bc..18a1225031f7ff11381c60dc79fc0a3ba8c89b0e 100644 --- a/README.md +++ b/README.md @@ -25,11 +25,11 @@ Shining Valiant is a chat model built on the Llama 2 architecture, finetuned on ## Version -The current version is **1.2**. +The current version is **1.3!** -**Version 1.3** is now **being validated for release!** +We're thrilled to bring you our newest release! -Previous versions remain available in the repository. New models will be released for everyone once our team's training and validation process is complete :) +Previous versions remain available in the repository. New models will be released for everyone once our team's training and validation process is complete. ## Evaluation @@ -39,6 +39,10 @@ Previous versions remain available in the repository. New models will be release | Llama 2 | 67.35 | 67.32 | 87.33 | 69.83 | 44.92 | | Llama 2 Chat | 66.80 | 64.59 | 85.88 | 63.91 | 52.80 | +**Shining Valiant 1.3** is awaiting full results from the Open LLM Leaderboard. + +SV 1.3 outperformed SV 1.2 on our internal testing. + ## Prompting Guide Shining Valiant uses the same prompt format as Llama 2 Chat - feel free to use your existing prompts and scripts! A few examples of different formats: @@ -50,7 +54,7 @@ A few examples of different formats: 3. [INST] << SYS >>You are an intelligent, helpful AI assistant.<< /SYS >>Deep dive about a country with interesting history: [/INST] ## The Model -Shining Valiant is built on top of Stellar Bright, which uses Llama 2's 70b parameter architecture and features upgraded general capability. (Stellar Bright uses public open source data only.) +Shining Valiant is built on top of Sunset Boulevard, which uses Llama 2's 70b parameter architecture and features upgraded general capability. From there, we've created Shining Valiant through multiple finetuning runs on different compositions of our private dataset. diff --git a/model-00001-of-00061.safetensors b/model-00001-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..27bf0febefbc748eae84eb33fb4bb2833e397a9c --- /dev/null +++ b/model-00001-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dc4ed54e736c5d13b2ced1c4c5d96d36854f823c2a80316c08674ac4dc12905 +size 4829678080 diff --git a/model-00002-of-00061.safetensors b/model-00002-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7aa534e8420f9a3b35d171eb362a1ff2739935cc --- /dev/null +++ b/model-00002-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46b3af357e97b16ff83a04c91b76957b58b7b4c398017c3fe23b7852a5fa973c +size 4642047760 diff --git a/model-00003-of-00061.safetensors b/model-00003-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7dfddaadd184815e24dc841b83d7583f68b57a84 --- /dev/null +++ b/model-00003-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c30362d287354bbfb8a03a1ac1f00f01d113388cb2ad5654dd1b5861820450b +size 4373612184 diff --git a/model-00004-of-00061.safetensors b/model-00004-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0175a1f997d883a721cb2d25905b6d2616c2d6c4 --- /dev/null +++ b/model-00004-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f6ec8050a5cec810bf98716a50c48587ea240f1a8ae1037587c4a069c0c5346 +size 4989127504 diff --git a/model-00005-of-00061.safetensors b/model-00005-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b2754ce80e0506532d16e09ea9d7b4e7386f2b7b --- /dev/null +++ b/model-00005-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f798e0c23342f96a09bbac844334d8542dece23627168c01819aefd8db1dbae4 +size 4373612184 diff --git a/model-00006-of-00061.safetensors b/model-00006-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a4c97584f42de111848e03d838bec33a3e1b435a --- /dev/null +++ b/model-00006-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77d2a7b76a95c7fda85e63b262e3d00cb0178306aaff700cea4603dd55f7977a +size 4373612184 diff --git a/model-00007-of-00061.safetensors b/model-00007-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..53011d63dc16f956baeccd8584538d60da4073b2 --- /dev/null +++ b/model-00007-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2b621b2600196832c8e85a4bf68d111d03159b0c08e9731dfd4f45df4ab8c10 +size 4989127504 diff --git a/model-00008-of-00061.safetensors b/model-00008-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bc08a8827586155fce7c209332e6e775d2322f2c --- /dev/null +++ b/model-00008-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc3bce3873e52561fdd0d3d3594d9be7e02256d60ace42be717aba51b896e096 +size 4373612200 diff --git a/model-00009-of-00061.safetensors b/model-00009-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..397b0ab4052859baa6c4e82fd58799bb74d5ba2c --- /dev/null +++ b/model-00009-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1f7c5ef4e3f7979037b25a5c6aa8e0824ec80793d7be37a4586bbab81036c40 +size 4373612200 diff --git a/model-00010-of-00061.safetensors b/model-00010-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dd68cd01cc2baefb048923b5566ef31e991d17b0 --- /dev/null +++ b/model-00010-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eaf35b1f54fea6f42dad09ae0c359a7cd8c3a6e937ed0677536f5cde9a04f322 +size 4989127528 diff --git a/model-00011-of-00061.safetensors b/model-00011-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c405a1fbd83e8d5482139815efea04fa3fc8738e --- /dev/null +++ b/model-00011-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5010617f685d5c77ac291540a3e746b585264f54dfd84e715e572eed2dbd76d6 +size 4373612200 diff --git a/model-00012-of-00061.safetensors b/model-00012-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ca11d39d6614082c63db03d30523658cc42b0e37 --- /dev/null +++ b/model-00012-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5247cb589001678aa6fcabe321eaff2b99ae98c56e940296b8ee900d7dbc0a7 +size 4373612200 diff --git a/model-00013-of-00061.safetensors b/model-00013-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..552ea748c20d97e8f67389c34f7e1e2b852d4b1a --- /dev/null +++ b/model-00013-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53b7f85259c4576142f2a8700fbc12809758340b59f16a649d8ace3093fc3d9f +size 4989127528 diff --git a/model-00014-of-00061.safetensors b/model-00014-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..421305f621134a2e289c4c4afe6c955e709711af --- /dev/null +++ b/model-00014-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94d96ea5b1bd5ea70fefdb276b5071e2822dc3b56db3211ec8397d701d8230b4 +size 4373612200 diff --git a/model-00015-of-00061.safetensors b/model-00015-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2edcbe9641c6b951410bb73e4567bea29e4a9c4c --- /dev/null +++ b/model-00015-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:894d89e5b9b97d924101027848e58c0ac9f79b4436eff5dd9de45d37d4152cdc +size 4373612200 diff --git a/model-00016-of-00061.safetensors b/model-00016-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9fdcf4be9ff1a89008717d5e992cb6d7b16cc9b5 --- /dev/null +++ b/model-00016-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46d0b2a79ab34a91fa8b8f279ff88cf24d84a96fce7fe8cd98f8357701205d5f +size 4989127528 diff --git a/model-00017-of-00061.safetensors b/model-00017-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fe80876611320bc1f5306d4e62ac3a6cbc304d9c --- /dev/null +++ b/model-00017-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3bf163933cbe771e0836a82aa38473da7a1803fb059e4cb04e69504d29862ad +size 4373612200 diff --git a/model-00018-of-00061.safetensors b/model-00018-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d0bcfae60eb0c093eef5e9d36ab11cecf10ba536 --- /dev/null +++ b/model-00018-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9d9390a0c9dd22dc16b1e1ed30a0f8595d8f12bf52a498b2b29a6c62b6467e1 +size 4373612200 diff --git a/model-00019-of-00061.safetensors b/model-00019-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..914171cbdc0e4950de9c439c82514901ed420742 --- /dev/null +++ b/model-00019-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4529d07c0e6dcdf8a77a2627d117ee922abeee618069569451849ecc7aa339d +size 4989127528 diff --git a/model-00020-of-00061.safetensors b/model-00020-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7603cbe0d5149ea3bc3fa144e4e26a582963e64a --- /dev/null +++ b/model-00020-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3bd585996c4d5af5587b7a807bdf2fd2170981f041db89e8b2923469ec940d9d +size 4373612200 diff --git a/model-00021-of-00061.safetensors b/model-00021-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..39cfd69e53624e181a11ffd465c4f37d1b1e52fa --- /dev/null +++ b/model-00021-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30854cb22ca8777646ba273e0591f4cf1690e4078c9298077c50cf0e2dde583c +size 4373612200 diff --git a/model-00022-of-00061.safetensors b/model-00022-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..88108163b201cd44db1ee274b0b869bfdfcf4020 --- /dev/null +++ b/model-00022-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:651b0e0d6d0726aa720f4ce68c67bd4b47a74b22d370590510ddb6348c69ab5a +size 4989127528 diff --git a/model-00023-of-00061.safetensors b/model-00023-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..995cb0e735007a720cdcadef9f66e114a1fe373b --- /dev/null +++ b/model-00023-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e4fd8b76d520900d420b1dd619c17d6dbca9db12c3f9dfc543811c3613ae08b +size 4373612200 diff --git a/model-00024-of-00061.safetensors b/model-00024-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bb64f46de9ac6e92f7eeca3ade47f95a0c4b7d54 --- /dev/null +++ b/model-00024-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf67fc00883bbe333f29a5fe93a481f9ac8c548705a0b96b79f1b2072af2e391 +size 4373612200 diff --git a/model-00025-of-00061.safetensors b/model-00025-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b48e6cb3f1a8d109d7088d052c89cbda7347ed05 --- /dev/null +++ b/model-00025-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:383ad969a694badc469b5d313ccffb384c25e5a9280a69772aca961968ced11a +size 4989127528 diff --git a/model-00026-of-00061.safetensors b/model-00026-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c6e781b9a3740c89103c96f05fcc0659c0f00574 --- /dev/null +++ b/model-00026-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a305dad26ba99db15c0d788f0703fb8d8652289b9a1d83f906eabc04fa2a1965 +size 4373612200 diff --git a/model-00027-of-00061.safetensors b/model-00027-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ebc325bc6862a0d89542b3c198724472042049a3 --- /dev/null +++ b/model-00027-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a78b51174496d7caf102675247fd4cc7d25c6b7c08bb10eea6e913b812ec42fe +size 4373612200 diff --git a/model-00028-of-00061.safetensors b/model-00028-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..75db7d3c699bb2bbe5a7f6873cde91dae8cf35fe --- /dev/null +++ b/model-00028-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80c35fafdad91444fd97f5b7c18b0eb704bf1a8d1e60508ec8fa4d365ad7202d +size 4989127528 diff --git a/model-00029-of-00061.safetensors b/model-00029-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..42764f0af40b011edebf42ef112993c6d163252c --- /dev/null +++ b/model-00029-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c0633b845bccf0924cdf6f723b42b89cb3a48eda645e1d57452300037911524 +size 4373612200 diff --git a/model-00030-of-00061.safetensors b/model-00030-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2f433bf7682682fe8f89fe5fb5686149af7f9c8c --- /dev/null +++ b/model-00030-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e946e96b50482a20eb7df6d2584a7bc6f23471e4d7890111fd01b92812d1aa1f +size 4373612200 diff --git a/model-00031-of-00061.safetensors b/model-00031-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7e3f69b4d66c67c2080fb0af2606e01b66a8a1fb --- /dev/null +++ b/model-00031-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5543aa4273bf925573330575ec83dd17483a6c5bbf5864cc9429ca6e61729fa0 +size 4989127528 diff --git a/model-00032-of-00061.safetensors b/model-00032-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..325a1b15b4e5d77f7901aa2f2f186c681c4f84b3 --- /dev/null +++ b/model-00032-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2a1adee9e1983d93ddd08d8fc5b0d2cc2a701f441afe7a018f88414d6b181e7 +size 4373612200 diff --git a/model-00033-of-00061.safetensors b/model-00033-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2e704a79a1558ccb2f645de01d0d6a3333783986 --- /dev/null +++ b/model-00033-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9432a07eea4f92853667c83f60f662b36f6e0fd1adfd92e9e6c7d52f87648d52 +size 4373612200 diff --git a/model-00034-of-00061.safetensors b/model-00034-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f5ab6125ca9b805ce2d4e7e6b91ed10aa73a67d9 --- /dev/null +++ b/model-00034-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b9cc6274a106b89b659f7952497e3958354650e640dc8051376abbf3e3e293b +size 4989127528 diff --git a/model-00035-of-00061.safetensors b/model-00035-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8545b2f5e09122e88b5a3e2f162066b4d29f5e48 --- /dev/null +++ b/model-00035-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f17095616c39cc562f354d348f7a2fa9ca851213831e1b78c0f958ec4e4ab0c +size 4373612200 diff --git a/model-00036-of-00061.safetensors b/model-00036-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eeeb1abfc8c8cd55f8579be0e72a4ed8ed1a705a --- /dev/null +++ b/model-00036-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a006ebb9c9a48e49d021626ff12eb1a81bd7eb41298d6e2633372624e6764ae +size 4373612200 diff --git a/model-00037-of-00061.safetensors b/model-00037-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..30c83e554007e7ed1f782c2d6c57e81a7a03769b --- /dev/null +++ b/model-00037-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f0cba34fb68521ca64535aa1b4926c8dc42def86990d88c5ff9ec2f361ef403 +size 4989127528 diff --git a/model-00038-of-00061.safetensors b/model-00038-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..13eca929d487b195a2f3536fea3b765e2d7c1fc7 --- /dev/null +++ b/model-00038-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc3bcfc5feeacb1fd29e55c4e06ff376a41706b1274dfb34fb6ef388ed36da5e +size 4373612200 diff --git a/model-00039-of-00061.safetensors b/model-00039-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..daf3fa2854e1d84c1d2b351476dee913012d4406 --- /dev/null +++ b/model-00039-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52dd5434217e8956e78be5bb6f8762809a565812675e8c2a72077fccaaeadf33 +size 4373612200 diff --git a/model-00040-of-00061.safetensors b/model-00040-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..19be479b5c41da1610355f768b9a3cccd58ecfe0 --- /dev/null +++ b/model-00040-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81a3f6dfe3126fa87cb4911adcf1f78bf879d529a328c53a1cde2e3d4dea241e +size 4989127528 diff --git a/model-00041-of-00061.safetensors b/model-00041-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7cdcee7f1a974f2b89d9a6e738ffe2bb1708c1c7 --- /dev/null +++ b/model-00041-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:546a8fab65a3d70e826ef3255138a0fd5d9bff51ef6c146c01fcb0227a4d2153 +size 4373612200 diff --git a/model-00042-of-00061.safetensors b/model-00042-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..261fce9d8e002a80f38a79cc9f9d694495f597d3 --- /dev/null +++ b/model-00042-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fdb73cb8ebdef799f5d331b44e6246a68b38908f50d9dc4ef3a99a98f1ff99ea +size 4373612200 diff --git a/model-00043-of-00061.safetensors b/model-00043-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..57ec7bae2c2ded9c51cafdd8cd2a6f1288767638 --- /dev/null +++ b/model-00043-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b4944c19cbf313aacfa5b8d70a47cd9c052b34adc15be12cf80cb8838b8ca6e +size 4989127528 diff --git a/model-00044-of-00061.safetensors b/model-00044-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2a18c4ea8ca4d17cf55f1b766f87eefe10a6b99f --- /dev/null +++ b/model-00044-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a7799868fac8c86eea07b50046b46545f1631a47411bfec132a04d646a5a0bd +size 4373612200 diff --git a/model-00045-of-00061.safetensors b/model-00045-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..03683a845d16b1646f73fe1a6668b074d3200612 --- /dev/null +++ b/model-00045-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f90c6d73620da6c3adf1312fb9ee60cf86b87e0f8c6ce87e81b8ba46d3fbda70 +size 4373612200 diff --git a/model-00046-of-00061.safetensors b/model-00046-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f66d1b3bbdaae38410ed8bd9f4fb530c2658c2f5 --- /dev/null +++ b/model-00046-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d9a4db2a1fec6341c31afabd7098d2c609a9d8cc7e461f25ad8642d39d48ba0 +size 4989127528 diff --git a/model-00047-of-00061.safetensors b/model-00047-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..920d342c249aeaa7fa1ad578d3b5b15f83f5fac2 --- /dev/null +++ b/model-00047-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fde1a6af79a036be96b7e7c385900bdc295d0f8e20e5d3814aaa907c162152f5 +size 4373612200 diff --git a/model-00048-of-00061.safetensors b/model-00048-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6ee9ca2af5d6ae912eed6206272ffe6fd06e0ed8 --- /dev/null +++ b/model-00048-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce67f6a4d4b9fe331baedc3133abef8a260a60fe37ed8413a379cdd0dd2ae575 +size 4373612200 diff --git a/model-00049-of-00061.safetensors b/model-00049-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..72a97e0f0c626c8d52cda3096dfd23fd9151d209 --- /dev/null +++ b/model-00049-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bfc59ca81c4a8f85dd723149149a3c585bada9c0da7996a36b8ee81fca162e12 +size 4989127528 diff --git a/model-00050-of-00061.safetensors b/model-00050-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6568bd784169eb8477fa0f8da6e71d3a8ed9886e --- /dev/null +++ b/model-00050-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eefff4df3f049d652b9bc9c24a5112107067c2e07a21b891b496ea38d2b7c689 +size 4373612200 diff --git a/model-00051-of-00061.safetensors b/model-00051-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c523205186ca88bbcf2560881979836954e2d44c --- /dev/null +++ b/model-00051-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:012b4bc81e1f439f67920ca9109ccabe10cdab0182c3e687bd389ec6663b1fcd +size 4373612200 diff --git a/model-00052-of-00061.safetensors b/model-00052-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..47e4bbb4a243352cf63d5075168a6246af925c25 --- /dev/null +++ b/model-00052-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5be8a68f8243469be9feeedfce196d133b28d5ee239af2e227c22aeb018e487b +size 4989127528 diff --git a/model-00053-of-00061.safetensors b/model-00053-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fca1fe02748a0cac85d8e95512c0cf32aa6dabd0 --- /dev/null +++ b/model-00053-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9637c9df48a125275b77e534fb5d44ed4587ad484b58224b1d4218035bde7e22 +size 4373612200 diff --git a/model-00054-of-00061.safetensors b/model-00054-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f09c4a8b118c473eac1c1aea18d562285c263470 --- /dev/null +++ b/model-00054-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc5774a350a4022293ec74dea3c132433aa2ac621fa3b1528506034a4567e7de +size 4373612200 diff --git a/model-00055-of-00061.safetensors b/model-00055-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c920f493ad271a24a98cd0a01905c7ff4f9e5e4b --- /dev/null +++ b/model-00055-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3fa07dea54956d3e871a662f62c0681f35fe347a0794dd9ab48af98e2d93f1f +size 4989127528 diff --git a/model-00056-of-00061.safetensors b/model-00056-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bb0e7679fc55709b2dae071fd2fadfca9eb4e806 --- /dev/null +++ b/model-00056-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:878b0eeed0949288ed3e6a52b41e572b324c55ae89709e8a863f605448aad70c +size 4373612200 diff --git a/model-00057-of-00061.safetensors b/model-00057-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7c0f690a12f02b6084e38d239578544c9acc8d9e --- /dev/null +++ b/model-00057-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a1bab1b9611953a58483129712df3d7ffb17d6171f23df3b2ccc4fee67c96b6 +size 4373612200 diff --git a/model-00058-of-00061.safetensors b/model-00058-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..28b7d53315dfce763acee29f40f5891be05835f8 --- /dev/null +++ b/model-00058-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0bdfefa9c2d6b65477995975c1a3e22860152d63b6f51412be6b5b27f664391 +size 4989127528 diff --git a/model-00059-of-00061.safetensors b/model-00059-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c3bab595346d3d5513d10b296b4c35c6488c8003 --- /dev/null +++ b/model-00059-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3e95bfa154015e1b7b6d8d3d0b2849afb131ae960a9eb7e3ca907da47ead523 +size 4373612200 diff --git a/model-00060-of-00061.safetensors b/model-00060-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..28388d212ba8778213311c0dba096f5720c630a0 --- /dev/null +++ b/model-00060-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e5f9c3c6844cdb028e95f33f6871a7d7355eca6a12668174070fa211ee565a0c +size 4373612200 diff --git a/model-00061-of-00061.safetensors b/model-00061-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eed51e430e97073d5a3526ee1619767e1084813d --- /dev/null +++ b/model-00061-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c25429c3da71e1db5ad2b6d7cd64f22675a117f9d9967881bba38219ea7d5068 +size 1988198960 diff --git a/model.safetensors.index.json b/model.safetensors.index.json index 542634137f88065bab104ac6c698759b2377762f..efccb6e9b0e614705696e3583f0201b6fc7c1f8a 100644 --- a/model.safetensors.index.json +++ b/model.safetensors.index.json @@ -1,1050 +1,1050 @@ { "metadata": { - "total_size": 276955168768 + "total_size": 276824096768 }, "weight_map": { - "lm_head.weight": "model-00030-of-00030.safetensors", - "model.embed_tokens.weight": "model-00001-of-00030.safetensors", - "model.layers.0.input_layernorm.weight": "model-00001-of-00030.safetensors", - "model.layers.0.mlp.down_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.0.mlp.up_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00030.safetensors", - "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.0.self_attn.q_proj.lora_A.default.weight": "model-00001-of-00030.safetensors", - "model.layers.0.self_attn.q_proj.lora_B.default.weight": "model-00001-of-00030.safetensors", - "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.0.self_attn.v_proj.lora_A.default.weight": "model-00001-of-00030.safetensors", - "model.layers.0.self_attn.v_proj.lora_B.default.weight": "model-00001-of-00030.safetensors", - "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.1.input_layernorm.weight": "model-00001-of-00030.safetensors", - "model.layers.1.mlp.down_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.1.mlp.up_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00030.safetensors", - "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.1.self_attn.q_proj.lora_A.default.weight": "model-00001-of-00030.safetensors", - "model.layers.1.self_attn.q_proj.lora_B.default.weight": "model-00001-of-00030.safetensors", - "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.1.self_attn.v_proj.lora_A.default.weight": "model-00001-of-00030.safetensors", - "model.layers.1.self_attn.v_proj.lora_B.default.weight": "model-00001-of-00030.safetensors", - "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.10.input_layernorm.weight": "model-00005-of-00030.safetensors", - "model.layers.10.mlp.down_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.10.mlp.gate_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.10.mlp.up_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.10.post_attention_layernorm.weight": "model-00005-of-00030.safetensors", - "model.layers.10.self_attn.k_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.10.self_attn.o_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.10.self_attn.q_proj.lora_A.default.weight": "model-00004-of-00030.safetensors", - "model.layers.10.self_attn.q_proj.lora_B.default.weight": "model-00004-of-00030.safetensors", - "model.layers.10.self_attn.q_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.10.self_attn.v_proj.lora_A.default.weight": "model-00004-of-00030.safetensors", - "model.layers.10.self_attn.v_proj.lora_B.default.weight": "model-00004-of-00030.safetensors", - "model.layers.10.self_attn.v_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.11.input_layernorm.weight": "model-00005-of-00030.safetensors", - "model.layers.11.mlp.down_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.11.mlp.gate_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.11.mlp.up_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.11.post_attention_layernorm.weight": "model-00005-of-00030.safetensors", - "model.layers.11.self_attn.k_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.11.self_attn.o_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.11.self_attn.q_proj.lora_A.default.weight": "model-00005-of-00030.safetensors", - "model.layers.11.self_attn.q_proj.lora_B.default.weight": "model-00005-of-00030.safetensors", - "model.layers.11.self_attn.q_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.11.self_attn.v_proj.lora_A.default.weight": "model-00005-of-00030.safetensors", - "model.layers.11.self_attn.v_proj.lora_B.default.weight": "model-00005-of-00030.safetensors", - "model.layers.11.self_attn.v_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.12.input_layernorm.weight": "model-00005-of-00030.safetensors", - "model.layers.12.mlp.down_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.12.mlp.gate_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.12.mlp.up_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.12.post_attention_layernorm.weight": "model-00005-of-00030.safetensors", - "model.layers.12.self_attn.k_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.12.self_attn.o_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.12.self_attn.q_proj.lora_A.default.weight": "model-00005-of-00030.safetensors", - "model.layers.12.self_attn.q_proj.lora_B.default.weight": "model-00005-of-00030.safetensors", - "model.layers.12.self_attn.q_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.12.self_attn.v_proj.lora_A.default.weight": "model-00005-of-00030.safetensors", - "model.layers.12.self_attn.v_proj.lora_B.default.weight": "model-00005-of-00030.safetensors", - "model.layers.12.self_attn.v_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.13.input_layernorm.weight": "model-00006-of-00030.safetensors", - "model.layers.13.mlp.down_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.13.mlp.gate_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.13.mlp.up_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.13.post_attention_layernorm.weight": "model-00006-of-00030.safetensors", - "model.layers.13.self_attn.k_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.13.self_attn.o_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.13.self_attn.q_proj.lora_A.default.weight": "model-00005-of-00030.safetensors", - "model.layers.13.self_attn.q_proj.lora_B.default.weight": "model-00005-of-00030.safetensors", - "model.layers.13.self_attn.q_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.13.self_attn.v_proj.lora_A.default.weight": "model-00005-of-00030.safetensors", - "model.layers.13.self_attn.v_proj.lora_B.default.weight": "model-00005-of-00030.safetensors", - "model.layers.13.self_attn.v_proj.weight": "model-00005-of-00030.safetensors", - "model.layers.14.input_layernorm.weight": "model-00006-of-00030.safetensors", - "model.layers.14.mlp.down_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.14.mlp.gate_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.14.mlp.up_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.14.post_attention_layernorm.weight": "model-00006-of-00030.safetensors", - "model.layers.14.self_attn.k_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.14.self_attn.o_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.14.self_attn.q_proj.lora_A.default.weight": "model-00006-of-00030.safetensors", - "model.layers.14.self_attn.q_proj.lora_B.default.weight": "model-00006-of-00030.safetensors", - "model.layers.14.self_attn.q_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.14.self_attn.v_proj.lora_A.default.weight": "model-00006-of-00030.safetensors", - "model.layers.14.self_attn.v_proj.lora_B.default.weight": "model-00006-of-00030.safetensors", - "model.layers.14.self_attn.v_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.15.input_layernorm.weight": "model-00006-of-00030.safetensors", - "model.layers.15.mlp.down_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.15.mlp.gate_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.15.mlp.up_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.15.post_attention_layernorm.weight": "model-00006-of-00030.safetensors", - "model.layers.15.self_attn.k_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.15.self_attn.o_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.15.self_attn.q_proj.lora_A.default.weight": "model-00006-of-00030.safetensors", - "model.layers.15.self_attn.q_proj.lora_B.default.weight": "model-00006-of-00030.safetensors", - "model.layers.15.self_attn.q_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.15.self_attn.v_proj.lora_A.default.weight": "model-00006-of-00030.safetensors", - "model.layers.15.self_attn.v_proj.lora_B.default.weight": "model-00006-of-00030.safetensors", - "model.layers.15.self_attn.v_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.16.input_layernorm.weight": "model-00007-of-00030.safetensors", - "model.layers.16.mlp.down_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.16.mlp.gate_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.16.mlp.up_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.16.post_attention_layernorm.weight": "model-00007-of-00030.safetensors", - "model.layers.16.self_attn.k_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.16.self_attn.o_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.16.self_attn.q_proj.lora_A.default.weight": "model-00006-of-00030.safetensors", - "model.layers.16.self_attn.q_proj.lora_B.default.weight": "model-00006-of-00030.safetensors", - "model.layers.16.self_attn.q_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.16.self_attn.v_proj.lora_A.default.weight": "model-00006-of-00030.safetensors", - "model.layers.16.self_attn.v_proj.lora_B.default.weight": "model-00006-of-00030.safetensors", - "model.layers.16.self_attn.v_proj.weight": "model-00006-of-00030.safetensors", - "model.layers.17.input_layernorm.weight": "model-00007-of-00030.safetensors", - "model.layers.17.mlp.down_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.17.mlp.gate_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.17.mlp.up_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.17.post_attention_layernorm.weight": "model-00007-of-00030.safetensors", - "model.layers.17.self_attn.k_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.17.self_attn.o_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.17.self_attn.q_proj.lora_A.default.weight": "model-00007-of-00030.safetensors", - "model.layers.17.self_attn.q_proj.lora_B.default.weight": "model-00007-of-00030.safetensors", - "model.layers.17.self_attn.q_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.17.self_attn.v_proj.lora_A.default.weight": "model-00007-of-00030.safetensors", - "model.layers.17.self_attn.v_proj.lora_B.default.weight": "model-00007-of-00030.safetensors", - "model.layers.17.self_attn.v_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.18.input_layernorm.weight": "model-00007-of-00030.safetensors", - "model.layers.18.mlp.down_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.18.mlp.gate_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.18.mlp.up_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.18.post_attention_layernorm.weight": "model-00007-of-00030.safetensors", - "model.layers.18.self_attn.k_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.18.self_attn.o_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.18.self_attn.q_proj.lora_A.default.weight": "model-00007-of-00030.safetensors", - "model.layers.18.self_attn.q_proj.lora_B.default.weight": "model-00007-of-00030.safetensors", - "model.layers.18.self_attn.q_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.18.self_attn.v_proj.lora_A.default.weight": "model-00007-of-00030.safetensors", - "model.layers.18.self_attn.v_proj.lora_B.default.weight": "model-00007-of-00030.safetensors", - "model.layers.18.self_attn.v_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.19.input_layernorm.weight": "model-00008-of-00030.safetensors", - "model.layers.19.mlp.down_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.19.mlp.gate_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.19.mlp.up_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.19.post_attention_layernorm.weight": "model-00008-of-00030.safetensors", - "model.layers.19.self_attn.k_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.19.self_attn.o_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.19.self_attn.q_proj.lora_A.default.weight": "model-00007-of-00030.safetensors", - "model.layers.19.self_attn.q_proj.lora_B.default.weight": "model-00007-of-00030.safetensors", - "model.layers.19.self_attn.q_proj.weight": "model-00007-of-00030.safetensors", - "model.layers.19.self_attn.v_proj.lora_A.default.weight": "model-00008-of-00030.safetensors", - "model.layers.19.self_attn.v_proj.lora_B.default.weight": "model-00008-of-00030.safetensors", - "model.layers.19.self_attn.v_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.2.input_layernorm.weight": "model-00002-of-00030.safetensors", - "model.layers.2.mlp.down_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.2.mlp.up_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00030.safetensors", - "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.2.self_attn.q_proj.lora_A.default.weight": "model-00001-of-00030.safetensors", - "model.layers.2.self_attn.q_proj.lora_B.default.weight": "model-00001-of-00030.safetensors", - "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.2.self_attn.v_proj.lora_A.default.weight": "model-00001-of-00030.safetensors", - "model.layers.2.self_attn.v_proj.lora_B.default.weight": "model-00001-of-00030.safetensors", - "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00030.safetensors", - "model.layers.20.input_layernorm.weight": "model-00008-of-00030.safetensors", - "model.layers.20.mlp.down_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.20.mlp.gate_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.20.mlp.up_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.20.post_attention_layernorm.weight": "model-00008-of-00030.safetensors", - "model.layers.20.self_attn.k_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.20.self_attn.o_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.20.self_attn.q_proj.lora_A.default.weight": "model-00008-of-00030.safetensors", - "model.layers.20.self_attn.q_proj.lora_B.default.weight": "model-00008-of-00030.safetensors", - "model.layers.20.self_attn.q_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.20.self_attn.v_proj.lora_A.default.weight": "model-00008-of-00030.safetensors", - "model.layers.20.self_attn.v_proj.lora_B.default.weight": "model-00008-of-00030.safetensors", - "model.layers.20.self_attn.v_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.21.input_layernorm.weight": "model-00009-of-00030.safetensors", - "model.layers.21.mlp.down_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.21.mlp.gate_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.21.mlp.up_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.21.post_attention_layernorm.weight": "model-00009-of-00030.safetensors", - "model.layers.21.self_attn.k_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.21.self_attn.o_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.21.self_attn.q_proj.lora_A.default.weight": "model-00008-of-00030.safetensors", - "model.layers.21.self_attn.q_proj.lora_B.default.weight": "model-00008-of-00030.safetensors", - "model.layers.21.self_attn.q_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.21.self_attn.v_proj.lora_A.default.weight": "model-00008-of-00030.safetensors", - "model.layers.21.self_attn.v_proj.lora_B.default.weight": "model-00008-of-00030.safetensors", - "model.layers.21.self_attn.v_proj.weight": "model-00008-of-00030.safetensors", - "model.layers.22.input_layernorm.weight": "model-00009-of-00030.safetensors", - "model.layers.22.mlp.down_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.22.mlp.gate_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.22.mlp.up_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.22.post_attention_layernorm.weight": "model-00009-of-00030.safetensors", - "model.layers.22.self_attn.k_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.22.self_attn.o_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.22.self_attn.q_proj.lora_A.default.weight": "model-00009-of-00030.safetensors", - "model.layers.22.self_attn.q_proj.lora_B.default.weight": "model-00009-of-00030.safetensors", - "model.layers.22.self_attn.q_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.22.self_attn.v_proj.lora_A.default.weight": "model-00009-of-00030.safetensors", - "model.layers.22.self_attn.v_proj.lora_B.default.weight": "model-00009-of-00030.safetensors", - "model.layers.22.self_attn.v_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.23.input_layernorm.weight": "model-00009-of-00030.safetensors", - "model.layers.23.mlp.down_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.23.mlp.gate_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.23.mlp.up_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.23.post_attention_layernorm.weight": "model-00009-of-00030.safetensors", - "model.layers.23.self_attn.k_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.23.self_attn.o_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.23.self_attn.q_proj.lora_A.default.weight": "model-00009-of-00030.safetensors", - "model.layers.23.self_attn.q_proj.lora_B.default.weight": "model-00009-of-00030.safetensors", - "model.layers.23.self_attn.q_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.23.self_attn.v_proj.lora_A.default.weight": "model-00009-of-00030.safetensors", - "model.layers.23.self_attn.v_proj.lora_B.default.weight": "model-00009-of-00030.safetensors", - "model.layers.23.self_attn.v_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.24.input_layernorm.weight": "model-00010-of-00030.safetensors", - "model.layers.24.mlp.down_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.24.mlp.gate_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.24.mlp.up_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.24.post_attention_layernorm.weight": "model-00010-of-00030.safetensors", - "model.layers.24.self_attn.k_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.24.self_attn.o_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.24.self_attn.q_proj.lora_A.default.weight": "model-00009-of-00030.safetensors", - "model.layers.24.self_attn.q_proj.lora_B.default.weight": "model-00009-of-00030.safetensors", - "model.layers.24.self_attn.q_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.24.self_attn.v_proj.lora_A.default.weight": "model-00009-of-00030.safetensors", - "model.layers.24.self_attn.v_proj.lora_B.default.weight": "model-00009-of-00030.safetensors", - "model.layers.24.self_attn.v_proj.weight": "model-00009-of-00030.safetensors", - "model.layers.25.input_layernorm.weight": "model-00010-of-00030.safetensors", - "model.layers.25.mlp.down_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.25.mlp.gate_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.25.mlp.up_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.25.post_attention_layernorm.weight": "model-00010-of-00030.safetensors", - "model.layers.25.self_attn.k_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.25.self_attn.o_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.25.self_attn.q_proj.lora_A.default.weight": "model-00010-of-00030.safetensors", - "model.layers.25.self_attn.q_proj.lora_B.default.weight": "model-00010-of-00030.safetensors", - "model.layers.25.self_attn.q_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.25.self_attn.v_proj.lora_A.default.weight": "model-00010-of-00030.safetensors", - "model.layers.25.self_attn.v_proj.lora_B.default.weight": "model-00010-of-00030.safetensors", - "model.layers.25.self_attn.v_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.26.input_layernorm.weight": "model-00010-of-00030.safetensors", - "model.layers.26.mlp.down_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.26.mlp.gate_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.26.mlp.up_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.26.post_attention_layernorm.weight": "model-00010-of-00030.safetensors", - "model.layers.26.self_attn.k_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.26.self_attn.o_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.26.self_attn.q_proj.lora_A.default.weight": "model-00010-of-00030.safetensors", - "model.layers.26.self_attn.q_proj.lora_B.default.weight": "model-00010-of-00030.safetensors", - "model.layers.26.self_attn.q_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.26.self_attn.v_proj.lora_A.default.weight": "model-00010-of-00030.safetensors", - "model.layers.26.self_attn.v_proj.lora_B.default.weight": "model-00010-of-00030.safetensors", - "model.layers.26.self_attn.v_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.27.input_layernorm.weight": "model-00011-of-00030.safetensors", - "model.layers.27.mlp.down_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.27.mlp.gate_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.27.mlp.up_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.27.post_attention_layernorm.weight": "model-00011-of-00030.safetensors", - "model.layers.27.self_attn.k_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.27.self_attn.o_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.27.self_attn.q_proj.lora_A.default.weight": "model-00010-of-00030.safetensors", - "model.layers.27.self_attn.q_proj.lora_B.default.weight": "model-00010-of-00030.safetensors", - "model.layers.27.self_attn.q_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.27.self_attn.v_proj.lora_A.default.weight": "model-00010-of-00030.safetensors", - "model.layers.27.self_attn.v_proj.lora_B.default.weight": "model-00010-of-00030.safetensors", - "model.layers.27.self_attn.v_proj.weight": "model-00010-of-00030.safetensors", - "model.layers.28.input_layernorm.weight": "model-00011-of-00030.safetensors", - "model.layers.28.mlp.down_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.28.mlp.gate_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.28.mlp.up_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.28.post_attention_layernorm.weight": "model-00011-of-00030.safetensors", - "model.layers.28.self_attn.k_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.28.self_attn.o_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.28.self_attn.q_proj.lora_A.default.weight": "model-00011-of-00030.safetensors", - "model.layers.28.self_attn.q_proj.lora_B.default.weight": "model-00011-of-00030.safetensors", - "model.layers.28.self_attn.q_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.28.self_attn.v_proj.lora_A.default.weight": "model-00011-of-00030.safetensors", - "model.layers.28.self_attn.v_proj.lora_B.default.weight": "model-00011-of-00030.safetensors", - "model.layers.28.self_attn.v_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.29.input_layernorm.weight": "model-00011-of-00030.safetensors", - "model.layers.29.mlp.down_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.29.mlp.gate_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.29.mlp.up_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.29.post_attention_layernorm.weight": "model-00011-of-00030.safetensors", - "model.layers.29.self_attn.k_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.29.self_attn.o_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.29.self_attn.q_proj.lora_A.default.weight": "model-00011-of-00030.safetensors", - "model.layers.29.self_attn.q_proj.lora_B.default.weight": "model-00011-of-00030.safetensors", - "model.layers.29.self_attn.q_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.29.self_attn.v_proj.lora_A.default.weight": "model-00011-of-00030.safetensors", - "model.layers.29.self_attn.v_proj.lora_B.default.weight": "model-00011-of-00030.safetensors", - "model.layers.29.self_attn.v_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.3.input_layernorm.weight": "model-00002-of-00030.safetensors", - "model.layers.3.mlp.down_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.3.mlp.up_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00030.safetensors", - "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.3.self_attn.q_proj.lora_A.default.weight": "model-00002-of-00030.safetensors", - "model.layers.3.self_attn.q_proj.lora_B.default.weight": "model-00002-of-00030.safetensors", - "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.3.self_attn.v_proj.lora_A.default.weight": "model-00002-of-00030.safetensors", - "model.layers.3.self_attn.v_proj.lora_B.default.weight": "model-00002-of-00030.safetensors", - "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.30.input_layernorm.weight": "model-00012-of-00030.safetensors", - "model.layers.30.mlp.down_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.30.mlp.gate_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.30.mlp.up_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.30.post_attention_layernorm.weight": "model-00012-of-00030.safetensors", - "model.layers.30.self_attn.k_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.30.self_attn.o_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.30.self_attn.q_proj.lora_A.default.weight": "model-00011-of-00030.safetensors", - "model.layers.30.self_attn.q_proj.lora_B.default.weight": "model-00011-of-00030.safetensors", - "model.layers.30.self_attn.q_proj.weight": "model-00011-of-00030.safetensors", - "model.layers.30.self_attn.v_proj.lora_A.default.weight": "model-00012-of-00030.safetensors", - "model.layers.30.self_attn.v_proj.lora_B.default.weight": "model-00012-of-00030.safetensors", - "model.layers.30.self_attn.v_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.31.input_layernorm.weight": "model-00012-of-00030.safetensors", - "model.layers.31.mlp.down_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.31.mlp.gate_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.31.mlp.up_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.31.post_attention_layernorm.weight": "model-00012-of-00030.safetensors", - "model.layers.31.self_attn.k_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.31.self_attn.o_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.31.self_attn.q_proj.lora_A.default.weight": "model-00012-of-00030.safetensors", - "model.layers.31.self_attn.q_proj.lora_B.default.weight": "model-00012-of-00030.safetensors", - "model.layers.31.self_attn.q_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.31.self_attn.v_proj.lora_A.default.weight": "model-00012-of-00030.safetensors", - "model.layers.31.self_attn.v_proj.lora_B.default.weight": "model-00012-of-00030.safetensors", - "model.layers.31.self_attn.v_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.32.input_layernorm.weight": "model-00013-of-00030.safetensors", - "model.layers.32.mlp.down_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.32.mlp.gate_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.32.mlp.up_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.32.post_attention_layernorm.weight": "model-00013-of-00030.safetensors", - "model.layers.32.self_attn.k_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.32.self_attn.o_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.32.self_attn.q_proj.lora_A.default.weight": "model-00012-of-00030.safetensors", - "model.layers.32.self_attn.q_proj.lora_B.default.weight": "model-00012-of-00030.safetensors", - "model.layers.32.self_attn.q_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.32.self_attn.v_proj.lora_A.default.weight": "model-00012-of-00030.safetensors", - "model.layers.32.self_attn.v_proj.lora_B.default.weight": "model-00012-of-00030.safetensors", - "model.layers.32.self_attn.v_proj.weight": "model-00012-of-00030.safetensors", - "model.layers.33.input_layernorm.weight": "model-00013-of-00030.safetensors", - "model.layers.33.mlp.down_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.33.mlp.gate_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.33.mlp.up_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.33.post_attention_layernorm.weight": "model-00013-of-00030.safetensors", - "model.layers.33.self_attn.k_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.33.self_attn.o_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.33.self_attn.q_proj.lora_A.default.weight": "model-00013-of-00030.safetensors", - "model.layers.33.self_attn.q_proj.lora_B.default.weight": "model-00013-of-00030.safetensors", - "model.layers.33.self_attn.q_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.33.self_attn.v_proj.lora_A.default.weight": "model-00013-of-00030.safetensors", - "model.layers.33.self_attn.v_proj.lora_B.default.weight": "model-00013-of-00030.safetensors", - "model.layers.33.self_attn.v_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.34.input_layernorm.weight": "model-00013-of-00030.safetensors", - "model.layers.34.mlp.down_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.34.mlp.gate_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.34.mlp.up_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.34.post_attention_layernorm.weight": "model-00013-of-00030.safetensors", - "model.layers.34.self_attn.k_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.34.self_attn.o_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.34.self_attn.q_proj.lora_A.default.weight": "model-00013-of-00030.safetensors", - "model.layers.34.self_attn.q_proj.lora_B.default.weight": "model-00013-of-00030.safetensors", - "model.layers.34.self_attn.q_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.34.self_attn.v_proj.lora_A.default.weight": "model-00013-of-00030.safetensors", - "model.layers.34.self_attn.v_proj.lora_B.default.weight": "model-00013-of-00030.safetensors", - "model.layers.34.self_attn.v_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.35.input_layernorm.weight": "model-00014-of-00030.safetensors", - "model.layers.35.mlp.down_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.35.mlp.gate_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.35.mlp.up_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.35.post_attention_layernorm.weight": "model-00014-of-00030.safetensors", - "model.layers.35.self_attn.k_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.35.self_attn.o_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.35.self_attn.q_proj.lora_A.default.weight": "model-00013-of-00030.safetensors", - "model.layers.35.self_attn.q_proj.lora_B.default.weight": "model-00013-of-00030.safetensors", - "model.layers.35.self_attn.q_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.35.self_attn.v_proj.lora_A.default.weight": "model-00013-of-00030.safetensors", - "model.layers.35.self_attn.v_proj.lora_B.default.weight": "model-00013-of-00030.safetensors", - "model.layers.35.self_attn.v_proj.weight": "model-00013-of-00030.safetensors", - "model.layers.36.input_layernorm.weight": "model-00014-of-00030.safetensors", - "model.layers.36.mlp.down_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.36.mlp.gate_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.36.mlp.up_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.36.post_attention_layernorm.weight": "model-00014-of-00030.safetensors", - "model.layers.36.self_attn.k_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.36.self_attn.o_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.36.self_attn.q_proj.lora_A.default.weight": "model-00014-of-00030.safetensors", - "model.layers.36.self_attn.q_proj.lora_B.default.weight": "model-00014-of-00030.safetensors", - "model.layers.36.self_attn.q_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.36.self_attn.v_proj.lora_A.default.weight": "model-00014-of-00030.safetensors", - "model.layers.36.self_attn.v_proj.lora_B.default.weight": "model-00014-of-00030.safetensors", - "model.layers.36.self_attn.v_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.37.input_layernorm.weight": "model-00014-of-00030.safetensors", - "model.layers.37.mlp.down_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.37.mlp.gate_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.37.mlp.up_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.37.post_attention_layernorm.weight": "model-00014-of-00030.safetensors", - "model.layers.37.self_attn.k_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.37.self_attn.o_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.37.self_attn.q_proj.lora_A.default.weight": "model-00014-of-00030.safetensors", - "model.layers.37.self_attn.q_proj.lora_B.default.weight": "model-00014-of-00030.safetensors", - "model.layers.37.self_attn.q_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.37.self_attn.v_proj.lora_A.default.weight": "model-00014-of-00030.safetensors", - "model.layers.37.self_attn.v_proj.lora_B.default.weight": "model-00014-of-00030.safetensors", - "model.layers.37.self_attn.v_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.38.input_layernorm.weight": "model-00015-of-00030.safetensors", - "model.layers.38.mlp.down_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.38.mlp.gate_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.38.mlp.up_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.38.post_attention_layernorm.weight": "model-00015-of-00030.safetensors", - "model.layers.38.self_attn.k_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.38.self_attn.o_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.38.self_attn.q_proj.lora_A.default.weight": "model-00014-of-00030.safetensors", - "model.layers.38.self_attn.q_proj.lora_B.default.weight": "model-00014-of-00030.safetensors", - "model.layers.38.self_attn.q_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.38.self_attn.v_proj.lora_A.default.weight": "model-00014-of-00030.safetensors", - "model.layers.38.self_attn.v_proj.lora_B.default.weight": "model-00014-of-00030.safetensors", - "model.layers.38.self_attn.v_proj.weight": "model-00014-of-00030.safetensors", - "model.layers.39.input_layernorm.weight": "model-00015-of-00030.safetensors", - "model.layers.39.mlp.down_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.39.mlp.gate_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.39.mlp.up_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.39.post_attention_layernorm.weight": "model-00015-of-00030.safetensors", - "model.layers.39.self_attn.k_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.39.self_attn.o_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.39.self_attn.q_proj.lora_A.default.weight": "model-00015-of-00030.safetensors", - "model.layers.39.self_attn.q_proj.lora_B.default.weight": "model-00015-of-00030.safetensors", - "model.layers.39.self_attn.q_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.39.self_attn.v_proj.lora_A.default.weight": "model-00015-of-00030.safetensors", - "model.layers.39.self_attn.v_proj.lora_B.default.weight": "model-00015-of-00030.safetensors", - "model.layers.39.self_attn.v_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.4.input_layernorm.weight": "model-00002-of-00030.safetensors", - "model.layers.4.mlp.down_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.4.mlp.gate_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.4.mlp.up_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00030.safetensors", - "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.4.self_attn.q_proj.lora_A.default.weight": "model-00002-of-00030.safetensors", - "model.layers.4.self_attn.q_proj.lora_B.default.weight": "model-00002-of-00030.safetensors", - "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.4.self_attn.v_proj.lora_A.default.weight": "model-00002-of-00030.safetensors", - "model.layers.4.self_attn.v_proj.lora_B.default.weight": "model-00002-of-00030.safetensors", - "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.40.input_layernorm.weight": "model-00015-of-00030.safetensors", - "model.layers.40.mlp.down_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.40.mlp.gate_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.40.mlp.up_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.40.post_attention_layernorm.weight": "model-00015-of-00030.safetensors", - "model.layers.40.self_attn.k_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.40.self_attn.o_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.40.self_attn.q_proj.lora_A.default.weight": "model-00015-of-00030.safetensors", - "model.layers.40.self_attn.q_proj.lora_B.default.weight": "model-00015-of-00030.safetensors", - "model.layers.40.self_attn.q_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.40.self_attn.v_proj.lora_A.default.weight": "model-00015-of-00030.safetensors", - "model.layers.40.self_attn.v_proj.lora_B.default.weight": "model-00015-of-00030.safetensors", - "model.layers.40.self_attn.v_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.41.input_layernorm.weight": "model-00016-of-00030.safetensors", - "model.layers.41.mlp.down_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.41.mlp.gate_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.41.mlp.up_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.41.post_attention_layernorm.weight": "model-00016-of-00030.safetensors", - "model.layers.41.self_attn.k_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.41.self_attn.o_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.41.self_attn.q_proj.lora_A.default.weight": "model-00015-of-00030.safetensors", - "model.layers.41.self_attn.q_proj.lora_B.default.weight": "model-00015-of-00030.safetensors", - "model.layers.41.self_attn.q_proj.weight": "model-00015-of-00030.safetensors", - "model.layers.41.self_attn.v_proj.lora_A.default.weight": "model-00016-of-00030.safetensors", - "model.layers.41.self_attn.v_proj.lora_B.default.weight": "model-00016-of-00030.safetensors", - "model.layers.41.self_attn.v_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.42.input_layernorm.weight": "model-00016-of-00030.safetensors", - "model.layers.42.mlp.down_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.42.mlp.gate_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.42.mlp.up_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.42.post_attention_layernorm.weight": "model-00016-of-00030.safetensors", - "model.layers.42.self_attn.k_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.42.self_attn.o_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.42.self_attn.q_proj.lora_A.default.weight": "model-00016-of-00030.safetensors", - "model.layers.42.self_attn.q_proj.lora_B.default.weight": "model-00016-of-00030.safetensors", - "model.layers.42.self_attn.q_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.42.self_attn.v_proj.lora_A.default.weight": "model-00016-of-00030.safetensors", - "model.layers.42.self_attn.v_proj.lora_B.default.weight": "model-00016-of-00030.safetensors", - "model.layers.42.self_attn.v_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.43.input_layernorm.weight": "model-00017-of-00030.safetensors", - "model.layers.43.mlp.down_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.43.mlp.gate_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.43.mlp.up_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.43.post_attention_layernorm.weight": "model-00017-of-00030.safetensors", - "model.layers.43.self_attn.k_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.43.self_attn.o_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.43.self_attn.q_proj.lora_A.default.weight": "model-00016-of-00030.safetensors", - "model.layers.43.self_attn.q_proj.lora_B.default.weight": "model-00016-of-00030.safetensors", - "model.layers.43.self_attn.q_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.43.self_attn.v_proj.lora_A.default.weight": "model-00016-of-00030.safetensors", - "model.layers.43.self_attn.v_proj.lora_B.default.weight": "model-00016-of-00030.safetensors", - "model.layers.43.self_attn.v_proj.weight": "model-00016-of-00030.safetensors", - "model.layers.44.input_layernorm.weight": "model-00017-of-00030.safetensors", - "model.layers.44.mlp.down_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.44.mlp.gate_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.44.mlp.up_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.44.post_attention_layernorm.weight": "model-00017-of-00030.safetensors", - "model.layers.44.self_attn.k_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.44.self_attn.o_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.44.self_attn.q_proj.lora_A.default.weight": "model-00017-of-00030.safetensors", - "model.layers.44.self_attn.q_proj.lora_B.default.weight": "model-00017-of-00030.safetensors", - "model.layers.44.self_attn.q_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.44.self_attn.v_proj.lora_A.default.weight": "model-00017-of-00030.safetensors", - "model.layers.44.self_attn.v_proj.lora_B.default.weight": "model-00017-of-00030.safetensors", - "model.layers.44.self_attn.v_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.45.input_layernorm.weight": "model-00017-of-00030.safetensors", - "model.layers.45.mlp.down_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.45.mlp.gate_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.45.mlp.up_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.45.post_attention_layernorm.weight": "model-00017-of-00030.safetensors", - "model.layers.45.self_attn.k_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.45.self_attn.o_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.45.self_attn.q_proj.lora_A.default.weight": "model-00017-of-00030.safetensors", - "model.layers.45.self_attn.q_proj.lora_B.default.weight": "model-00017-of-00030.safetensors", - "model.layers.45.self_attn.q_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.45.self_attn.v_proj.lora_A.default.weight": "model-00017-of-00030.safetensors", - "model.layers.45.self_attn.v_proj.lora_B.default.weight": "model-00017-of-00030.safetensors", - "model.layers.45.self_attn.v_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.46.input_layernorm.weight": "model-00018-of-00030.safetensors", - "model.layers.46.mlp.down_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.46.mlp.gate_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.46.mlp.up_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.46.post_attention_layernorm.weight": "model-00018-of-00030.safetensors", - "model.layers.46.self_attn.k_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.46.self_attn.o_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.46.self_attn.q_proj.lora_A.default.weight": "model-00017-of-00030.safetensors", - "model.layers.46.self_attn.q_proj.lora_B.default.weight": "model-00017-of-00030.safetensors", - "model.layers.46.self_attn.q_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.46.self_attn.v_proj.lora_A.default.weight": "model-00017-of-00030.safetensors", - "model.layers.46.self_attn.v_proj.lora_B.default.weight": "model-00017-of-00030.safetensors", - "model.layers.46.self_attn.v_proj.weight": "model-00017-of-00030.safetensors", - "model.layers.47.input_layernorm.weight": "model-00018-of-00030.safetensors", - "model.layers.47.mlp.down_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.47.mlp.gate_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.47.mlp.up_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.47.post_attention_layernorm.weight": "model-00018-of-00030.safetensors", - "model.layers.47.self_attn.k_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.47.self_attn.o_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.47.self_attn.q_proj.lora_A.default.weight": "model-00018-of-00030.safetensors", - "model.layers.47.self_attn.q_proj.lora_B.default.weight": "model-00018-of-00030.safetensors", - "model.layers.47.self_attn.q_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.47.self_attn.v_proj.lora_A.default.weight": "model-00018-of-00030.safetensors", - "model.layers.47.self_attn.v_proj.lora_B.default.weight": "model-00018-of-00030.safetensors", - "model.layers.47.self_attn.v_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.48.input_layernorm.weight": "model-00018-of-00030.safetensors", - "model.layers.48.mlp.down_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.48.mlp.gate_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.48.mlp.up_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.48.post_attention_layernorm.weight": "model-00018-of-00030.safetensors", - "model.layers.48.self_attn.k_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.48.self_attn.o_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.48.self_attn.q_proj.lora_A.default.weight": "model-00018-of-00030.safetensors", - "model.layers.48.self_attn.q_proj.lora_B.default.weight": "model-00018-of-00030.safetensors", - "model.layers.48.self_attn.q_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.48.self_attn.v_proj.lora_A.default.weight": "model-00018-of-00030.safetensors", - "model.layers.48.self_attn.v_proj.lora_B.default.weight": "model-00018-of-00030.safetensors", - "model.layers.48.self_attn.v_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.49.input_layernorm.weight": "model-00019-of-00030.safetensors", - "model.layers.49.mlp.down_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.49.mlp.gate_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.49.mlp.up_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.49.post_attention_layernorm.weight": "model-00019-of-00030.safetensors", - "model.layers.49.self_attn.k_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.49.self_attn.o_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.49.self_attn.q_proj.lora_A.default.weight": "model-00018-of-00030.safetensors", - "model.layers.49.self_attn.q_proj.lora_B.default.weight": "model-00018-of-00030.safetensors", - "model.layers.49.self_attn.q_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.49.self_attn.v_proj.lora_A.default.weight": "model-00018-of-00030.safetensors", - "model.layers.49.self_attn.v_proj.lora_B.default.weight": "model-00018-of-00030.safetensors", - "model.layers.49.self_attn.v_proj.weight": "model-00018-of-00030.safetensors", - "model.layers.5.input_layernorm.weight": "model-00003-of-00030.safetensors", - "model.layers.5.mlp.down_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.5.mlp.gate_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.5.mlp.up_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.5.post_attention_layernorm.weight": "model-00003-of-00030.safetensors", - "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.5.self_attn.q_proj.lora_A.default.weight": "model-00002-of-00030.safetensors", - "model.layers.5.self_attn.q_proj.lora_B.default.weight": "model-00002-of-00030.safetensors", - "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.5.self_attn.v_proj.lora_A.default.weight": "model-00002-of-00030.safetensors", - "model.layers.5.self_attn.v_proj.lora_B.default.weight": "model-00002-of-00030.safetensors", - "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00030.safetensors", - "model.layers.50.input_layernorm.weight": "model-00019-of-00030.safetensors", - "model.layers.50.mlp.down_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.50.mlp.gate_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.50.mlp.up_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.50.post_attention_layernorm.weight": "model-00019-of-00030.safetensors", - "model.layers.50.self_attn.k_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.50.self_attn.o_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.50.self_attn.q_proj.lora_A.default.weight": "model-00019-of-00030.safetensors", - "model.layers.50.self_attn.q_proj.lora_B.default.weight": "model-00019-of-00030.safetensors", - "model.layers.50.self_attn.q_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.50.self_attn.v_proj.lora_A.default.weight": "model-00019-of-00030.safetensors", - "model.layers.50.self_attn.v_proj.lora_B.default.weight": "model-00019-of-00030.safetensors", - "model.layers.50.self_attn.v_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.51.input_layernorm.weight": "model-00019-of-00030.safetensors", - "model.layers.51.mlp.down_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.51.mlp.gate_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.51.mlp.up_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.51.post_attention_layernorm.weight": "model-00019-of-00030.safetensors", - "model.layers.51.self_attn.k_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.51.self_attn.o_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.51.self_attn.q_proj.lora_A.default.weight": "model-00019-of-00030.safetensors", - "model.layers.51.self_attn.q_proj.lora_B.default.weight": "model-00019-of-00030.safetensors", - "model.layers.51.self_attn.q_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.51.self_attn.v_proj.lora_A.default.weight": "model-00019-of-00030.safetensors", - "model.layers.51.self_attn.v_proj.lora_B.default.weight": "model-00019-of-00030.safetensors", - "model.layers.51.self_attn.v_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.52.input_layernorm.weight": "model-00020-of-00030.safetensors", - "model.layers.52.mlp.down_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.52.mlp.gate_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.52.mlp.up_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.52.post_attention_layernorm.weight": "model-00020-of-00030.safetensors", - "model.layers.52.self_attn.k_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.52.self_attn.o_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.52.self_attn.q_proj.lora_A.default.weight": "model-00019-of-00030.safetensors", - "model.layers.52.self_attn.q_proj.lora_B.default.weight": "model-00019-of-00030.safetensors", - "model.layers.52.self_attn.q_proj.weight": "model-00019-of-00030.safetensors", - "model.layers.52.self_attn.v_proj.lora_A.default.weight": "model-00020-of-00030.safetensors", - "model.layers.52.self_attn.v_proj.lora_B.default.weight": "model-00020-of-00030.safetensors", - "model.layers.52.self_attn.v_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.53.input_layernorm.weight": "model-00020-of-00030.safetensors", - "model.layers.53.mlp.down_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.53.mlp.gate_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.53.mlp.up_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.53.post_attention_layernorm.weight": "model-00020-of-00030.safetensors", - "model.layers.53.self_attn.k_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.53.self_attn.o_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.53.self_attn.q_proj.lora_A.default.weight": "model-00020-of-00030.safetensors", - "model.layers.53.self_attn.q_proj.lora_B.default.weight": "model-00020-of-00030.safetensors", - "model.layers.53.self_attn.q_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.53.self_attn.v_proj.lora_A.default.weight": "model-00020-of-00030.safetensors", - "model.layers.53.self_attn.v_proj.lora_B.default.weight": "model-00020-of-00030.safetensors", - "model.layers.53.self_attn.v_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.54.input_layernorm.weight": "model-00021-of-00030.safetensors", - "model.layers.54.mlp.down_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.54.mlp.gate_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.54.mlp.up_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.54.post_attention_layernorm.weight": "model-00021-of-00030.safetensors", - "model.layers.54.self_attn.k_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.54.self_attn.o_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.54.self_attn.q_proj.lora_A.default.weight": "model-00020-of-00030.safetensors", - "model.layers.54.self_attn.q_proj.lora_B.default.weight": "model-00020-of-00030.safetensors", - "model.layers.54.self_attn.q_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.54.self_attn.v_proj.lora_A.default.weight": "model-00020-of-00030.safetensors", - "model.layers.54.self_attn.v_proj.lora_B.default.weight": "model-00020-of-00030.safetensors", - "model.layers.54.self_attn.v_proj.weight": "model-00020-of-00030.safetensors", - "model.layers.55.input_layernorm.weight": "model-00021-of-00030.safetensors", - "model.layers.55.mlp.down_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.55.mlp.gate_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.55.mlp.up_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.55.post_attention_layernorm.weight": "model-00021-of-00030.safetensors", - "model.layers.55.self_attn.k_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.55.self_attn.o_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.55.self_attn.q_proj.lora_A.default.weight": "model-00021-of-00030.safetensors", - "model.layers.55.self_attn.q_proj.lora_B.default.weight": "model-00021-of-00030.safetensors", - "model.layers.55.self_attn.q_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.55.self_attn.v_proj.lora_A.default.weight": "model-00021-of-00030.safetensors", - "model.layers.55.self_attn.v_proj.lora_B.default.weight": "model-00021-of-00030.safetensors", - "model.layers.55.self_attn.v_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.56.input_layernorm.weight": "model-00021-of-00030.safetensors", - "model.layers.56.mlp.down_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.56.mlp.gate_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.56.mlp.up_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.56.post_attention_layernorm.weight": "model-00021-of-00030.safetensors", - "model.layers.56.self_attn.k_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.56.self_attn.o_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.56.self_attn.q_proj.lora_A.default.weight": "model-00021-of-00030.safetensors", - "model.layers.56.self_attn.q_proj.lora_B.default.weight": "model-00021-of-00030.safetensors", - "model.layers.56.self_attn.q_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.56.self_attn.v_proj.lora_A.default.weight": "model-00021-of-00030.safetensors", - "model.layers.56.self_attn.v_proj.lora_B.default.weight": "model-00021-of-00030.safetensors", - "model.layers.56.self_attn.v_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.57.input_layernorm.weight": "model-00022-of-00030.safetensors", - "model.layers.57.mlp.down_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.57.mlp.gate_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.57.mlp.up_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.57.post_attention_layernorm.weight": "model-00022-of-00030.safetensors", - "model.layers.57.self_attn.k_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.57.self_attn.o_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.57.self_attn.q_proj.lora_A.default.weight": "model-00021-of-00030.safetensors", - "model.layers.57.self_attn.q_proj.lora_B.default.weight": "model-00021-of-00030.safetensors", - "model.layers.57.self_attn.q_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.57.self_attn.v_proj.lora_A.default.weight": "model-00021-of-00030.safetensors", - "model.layers.57.self_attn.v_proj.lora_B.default.weight": "model-00021-of-00030.safetensors", - "model.layers.57.self_attn.v_proj.weight": "model-00021-of-00030.safetensors", - "model.layers.58.input_layernorm.weight": "model-00022-of-00030.safetensors", - "model.layers.58.mlp.down_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.58.mlp.gate_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.58.mlp.up_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.58.post_attention_layernorm.weight": "model-00022-of-00030.safetensors", - "model.layers.58.self_attn.k_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.58.self_attn.o_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.58.self_attn.q_proj.lora_A.default.weight": "model-00022-of-00030.safetensors", - "model.layers.58.self_attn.q_proj.lora_B.default.weight": "model-00022-of-00030.safetensors", - "model.layers.58.self_attn.q_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.58.self_attn.v_proj.lora_A.default.weight": "model-00022-of-00030.safetensors", - "model.layers.58.self_attn.v_proj.lora_B.default.weight": "model-00022-of-00030.safetensors", - "model.layers.58.self_attn.v_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.59.input_layernorm.weight": "model-00022-of-00030.safetensors", - "model.layers.59.mlp.down_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.59.mlp.gate_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.59.mlp.up_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.59.post_attention_layernorm.weight": "model-00022-of-00030.safetensors", - "model.layers.59.self_attn.k_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.59.self_attn.o_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.59.self_attn.q_proj.lora_A.default.weight": "model-00022-of-00030.safetensors", - "model.layers.59.self_attn.q_proj.lora_B.default.weight": "model-00022-of-00030.safetensors", - "model.layers.59.self_attn.q_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.59.self_attn.v_proj.lora_A.default.weight": "model-00022-of-00030.safetensors", - "model.layers.59.self_attn.v_proj.lora_B.default.weight": "model-00022-of-00030.safetensors", - "model.layers.59.self_attn.v_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.6.input_layernorm.weight": "model-00003-of-00030.safetensors", - "model.layers.6.mlp.down_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.6.mlp.gate_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.6.mlp.up_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.6.post_attention_layernorm.weight": "model-00003-of-00030.safetensors", - "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.6.self_attn.o_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.6.self_attn.q_proj.lora_A.default.weight": "model-00003-of-00030.safetensors", - "model.layers.6.self_attn.q_proj.lora_B.default.weight": "model-00003-of-00030.safetensors", - "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.6.self_attn.v_proj.lora_A.default.weight": "model-00003-of-00030.safetensors", - "model.layers.6.self_attn.v_proj.lora_B.default.weight": "model-00003-of-00030.safetensors", - "model.layers.6.self_attn.v_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.60.input_layernorm.weight": "model-00023-of-00030.safetensors", - "model.layers.60.mlp.down_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.60.mlp.gate_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.60.mlp.up_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.60.post_attention_layernorm.weight": "model-00023-of-00030.safetensors", - "model.layers.60.self_attn.k_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.60.self_attn.o_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.60.self_attn.q_proj.lora_A.default.weight": "model-00022-of-00030.safetensors", - "model.layers.60.self_attn.q_proj.lora_B.default.weight": "model-00022-of-00030.safetensors", - "model.layers.60.self_attn.q_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.60.self_attn.v_proj.lora_A.default.weight": "model-00022-of-00030.safetensors", - "model.layers.60.self_attn.v_proj.lora_B.default.weight": "model-00022-of-00030.safetensors", - "model.layers.60.self_attn.v_proj.weight": "model-00022-of-00030.safetensors", - "model.layers.61.input_layernorm.weight": "model-00023-of-00030.safetensors", - "model.layers.61.mlp.down_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.61.mlp.gate_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.61.mlp.up_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.61.post_attention_layernorm.weight": "model-00023-of-00030.safetensors", - "model.layers.61.self_attn.k_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.61.self_attn.o_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.61.self_attn.q_proj.lora_A.default.weight": "model-00023-of-00030.safetensors", - "model.layers.61.self_attn.q_proj.lora_B.default.weight": "model-00023-of-00030.safetensors", - "model.layers.61.self_attn.q_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.61.self_attn.v_proj.lora_A.default.weight": "model-00023-of-00030.safetensors", - "model.layers.61.self_attn.v_proj.lora_B.default.weight": "model-00023-of-00030.safetensors", - "model.layers.61.self_attn.v_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.62.input_layernorm.weight": "model-00023-of-00030.safetensors", - "model.layers.62.mlp.down_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.62.mlp.gate_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.62.mlp.up_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.62.post_attention_layernorm.weight": "model-00023-of-00030.safetensors", - "model.layers.62.self_attn.k_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.62.self_attn.o_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.62.self_attn.q_proj.lora_A.default.weight": "model-00023-of-00030.safetensors", - "model.layers.62.self_attn.q_proj.lora_B.default.weight": "model-00023-of-00030.safetensors", - "model.layers.62.self_attn.q_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.62.self_attn.v_proj.lora_A.default.weight": "model-00023-of-00030.safetensors", - "model.layers.62.self_attn.v_proj.lora_B.default.weight": "model-00023-of-00030.safetensors", - "model.layers.62.self_attn.v_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.63.input_layernorm.weight": "model-00024-of-00030.safetensors", - "model.layers.63.mlp.down_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.63.mlp.gate_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.63.mlp.up_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.63.post_attention_layernorm.weight": "model-00024-of-00030.safetensors", - "model.layers.63.self_attn.k_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.63.self_attn.o_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.63.self_attn.q_proj.lora_A.default.weight": "model-00023-of-00030.safetensors", - "model.layers.63.self_attn.q_proj.lora_B.default.weight": "model-00023-of-00030.safetensors", - "model.layers.63.self_attn.q_proj.weight": "model-00023-of-00030.safetensors", - "model.layers.63.self_attn.v_proj.lora_A.default.weight": "model-00024-of-00030.safetensors", - "model.layers.63.self_attn.v_proj.lora_B.default.weight": "model-00024-of-00030.safetensors", - "model.layers.63.self_attn.v_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.64.input_layernorm.weight": "model-00024-of-00030.safetensors", - "model.layers.64.mlp.down_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.64.mlp.gate_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.64.mlp.up_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.64.post_attention_layernorm.weight": "model-00024-of-00030.safetensors", - "model.layers.64.self_attn.k_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.64.self_attn.o_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.64.self_attn.q_proj.lora_A.default.weight": "model-00024-of-00030.safetensors", - "model.layers.64.self_attn.q_proj.lora_B.default.weight": "model-00024-of-00030.safetensors", - "model.layers.64.self_attn.q_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.64.self_attn.v_proj.lora_A.default.weight": "model-00024-of-00030.safetensors", - "model.layers.64.self_attn.v_proj.lora_B.default.weight": "model-00024-of-00030.safetensors", - "model.layers.64.self_attn.v_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.65.input_layernorm.weight": "model-00025-of-00030.safetensors", - "model.layers.65.mlp.down_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.65.mlp.gate_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.65.mlp.up_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.65.post_attention_layernorm.weight": "model-00025-of-00030.safetensors", - "model.layers.65.self_attn.k_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.65.self_attn.o_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.65.self_attn.q_proj.lora_A.default.weight": "model-00024-of-00030.safetensors", - "model.layers.65.self_attn.q_proj.lora_B.default.weight": "model-00024-of-00030.safetensors", - "model.layers.65.self_attn.q_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.65.self_attn.v_proj.lora_A.default.weight": "model-00024-of-00030.safetensors", - "model.layers.65.self_attn.v_proj.lora_B.default.weight": "model-00024-of-00030.safetensors", - "model.layers.65.self_attn.v_proj.weight": "model-00024-of-00030.safetensors", - "model.layers.66.input_layernorm.weight": "model-00025-of-00030.safetensors", - "model.layers.66.mlp.down_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.66.mlp.gate_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.66.mlp.up_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.66.post_attention_layernorm.weight": "model-00025-of-00030.safetensors", - "model.layers.66.self_attn.k_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.66.self_attn.o_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.66.self_attn.q_proj.lora_A.default.weight": "model-00025-of-00030.safetensors", - "model.layers.66.self_attn.q_proj.lora_B.default.weight": "model-00025-of-00030.safetensors", - "model.layers.66.self_attn.q_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.66.self_attn.v_proj.lora_A.default.weight": "model-00025-of-00030.safetensors", - "model.layers.66.self_attn.v_proj.lora_B.default.weight": "model-00025-of-00030.safetensors", - "model.layers.66.self_attn.v_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.67.input_layernorm.weight": "model-00025-of-00030.safetensors", - "model.layers.67.mlp.down_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.67.mlp.gate_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.67.mlp.up_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.67.post_attention_layernorm.weight": "model-00025-of-00030.safetensors", - "model.layers.67.self_attn.k_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.67.self_attn.o_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.67.self_attn.q_proj.lora_A.default.weight": "model-00025-of-00030.safetensors", - "model.layers.67.self_attn.q_proj.lora_B.default.weight": "model-00025-of-00030.safetensors", - "model.layers.67.self_attn.q_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.67.self_attn.v_proj.lora_A.default.weight": "model-00025-of-00030.safetensors", - "model.layers.67.self_attn.v_proj.lora_B.default.weight": "model-00025-of-00030.safetensors", - "model.layers.67.self_attn.v_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.68.input_layernorm.weight": "model-00026-of-00030.safetensors", - "model.layers.68.mlp.down_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.68.mlp.gate_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.68.mlp.up_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.68.post_attention_layernorm.weight": "model-00026-of-00030.safetensors", - "model.layers.68.self_attn.k_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.68.self_attn.o_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.68.self_attn.q_proj.lora_A.default.weight": "model-00025-of-00030.safetensors", - "model.layers.68.self_attn.q_proj.lora_B.default.weight": "model-00025-of-00030.safetensors", - "model.layers.68.self_attn.q_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.68.self_attn.v_proj.lora_A.default.weight": "model-00025-of-00030.safetensors", - "model.layers.68.self_attn.v_proj.lora_B.default.weight": "model-00025-of-00030.safetensors", - "model.layers.68.self_attn.v_proj.weight": "model-00025-of-00030.safetensors", - "model.layers.69.input_layernorm.weight": "model-00026-of-00030.safetensors", - "model.layers.69.mlp.down_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.69.mlp.gate_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.69.mlp.up_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.69.post_attention_layernorm.weight": "model-00026-of-00030.safetensors", - "model.layers.69.self_attn.k_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.69.self_attn.o_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.69.self_attn.q_proj.lora_A.default.weight": "model-00026-of-00030.safetensors", - "model.layers.69.self_attn.q_proj.lora_B.default.weight": "model-00026-of-00030.safetensors", - "model.layers.69.self_attn.q_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.69.self_attn.v_proj.lora_A.default.weight": "model-00026-of-00030.safetensors", - "model.layers.69.self_attn.v_proj.lora_B.default.weight": "model-00026-of-00030.safetensors", - "model.layers.69.self_attn.v_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.7.input_layernorm.weight": "model-00003-of-00030.safetensors", - "model.layers.7.mlp.down_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.7.mlp.gate_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.7.mlp.up_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.7.post_attention_layernorm.weight": "model-00003-of-00030.safetensors", - "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.7.self_attn.o_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.7.self_attn.q_proj.lora_A.default.weight": "model-00003-of-00030.safetensors", - "model.layers.7.self_attn.q_proj.lora_B.default.weight": "model-00003-of-00030.safetensors", - "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.7.self_attn.v_proj.lora_A.default.weight": "model-00003-of-00030.safetensors", - "model.layers.7.self_attn.v_proj.lora_B.default.weight": "model-00003-of-00030.safetensors", - "model.layers.7.self_attn.v_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.70.input_layernorm.weight": "model-00026-of-00030.safetensors", - "model.layers.70.mlp.down_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.70.mlp.gate_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.70.mlp.up_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.70.post_attention_layernorm.weight": "model-00026-of-00030.safetensors", - "model.layers.70.self_attn.k_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.70.self_attn.o_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.70.self_attn.q_proj.lora_A.default.weight": "model-00026-of-00030.safetensors", - "model.layers.70.self_attn.q_proj.lora_B.default.weight": "model-00026-of-00030.safetensors", - "model.layers.70.self_attn.q_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.70.self_attn.v_proj.lora_A.default.weight": "model-00026-of-00030.safetensors", - "model.layers.70.self_attn.v_proj.lora_B.default.weight": "model-00026-of-00030.safetensors", - "model.layers.70.self_attn.v_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.71.input_layernorm.weight": "model-00027-of-00030.safetensors", - "model.layers.71.mlp.down_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.71.mlp.gate_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.71.mlp.up_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.71.post_attention_layernorm.weight": "model-00027-of-00030.safetensors", - "model.layers.71.self_attn.k_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.71.self_attn.o_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.71.self_attn.q_proj.lora_A.default.weight": "model-00026-of-00030.safetensors", - "model.layers.71.self_attn.q_proj.lora_B.default.weight": "model-00026-of-00030.safetensors", - "model.layers.71.self_attn.q_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.71.self_attn.v_proj.lora_A.default.weight": "model-00026-of-00030.safetensors", - "model.layers.71.self_attn.v_proj.lora_B.default.weight": "model-00026-of-00030.safetensors", - "model.layers.71.self_attn.v_proj.weight": "model-00026-of-00030.safetensors", - "model.layers.72.input_layernorm.weight": "model-00027-of-00030.safetensors", - "model.layers.72.mlp.down_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.72.mlp.gate_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.72.mlp.up_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.72.post_attention_layernorm.weight": "model-00027-of-00030.safetensors", - "model.layers.72.self_attn.k_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.72.self_attn.o_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.72.self_attn.q_proj.lora_A.default.weight": "model-00027-of-00030.safetensors", - "model.layers.72.self_attn.q_proj.lora_B.default.weight": "model-00027-of-00030.safetensors", - "model.layers.72.self_attn.q_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.72.self_attn.v_proj.lora_A.default.weight": "model-00027-of-00030.safetensors", - "model.layers.72.self_attn.v_proj.lora_B.default.weight": "model-00027-of-00030.safetensors", - "model.layers.72.self_attn.v_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.73.input_layernorm.weight": "model-00027-of-00030.safetensors", - "model.layers.73.mlp.down_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.73.mlp.gate_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.73.mlp.up_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.73.post_attention_layernorm.weight": "model-00027-of-00030.safetensors", - "model.layers.73.self_attn.k_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.73.self_attn.o_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.73.self_attn.q_proj.lora_A.default.weight": "model-00027-of-00030.safetensors", - "model.layers.73.self_attn.q_proj.lora_B.default.weight": "model-00027-of-00030.safetensors", - "model.layers.73.self_attn.q_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.73.self_attn.v_proj.lora_A.default.weight": "model-00027-of-00030.safetensors", - "model.layers.73.self_attn.v_proj.lora_B.default.weight": "model-00027-of-00030.safetensors", - "model.layers.73.self_attn.v_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.74.input_layernorm.weight": "model-00028-of-00030.safetensors", - "model.layers.74.mlp.down_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.74.mlp.gate_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.74.mlp.up_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.74.post_attention_layernorm.weight": "model-00028-of-00030.safetensors", - "model.layers.74.self_attn.k_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.74.self_attn.o_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.74.self_attn.q_proj.lora_A.default.weight": "model-00027-of-00030.safetensors", - "model.layers.74.self_attn.q_proj.lora_B.default.weight": "model-00027-of-00030.safetensors", - "model.layers.74.self_attn.q_proj.weight": "model-00027-of-00030.safetensors", - "model.layers.74.self_attn.v_proj.lora_A.default.weight": "model-00028-of-00030.safetensors", - "model.layers.74.self_attn.v_proj.lora_B.default.weight": "model-00028-of-00030.safetensors", - "model.layers.74.self_attn.v_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.75.input_layernorm.weight": "model-00028-of-00030.safetensors", - "model.layers.75.mlp.down_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.75.mlp.gate_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.75.mlp.up_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.75.post_attention_layernorm.weight": "model-00028-of-00030.safetensors", - "model.layers.75.self_attn.k_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.75.self_attn.o_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.75.self_attn.q_proj.lora_A.default.weight": "model-00028-of-00030.safetensors", - "model.layers.75.self_attn.q_proj.lora_B.default.weight": "model-00028-of-00030.safetensors", - "model.layers.75.self_attn.q_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.75.self_attn.v_proj.lora_A.default.weight": "model-00028-of-00030.safetensors", - "model.layers.75.self_attn.v_proj.lora_B.default.weight": "model-00028-of-00030.safetensors", - "model.layers.75.self_attn.v_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.76.input_layernorm.weight": "model-00029-of-00030.safetensors", - "model.layers.76.mlp.down_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.76.mlp.gate_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.76.mlp.up_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.76.post_attention_layernorm.weight": "model-00029-of-00030.safetensors", - "model.layers.76.self_attn.k_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.76.self_attn.o_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.76.self_attn.q_proj.lora_A.default.weight": "model-00028-of-00030.safetensors", - "model.layers.76.self_attn.q_proj.lora_B.default.weight": "model-00028-of-00030.safetensors", - "model.layers.76.self_attn.q_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.76.self_attn.v_proj.lora_A.default.weight": "model-00028-of-00030.safetensors", - "model.layers.76.self_attn.v_proj.lora_B.default.weight": "model-00028-of-00030.safetensors", - "model.layers.76.self_attn.v_proj.weight": "model-00028-of-00030.safetensors", - "model.layers.77.input_layernorm.weight": "model-00029-of-00030.safetensors", - "model.layers.77.mlp.down_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.77.mlp.gate_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.77.mlp.up_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.77.post_attention_layernorm.weight": "model-00029-of-00030.safetensors", - "model.layers.77.self_attn.k_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.77.self_attn.o_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.77.self_attn.q_proj.lora_A.default.weight": "model-00029-of-00030.safetensors", - "model.layers.77.self_attn.q_proj.lora_B.default.weight": "model-00029-of-00030.safetensors", - "model.layers.77.self_attn.q_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.77.self_attn.v_proj.lora_A.default.weight": "model-00029-of-00030.safetensors", - "model.layers.77.self_attn.v_proj.lora_B.default.weight": "model-00029-of-00030.safetensors", - "model.layers.77.self_attn.v_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.78.input_layernorm.weight": "model-00029-of-00030.safetensors", - "model.layers.78.mlp.down_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.78.mlp.gate_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.78.mlp.up_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.78.post_attention_layernorm.weight": "model-00029-of-00030.safetensors", - "model.layers.78.self_attn.k_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.78.self_attn.o_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.78.self_attn.q_proj.lora_A.default.weight": "model-00029-of-00030.safetensors", - "model.layers.78.self_attn.q_proj.lora_B.default.weight": "model-00029-of-00030.safetensors", - "model.layers.78.self_attn.q_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.78.self_attn.v_proj.lora_A.default.weight": "model-00029-of-00030.safetensors", - "model.layers.78.self_attn.v_proj.lora_B.default.weight": "model-00029-of-00030.safetensors", - "model.layers.78.self_attn.v_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.79.input_layernorm.weight": "model-00030-of-00030.safetensors", - "model.layers.79.mlp.down_proj.weight": "model-00030-of-00030.safetensors", - "model.layers.79.mlp.gate_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.79.mlp.up_proj.weight": "model-00030-of-00030.safetensors", - "model.layers.79.post_attention_layernorm.weight": "model-00030-of-00030.safetensors", - "model.layers.79.self_attn.k_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.79.self_attn.o_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.79.self_attn.q_proj.lora_A.default.weight": "model-00029-of-00030.safetensors", - "model.layers.79.self_attn.q_proj.lora_B.default.weight": "model-00029-of-00030.safetensors", - "model.layers.79.self_attn.q_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.79.self_attn.v_proj.lora_A.default.weight": "model-00029-of-00030.safetensors", - "model.layers.79.self_attn.v_proj.lora_B.default.weight": "model-00029-of-00030.safetensors", - "model.layers.79.self_attn.v_proj.weight": "model-00029-of-00030.safetensors", - "model.layers.8.input_layernorm.weight": "model-00004-of-00030.safetensors", - "model.layers.8.mlp.down_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.8.mlp.gate_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.8.mlp.up_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.8.post_attention_layernorm.weight": "model-00004-of-00030.safetensors", - "model.layers.8.self_attn.k_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.8.self_attn.o_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.8.self_attn.q_proj.lora_A.default.weight": "model-00003-of-00030.safetensors", - "model.layers.8.self_attn.q_proj.lora_B.default.weight": "model-00003-of-00030.safetensors", - "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00030.safetensors", - "model.layers.8.self_attn.v_proj.lora_A.default.weight": "model-00004-of-00030.safetensors", - "model.layers.8.self_attn.v_proj.lora_B.default.weight": "model-00004-of-00030.safetensors", - "model.layers.8.self_attn.v_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.9.input_layernorm.weight": "model-00004-of-00030.safetensors", - "model.layers.9.mlp.down_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.9.mlp.gate_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.9.mlp.up_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.9.post_attention_layernorm.weight": "model-00004-of-00030.safetensors", - "model.layers.9.self_attn.k_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.9.self_attn.o_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.9.self_attn.q_proj.lora_A.default.weight": "model-00004-of-00030.safetensors", - "model.layers.9.self_attn.q_proj.lora_B.default.weight": "model-00004-of-00030.safetensors", - "model.layers.9.self_attn.q_proj.weight": "model-00004-of-00030.safetensors", - "model.layers.9.self_attn.v_proj.lora_A.default.weight": "model-00004-of-00030.safetensors", - "model.layers.9.self_attn.v_proj.lora_B.default.weight": "model-00004-of-00030.safetensors", - "model.layers.9.self_attn.v_proj.weight": "model-00004-of-00030.safetensors", - "model.norm.weight": "model-00030-of-00030.safetensors" + "lm_head.weight": "model-00061-of-00061.safetensors", + "model.embed_tokens.weight": "model-00001-of-00061.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.q_proj.lora_A.default.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.q_proj.lora_B.default.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.v_proj.lora_A.default.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.v_proj.lora_B.default.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00061.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.self_attn.q_proj.lora_A.default.weight": "model-00001-of-00061.safetensors", + "model.layers.1.self_attn.q_proj.lora_B.default.weight": "model-00001-of-00061.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.1.self_attn.v_proj.lora_A.default.weight": "model-00001-of-00061.safetensors", + "model.layers.1.self_attn.v_proj.lora_B.default.weight": "model-00001-of-00061.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.10.input_layernorm.weight": "model-00009-of-00061.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00061.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.q_proj.lora_A.default.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.q_proj.lora_B.default.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.v_proj.lora_A.default.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.v_proj.lora_B.default.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.11.input_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.q_proj.lora_A.default.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.q_proj.lora_B.default.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.v_proj.lora_A.default.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.v_proj.lora_B.default.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.12.input_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.q_proj.lora_A.default.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.q_proj.lora_B.default.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.v_proj.lora_A.default.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.v_proj.lora_B.default.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.input_layernorm.weight": "model-00011-of-00061.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00061.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.q_proj.lora_A.default.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.q_proj.lora_B.default.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.v_proj.lora_A.default.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.v_proj.lora_B.default.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.14.input_layernorm.weight": "model-00012-of-00061.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00012-of-00061.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.q_proj.lora_A.default.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.q_proj.lora_B.default.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.v_proj.lora_A.default.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.v_proj.lora_B.default.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.15.input_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.q_proj.lora_A.default.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.q_proj.lora_B.default.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.v_proj.lora_A.default.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.v_proj.lora_B.default.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.16.input_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.q_proj.lora_A.default.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.q_proj.lora_B.default.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.v_proj.lora_A.default.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.v_proj.lora_B.default.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.input_layernorm.weight": "model-00014-of-00061.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00014-of-00061.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.q_proj.lora_A.default.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.q_proj.lora_B.default.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.v_proj.lora_A.default.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.v_proj.lora_B.default.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.18.input_layernorm.weight": "model-00015-of-00061.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00015-of-00061.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.q_proj.lora_A.default.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.q_proj.lora_B.default.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.v_proj.lora_A.default.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.v_proj.lora_B.default.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.19.input_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.q_proj.lora_A.default.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.q_proj.lora_B.default.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.v_proj.lora_A.default.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.v_proj.lora_B.default.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00061.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00061.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.q_proj.lora_A.default.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.q_proj.lora_B.default.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.v_proj.lora_A.default.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.v_proj.lora_B.default.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.20.input_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.q_proj.lora_A.default.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.q_proj.lora_B.default.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.v_proj.lora_A.default.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.v_proj.lora_B.default.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.input_layernorm.weight": "model-00017-of-00061.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00017-of-00061.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.q_proj.lora_A.default.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.q_proj.lora_B.default.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.v_proj.lora_A.default.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.v_proj.lora_B.default.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.22.input_layernorm.weight": "model-00018-of-00061.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00061.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.q_proj.lora_A.default.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.q_proj.lora_B.default.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.v_proj.lora_A.default.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.v_proj.lora_B.default.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.23.input_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.q_proj.lora_A.default.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.q_proj.lora_B.default.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.v_proj.lora_A.default.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.v_proj.lora_B.default.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.24.input_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.q_proj.lora_A.default.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.q_proj.lora_B.default.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.v_proj.lora_A.default.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.v_proj.lora_B.default.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.input_layernorm.weight": "model-00020-of-00061.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00061.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.q_proj.lora_A.default.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.q_proj.lora_B.default.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.v_proj.lora_A.default.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.v_proj.lora_B.default.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.26.input_layernorm.weight": "model-00021-of-00061.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00061.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.q_proj.lora_A.default.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.q_proj.lora_B.default.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.v_proj.lora_A.default.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.v_proj.lora_B.default.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.27.input_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.q_proj.lora_A.default.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.q_proj.lora_B.default.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.v_proj.lora_A.default.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.v_proj.lora_B.default.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.28.input_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.q_proj.lora_A.default.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.q_proj.lora_B.default.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.v_proj.lora_A.default.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.v_proj.lora_B.default.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.input_layernorm.weight": "model-00023-of-00061.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00023-of-00061.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.q_proj.lora_A.default.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.q_proj.lora_B.default.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.v_proj.lora_A.default.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.v_proj.lora_B.default.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.q_proj.lora_A.default.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.q_proj.lora_B.default.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.v_proj.lora_A.default.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.v_proj.lora_B.default.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.30.input_layernorm.weight": "model-00024-of-00061.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00024-of-00061.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.q_proj.lora_A.default.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.q_proj.lora_B.default.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.v_proj.lora_A.default.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.v_proj.lora_B.default.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.31.input_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.q_proj.lora_A.default.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.q_proj.lora_B.default.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.v_proj.lora_A.default.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.v_proj.lora_B.default.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.32.input_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.q_proj.lora_A.default.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.q_proj.lora_B.default.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.v_proj.lora_A.default.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.v_proj.lora_B.default.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.input_layernorm.weight": "model-00026-of-00061.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00026-of-00061.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.q_proj.lora_A.default.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.q_proj.lora_B.default.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.v_proj.lora_A.default.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.v_proj.lora_B.default.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.34.input_layernorm.weight": "model-00027-of-00061.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00027-of-00061.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.q_proj.lora_A.default.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.q_proj.lora_B.default.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.v_proj.lora_A.default.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.v_proj.lora_B.default.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.35.input_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.q_proj.lora_A.default.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.q_proj.lora_B.default.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.v_proj.lora_A.default.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.v_proj.lora_B.default.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.36.input_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.q_proj.lora_A.default.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.q_proj.lora_B.default.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.v_proj.lora_A.default.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.v_proj.lora_B.default.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.input_layernorm.weight": "model-00029-of-00061.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00029-of-00061.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.q_proj.lora_A.default.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.q_proj.lora_B.default.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.v_proj.lora_A.default.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.v_proj.lora_B.default.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.38.input_layernorm.weight": "model-00030-of-00061.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00030-of-00061.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.q_proj.lora_A.default.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.q_proj.lora_B.default.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.v_proj.lora_A.default.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.v_proj.lora_B.default.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.39.input_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.q_proj.lora_A.default.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.q_proj.lora_B.default.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.v_proj.lora_A.default.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.v_proj.lora_B.default.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.4.input_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.q_proj.lora_A.default.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.q_proj.lora_B.default.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.v_proj.lora_A.default.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.v_proj.lora_B.default.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.40.input_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.q_proj.lora_A.default.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.q_proj.lora_B.default.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.v_proj.lora_A.default.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.v_proj.lora_B.default.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.input_layernorm.weight": "model-00032-of-00061.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00032-of-00061.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.q_proj.lora_A.default.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.q_proj.lora_B.default.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.v_proj.lora_A.default.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.v_proj.lora_B.default.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.42.input_layernorm.weight": "model-00033-of-00061.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00033-of-00061.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.q_proj.lora_A.default.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.q_proj.lora_B.default.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.v_proj.lora_A.default.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.v_proj.lora_B.default.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.43.input_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.q_proj.lora_A.default.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.q_proj.lora_B.default.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.v_proj.lora_A.default.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.v_proj.lora_B.default.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.44.input_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.q_proj.lora_A.default.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.q_proj.lora_B.default.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.v_proj.lora_A.default.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.v_proj.lora_B.default.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.input_layernorm.weight": "model-00035-of-00061.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00035-of-00061.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.q_proj.lora_A.default.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.q_proj.lora_B.default.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.v_proj.lora_A.default.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.v_proj.lora_B.default.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.46.input_layernorm.weight": "model-00036-of-00061.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00036-of-00061.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.q_proj.lora_A.default.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.q_proj.lora_B.default.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.v_proj.lora_A.default.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.v_proj.lora_B.default.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.47.input_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.q_proj.lora_A.default.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.q_proj.lora_B.default.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.v_proj.lora_A.default.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.v_proj.lora_B.default.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.48.input_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.q_proj.lora_A.default.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.q_proj.lora_B.default.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.v_proj.lora_A.default.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.v_proj.lora_B.default.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.input_layernorm.weight": "model-00038-of-00061.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00038-of-00061.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.q_proj.lora_A.default.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.q_proj.lora_B.default.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.v_proj.lora_A.default.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.v_proj.lora_B.default.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.5.input_layernorm.weight": "model-00005-of-00061.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00061.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.q_proj.lora_A.default.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.q_proj.lora_B.default.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.v_proj.lora_A.default.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.v_proj.lora_B.default.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.50.input_layernorm.weight": "model-00039-of-00061.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00039-of-00061.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.q_proj.lora_A.default.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.q_proj.lora_B.default.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.v_proj.lora_A.default.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.v_proj.lora_B.default.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.51.input_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.q_proj.lora_A.default.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.q_proj.lora_B.default.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.v_proj.lora_A.default.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.v_proj.lora_B.default.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.52.input_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.q_proj.lora_A.default.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.q_proj.lora_B.default.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.v_proj.lora_A.default.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.v_proj.lora_B.default.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.input_layernorm.weight": "model-00041-of-00061.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00041-of-00061.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.q_proj.lora_A.default.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.q_proj.lora_B.default.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.v_proj.lora_A.default.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.v_proj.lora_B.default.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.54.input_layernorm.weight": "model-00042-of-00061.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00042-of-00061.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.q_proj.lora_A.default.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.q_proj.lora_B.default.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.v_proj.lora_A.default.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.v_proj.lora_B.default.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.55.input_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.q_proj.lora_A.default.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.q_proj.lora_B.default.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.v_proj.lora_A.default.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.v_proj.lora_B.default.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.56.input_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.q_proj.lora_A.default.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.q_proj.lora_B.default.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.v_proj.lora_A.default.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.v_proj.lora_B.default.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.input_layernorm.weight": "model-00044-of-00061.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00044-of-00061.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.q_proj.lora_A.default.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.q_proj.lora_B.default.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.v_proj.lora_A.default.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.v_proj.lora_B.default.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.58.input_layernorm.weight": "model-00045-of-00061.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00045-of-00061.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.q_proj.lora_A.default.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.q_proj.lora_B.default.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.v_proj.lora_A.default.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.v_proj.lora_B.default.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.59.input_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.q_proj.lora_A.default.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.q_proj.lora_B.default.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.v_proj.lora_A.default.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.v_proj.lora_B.default.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.6.input_layernorm.weight": "model-00006-of-00061.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00061.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.q_proj.lora_A.default.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.q_proj.lora_B.default.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.v_proj.lora_A.default.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.v_proj.lora_B.default.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.60.input_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.q_proj.lora_A.default.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.q_proj.lora_B.default.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.v_proj.lora_A.default.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.v_proj.lora_B.default.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.input_layernorm.weight": "model-00047-of-00061.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00047-of-00061.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.q_proj.lora_A.default.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.q_proj.lora_B.default.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.v_proj.lora_A.default.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.v_proj.lora_B.default.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.62.input_layernorm.weight": "model-00048-of-00061.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00048-of-00061.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.q_proj.lora_A.default.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.q_proj.lora_B.default.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.v_proj.lora_A.default.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.v_proj.lora_B.default.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.63.input_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.q_proj.lora_A.default.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.q_proj.lora_B.default.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.v_proj.lora_A.default.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.v_proj.lora_B.default.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.64.input_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.q_proj.lora_A.default.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.q_proj.lora_B.default.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.v_proj.lora_A.default.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.v_proj.lora_B.default.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.input_layernorm.weight": "model-00050-of-00061.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00050-of-00061.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.q_proj.lora_A.default.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.q_proj.lora_B.default.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.v_proj.lora_A.default.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.v_proj.lora_B.default.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.66.input_layernorm.weight": "model-00051-of-00061.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00051-of-00061.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.q_proj.lora_A.default.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.q_proj.lora_B.default.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.v_proj.lora_A.default.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.v_proj.lora_B.default.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.67.input_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.q_proj.lora_A.default.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.q_proj.lora_B.default.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.v_proj.lora_A.default.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.v_proj.lora_B.default.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.68.input_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.q_proj.lora_A.default.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.q_proj.lora_B.default.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.v_proj.lora_A.default.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.v_proj.lora_B.default.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.input_layernorm.weight": "model-00053-of-00061.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00053-of-00061.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.q_proj.lora_A.default.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.q_proj.lora_B.default.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.v_proj.lora_A.default.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.v_proj.lora_B.default.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.7.input_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.q_proj.lora_A.default.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.q_proj.lora_B.default.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.v_proj.lora_A.default.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.v_proj.lora_B.default.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.70.input_layernorm.weight": "model-00054-of-00061.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00054-of-00061.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.q_proj.lora_A.default.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.q_proj.lora_B.default.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.v_proj.lora_A.default.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.v_proj.lora_B.default.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.71.input_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.q_proj.lora_A.default.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.q_proj.lora_B.default.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.v_proj.lora_A.default.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.v_proj.lora_B.default.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.72.input_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.q_proj.lora_A.default.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.q_proj.lora_B.default.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.v_proj.lora_A.default.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.v_proj.lora_B.default.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.input_layernorm.weight": "model-00056-of-00061.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00056-of-00061.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.q_proj.lora_A.default.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.q_proj.lora_B.default.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.v_proj.lora_A.default.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.v_proj.lora_B.default.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.74.input_layernorm.weight": "model-00057-of-00061.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00057-of-00061.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.q_proj.lora_A.default.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.q_proj.lora_B.default.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.v_proj.lora_A.default.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.v_proj.lora_B.default.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.75.input_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.q_proj.lora_A.default.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.q_proj.lora_B.default.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.v_proj.lora_A.default.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.v_proj.lora_B.default.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.76.input_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.q_proj.lora_A.default.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.q_proj.lora_B.default.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.v_proj.lora_A.default.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.v_proj.lora_B.default.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.input_layernorm.weight": "model-00059-of-00061.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00059-of-00061.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.q_proj.lora_A.default.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.q_proj.lora_B.default.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.v_proj.lora_A.default.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.v_proj.lora_B.default.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.78.input_layernorm.weight": "model-00060-of-00061.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00060-of-00061.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.q_proj.lora_A.default.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.q_proj.lora_B.default.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.v_proj.lora_A.default.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.v_proj.lora_B.default.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.79.input_layernorm.weight": "model-00061-of-00061.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00061-of-00061.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00061.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.q_proj.lora_A.default.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.q_proj.lora_B.default.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.v_proj.lora_A.default.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.v_proj.lora_B.default.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.8.input_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.q_proj.lora_A.default.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.q_proj.lora_B.default.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.v_proj.lora_A.default.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.v_proj.lora_B.default.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.input_layernorm.weight": "model-00008-of-00061.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00061.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.q_proj.lora_A.default.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.q_proj.lora_B.default.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.v_proj.lora_A.default.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.v_proj.lora_B.default.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00061.safetensors", + "model.norm.weight": "model-00061-of-00061.safetensors" } }