Upload folder using huggingface_hub (#1)
Browse files- 94af3249ee7d2f6b8be2c389a2a532e25abd3e942df5aa49194f06624d6ec7df (df15c8e7968db4a15de24735690f4a788b59ee08)
- 62ec0adb2e6eb61b84bf5753b8d55937675316bafc0e2fdf25d11d6795dfa25e (82f99016ca11306c72cf7870c5b47973a7025f84)
- 412d6cdaa4d4d74e8381bb114a41f7c6801177a2eb8de0cedd99f6dd1a6dbb6d (6c7dab0ec7726f28301dc8bbfa560ff13fdf7004)
- 848585d11936c4ae112bf93bdb59e5c759aa2a948a29f5b348a21692483ded0d (31c80b91151cec920f1e1e84442511cc0471a8be)
- b6261dfaf7c3a8aeabf16b4204d3fe0ab6a2973320e9698aa480f4c4fefc0bc8 (dc4a1e2d9e3026ae69a7331847617986f2b06484)
- b1190a762b308fca274cd5b94f07c056b291b57eb45e64c23c9f3eb0cb2eccb2 (e0cd4ed8bb9b1de55367bacd186b440055e3bc9c)
- 5c13dc3596290b66b847eb68af8eee320d5cdbedecd932ced9db3676f12116dd (7f5e4c9e784b130b829c0c3b06266221e6b80172)
- 63dc6b21fa0cb02b631da79d5c11168ccc310d4a7d03a71af390090340092d3c (4b295ca4ce303037c72b736676ac2627ebc91595)
- ebfd399034a4ddc90b959eb358013d1247f31ec237ee8b63fdf29bfaa0d440b5 (01ad1d15d43903ff0abd112ae43cec38e397663a)
- eb0a6533b36e84c8b2e7a36ebf5e5923b543c2a7653d628976333cbdc9e80aa0 (a6a3a3553a3406433b1adaa92c696fc41d3bec15)
- c20665c5f23317785cec4c9d0f7da76320d5236daba6c1a221334726d790eadd (1794d11175fcd29d7d50f1a897b64cfd9fab0137)
- ce858e8dc9990be378e9eb741afd4893a9a815a0e7f001cf9fbbc7a7a78984e9 (d1c1bd8260cb043c6c7ac7c6a295f49f3a42d7f9)
- ba3c47abdaaf7b18127dbedabac08e41a5615f84313f6e51150a6eb38ab4ec92 (1d4444e8c62513a7db86aa98c58d43dd86f2c250)
- 2a1440f1980ef647bc431f7d01a7fecd4311d733f161bad6670c4c06b3f93b02 (fbf5e925072bb59236fad297b687cfaad63bebec)
- 121a0045e764e701c2a4bb1b67be01fe422c5935e71f312e7cffdddf8c3ce91e (caebecd7f93cd533d7f519031cd0214d93e5d428)
- 00684905142fd5ca6093cae55cde0a0bf816a296294347cc63d59edfb648aed1 (f7e9ab61d55afdadd4ad9138a9f0f50da21c94bc)
- f2fca6496e27649572bb3d50976660aa6e1212bb29517e8bae7ea05766b57ad8 (40c9eeb13d757524bc0ba8e264ec44009e3a6ea0)
- 67271d62879cecdfd5ebe7180421781a589d111a944ae6915c018e1f2400884d (7c7cd95f70aae7657a1b8de6fc39ac09b1cce869)
- 3997b6ff6c69cb8c4fb2e8595c5d9aeb4cb813a9c26c482ad479d016f7d7e89b (ce4de135747ddf5e32415b6dc224d765b4e5ee2e)
- 09e08699082b878ed8d9a54acf2762db31c24d7edfa266a20640c042056ce41c (def6b545391b37f6e7d1eca6fa9a4d09f6114805)
- 9c2d4c0e364dbca0fff46e6fcc881682d7de168f3918563849a7fcf712d63ec0 (0a18b9779a44f50c0f9ed00cc4cb7e9dd5fd39de)
- 901c3c78b7f74b4022cd2745c381b08924dee6d5bb487b4eca4fe6ba382e8208 (7acd1f7688a10057cc3e2dcb68e710598a51d073)
- 8004e80664ae55918df6a582838a375c00a08fcd2d02a1d126f39c3a0742a36d (67aa82f8bfc304e58a2c17cdcf94970662d9567f)
- f8949fda5a613104702af6da5971c96c7fb3edd8c60d3eace9a6d55643480cce (3a03090755c5346ee85df97364d3a1bd736083b7)
- ebb0eb0f116d8704168107e0dc6b4448e1345782dee8d61d0a64cc46b0799466 (9fbe051ffbdc5e7bbd215575df943022018de970)
- 8785f45fc040fe79bd5cf21743c792e45d78f3197270e06da2d8c45e20533413 (698c045206713048d62b2c4d718b6a6ba99756ef)
- 741bdd74d3f2a553d717a94b1f0cd884ea959819cc455a33747ec14224ceace8 (1e9ced62984a72df54e2ad4f713e0a18c8b5ed68)
- ae44b464198fb42818814052f3f3b3881a7b33bbc5ebe633c1a71df83401fe38 (75d8e2058fcdd70fa913b1b28086e2fb9e8f86d4)
- 179c55eaa9f2e961752ab2fa32c63faa05f1e161396368566e663d641e9b2e95 (8345d3c8ab52c5fd8f25c4d1e9fd7f8baaddfc60)
- e2957fbd388fac7a32132f0e1b55975cbb769ba39f85a3e0db65af930604a609 (832219a75f6e30933a429bd59d0ddd7f2ffcd7d8)
- 6705812196fc694d4e1e9cf438d0e8e67ead54ecf0f0af5bdf5bf8e8f1e0eea1 (4d5629fe2ab00c24555feaa8793388d29463a265)
- 52837442612124c9c9bb5233b20375f107744a77620856b6fc0f1a60cc2b8c85 (7e871ffdb471d4cf98676d3ba288245692e12464)
- 13cafd840b0f1f35d94e8f743b29819b9b6763093851cd0a46a91aa1432e23be (a33e51a91c9f009a9d8fcb6f7bc107b202b67995)
- 3350024eeabd99f4dccc2c3824fc11da46ef3284199de7247fd1a079621696d7 (714c0a61e55148e9abeb0c26b00bb790e73ad78c)
- c9c9a18ead97df737c71124e5fcd02ac6e3360d0cf917c8a87e328f9c9fa1cc9 (e101b0649cd2ca454121c5c350aa85f04be16287)
- a103c940c961368b32ee2b306570bea3c6f02f223e3057d7a5e1ffb85412c700 (2b39a6facf0c93c37aa27b6a69ba31849b415f65)
- 372638ea9803605831171cac9add101fa21e78eb828b5d6c0cb5f794863e02e6 (b5436a2e078fe263ecef51ebc6dd8df2c21e557a)
- 039241a347695a6cbe01646105f544b8fafdf2c6b57a948bebe02b3b242e70c7 (b5ab169dbb361605c2956d14c90d38dfe16a24eb)
- d3464a3dda25bed02f33528d17eefa7a6174e2f310a42ff51cf081e59e18d00d (442f59e7e8208c650d6bcb2fb8a5e1b52d753f6b)
- 5860194d83aa6c31789555105659e6641529d8bf2ab998da9454a910a9ae2a2a (9f5ff064f5c0a48845b5b50b334202c81c1f39d1)
- fb8112f30fe67d8ed6127d87a546eaa5e2f4fcb41b1e60b0b615aaa13b75df20 (3f907711104336280aee10d3c92b400113b24115)
- 7399f9197a0ed39a9ae6fdffea87ef092d309c6f27dc096895929f20a32ab94f (c82f83089476a1ebef4bea363ed74562ae744140)
- ff1cf3a250ca5542cbeba0177fa3ff0c8e3cd8b106bce020d561bcc50c5255e7 (ebbb508ed49694f5a73451bd1c302e6005002777)
- b7340f108dfd263e1457bdf8cbf0747d97bf50e221a1b00ad9f6e066d1516ca3 (fee1e0f7582e1058d977d3c13f66a4ac5cfc5aaf)
- c64d4878e5b08398e3220973cfbcf653d51c7723c7945afaa5ac3189a7f866b3 (2335ab2c7a187d1bcc66df3ea354d098c0a2db10)
- fbe73ebc9846c7f284815109a4d5078243d7317feeaf9b18e4c1f892f565117f (0bd814d8a989082a4ecd8b451af509941d1c840b)
- cd5a0c152656f885ef2e0b70b9ef200ab1c1472f4d928e94b7a3098ac804b6bc (7b217ed8ea096c79bc41b61acb5938a259f38cad)
- a8cadbf0fe87dd61e102196e337da7b1a6e25ee009025ac6b159a583972cb55a (97c86e9dbc0be8598fbfa21a00e426529eb3df49)
- ac825f5998dec81cb71466d4bbc5c03cc83db88a27a9e870998928d45c245e58 (8cdcb62b0b3cfeb7abdd0d1e331c220deecd1dea)
- d5a8550526c35940bfcef1f0fb3af4065b96b847115e02c7d268ace2c98c0af3 (a73576e9b03235b4c4c657e6484eb5a68dfd7e93)
- fba5c2437f8af7d91bf3c30af0e9776747d0888ce69dcd3edd81c1097e7474e0 (7a2a960c2f1411cf506bf9181d1e900b48273e04)
- 90bb393b644f27c80abba53566ee0c8325ade8e86a08efc8b1f815a4fe6e8cdd (8948e27f1942cb9237360feb40f37061c2ad49a9)
- 3648ab95925cf7ea5ecba4525589ba464f4c321beca1f47b15154f075ba87fa1 (101e11c5a860fc7e8346c4da5f7bda4088ffbb72)
- fe904876f884f632ddcd034acec523bfb870b804775c3ed1a40389cd9dc67a7d (e8aae75634eb0b35aa811ecfad83cd123bec3c0e)
- 0024f1ce915d6e407473592476b05038f8cdb0709d2a4aa331c9f30a00d9e487 (92efb423c37759fd50aa0aea1db9e4c12f3aac98)
- 5998b27b46112b6ce6f2df515c9a1b9d212e3e75ffb7c35966680964b85a449d (a4a5d884f06a5cecee1395ff7a2d89d3283d646c)
- 61e15c2eff9b3564b8f076bcb650b9e9847d7aabc386eda7d77f10e8c49f1136 (856f0b2b9826d2b51bf171ea647d92e90fca715f)
- 3a2f406c7c2cc7de0cf4a198623824c979809a74ec4f26dad2aa47189b42215f (bb381012b149bd75e965712581ce299c0f544612)
- de3669e9b655c76dbc22feeee91d4caea1fbf028c9aacb565d435fdf2319c5d5 (8005c4cc90add8c3811683b2a55f0bb411165bc4)
- f0868b81529d08b281ef5b27f1d312282532a034d546729fd5d11129b47a1a51 (d4ba2af71396c2e88dcdcc7591e89b7b197d0d0d)
- 76d9822d3fe37ddc05aeede0658bd020de5c5917050cde9098ec4db175c12956 (1f64b191bbbca53912f938ed10b67e866c54a59d)
- .ipynb_checkpoints/config-checkpoint.json +29 -0
 - .ipynb_checkpoints/generation_config-checkpoint.json +8 -0
 - .ipynb_checkpoints/model.safetensors.index-checkpoint.json +730 -0
 - .ipynb_checkpoints/special_tokens_map-checkpoint.json +24 -0
 - .ipynb_checkpoints/tokenizer-checkpoint.json +0 -0
 - .ipynb_checkpoints/tokenizer_config-checkpoint.json +40 -0
 - config.json +29 -0
 - generation_config.json +8 -0
 - model-00001-of-00061.safetensors +3 -0
 - model-00002-of-00061.safetensors +3 -0
 - model-00003-of-00061.safetensors +3 -0
 - model-00004-of-00061.safetensors +3 -0
 - model-00005-of-00061.safetensors +3 -0
 - model-00006-of-00061.safetensors +3 -0
 - model-00007-of-00061.safetensors +3 -0
 - model-00008-of-00061.safetensors +3 -0
 - model-00009-of-00061.safetensors +3 -0
 - model-00010-of-00061.safetensors +3 -0
 - model-00011-of-00061.safetensors +3 -0
 - model-00012-of-00061.safetensors +3 -0
 - model-00013-of-00061.safetensors +3 -0
 - model-00014-of-00061.safetensors +3 -0
 - model-00015-of-00061.safetensors +3 -0
 - model-00016-of-00061.safetensors +3 -0
 - model-00017-of-00061.safetensors +3 -0
 - model-00018-of-00061.safetensors +3 -0
 - model-00019-of-00061.safetensors +3 -0
 - model-00020-of-00061.safetensors +3 -0
 - model-00021-of-00061.safetensors +3 -0
 - model-00022-of-00061.safetensors +3 -0
 - model-00023-of-00061.safetensors +3 -0
 - model-00024-of-00061.safetensors +3 -0
 - model-00025-of-00061.safetensors +3 -0
 - model-00026-of-00061.safetensors +3 -0
 - model-00027-of-00061.safetensors +3 -0
 - model-00028-of-00061.safetensors +3 -0
 - model-00029-of-00061.safetensors +3 -0
 - model-00030-of-00061.safetensors +3 -0
 - model-00031-of-00061.safetensors +3 -0
 - model-00032-of-00061.safetensors +3 -0
 - model-00033-of-00061.safetensors +3 -0
 - model-00034-of-00061.safetensors +3 -0
 - model-00035-of-00061.safetensors +3 -0
 - model-00036-of-00061.safetensors +3 -0
 - model-00037-of-00061.safetensors +3 -0
 - model-00038-of-00061.safetensors +3 -0
 - model-00039-of-00061.safetensors +3 -0
 - model-00040-of-00061.safetensors +3 -0
 - model-00041-of-00061.safetensors +3 -0
 - model-00042-of-00061.safetensors +3 -0
 
| 
         @@ -0,0 +1,29 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "_name_or_path": "sequelbox/SpellBlade",
         
     | 
| 3 | 
         
            +
              "architectures": [
         
     | 
| 4 | 
         
            +
                "LlamaForCausalLM"
         
     | 
| 5 | 
         
            +
              ],
         
     | 
| 6 | 
         
            +
              "attention_bias": false,
         
     | 
| 7 | 
         
            +
              "attention_dropout": 0.0,
         
     | 
| 8 | 
         
            +
              "bos_token_id": 1,
         
     | 
| 9 | 
         
            +
              "eos_token_id": 2,
         
     | 
| 10 | 
         
            +
              "hidden_act": "silu",
         
     | 
| 11 | 
         
            +
              "hidden_size": 8192,
         
     | 
| 12 | 
         
            +
              "initializer_range": 0.02,
         
     | 
| 13 | 
         
            +
              "intermediate_size": 28672,
         
     | 
| 14 | 
         
            +
              "max_position_embeddings": 4096,
         
     | 
| 15 | 
         
            +
              "model_type": "llama",
         
     | 
| 16 | 
         
            +
              "num_attention_heads": 64,
         
     | 
| 17 | 
         
            +
              "num_hidden_layers": 80,
         
     | 
| 18 | 
         
            +
              "num_key_value_heads": 8,
         
     | 
| 19 | 
         
            +
              "pad_token_id": 0,
         
     | 
| 20 | 
         
            +
              "pretraining_tp": 1,
         
     | 
| 21 | 
         
            +
              "rms_norm_eps": 1e-05,
         
     | 
| 22 | 
         
            +
              "rope_scaling": null,
         
     | 
| 23 | 
         
            +
              "rope_theta": 10000.0,
         
     | 
| 24 | 
         
            +
              "tie_word_embeddings": false,
         
     | 
| 25 | 
         
            +
              "torch_dtype": "float32",
         
     | 
| 26 | 
         
            +
              "transformers_version": "4.36.2",
         
     | 
| 27 | 
         
            +
              "use_cache": false,
         
     | 
| 28 | 
         
            +
              "vocab_size": 32000
         
     | 
| 29 | 
         
            +
            }
         
     | 
| 
         @@ -0,0 +1,8 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "_from_model_config": true,
         
     | 
| 3 | 
         
            +
              "bos_token_id": 1,
         
     | 
| 4 | 
         
            +
              "eos_token_id": 2,
         
     | 
| 5 | 
         
            +
              "pad_token_id": 0,
         
     | 
| 6 | 
         
            +
              "transformers_version": "4.36.2",
         
     | 
| 7 | 
         
            +
              "use_cache": false
         
     | 
| 8 | 
         
            +
            }
         
     | 
| 
         @@ -0,0 +1,730 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "metadata": {
         
     | 
| 3 | 
         
            +
                "total_size": 275906592768
         
     | 
| 4 | 
         
            +
              },
         
     | 
| 5 | 
         
            +
              "weight_map": {
         
     | 
| 6 | 
         
            +
                "lm_head.weight": "model-00061-of-00061.safetensors",
         
     | 
| 7 | 
         
            +
                "model.embed_tokens.weight": "model-00001-of-00061.safetensors",
         
     | 
| 8 | 
         
            +
                "model.layers.0.input_layernorm.weight": "model-00001-of-00061.safetensors",
         
     | 
| 9 | 
         
            +
                "model.layers.0.mlp.down_proj.weight": "model-00001-of-00061.safetensors",
         
     | 
| 10 | 
         
            +
                "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00061.safetensors",
         
     | 
| 11 | 
         
            +
                "model.layers.0.mlp.up_proj.weight": "model-00001-of-00061.safetensors",
         
     | 
| 12 | 
         
            +
                "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00061.safetensors",
         
     | 
| 13 | 
         
            +
                "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00061.safetensors",
         
     | 
| 14 | 
         
            +
                "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00061.safetensors",
         
     | 
| 15 | 
         
            +
                "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00061.safetensors",
         
     | 
| 16 | 
         
            +
                "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00061.safetensors",
         
     | 
| 17 | 
         
            +
                "model.layers.1.input_layernorm.weight": "model-00002-of-00061.safetensors",
         
     | 
| 18 | 
         
            +
                "model.layers.1.mlp.down_proj.weight": "model-00002-of-00061.safetensors",
         
     | 
| 19 | 
         
            +
                "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00061.safetensors",
         
     | 
| 20 | 
         
            +
                "model.layers.1.mlp.up_proj.weight": "model-00002-of-00061.safetensors",
         
     | 
| 21 | 
         
            +
                "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00061.safetensors",
         
     | 
| 22 | 
         
            +
                "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00061.safetensors",
         
     | 
| 23 | 
         
            +
                "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00061.safetensors",
         
     | 
| 24 | 
         
            +
                "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00061.safetensors",
         
     | 
| 25 | 
         
            +
                "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00061.safetensors",
         
     | 
| 26 | 
         
            +
                "model.layers.10.input_layernorm.weight": "model-00009-of-00061.safetensors",
         
     | 
| 27 | 
         
            +
                "model.layers.10.mlp.down_proj.weight": "model-00009-of-00061.safetensors",
         
     | 
| 28 | 
         
            +
                "model.layers.10.mlp.gate_proj.weight": "model-00008-of-00061.safetensors",
         
     | 
| 29 | 
         
            +
                "model.layers.10.mlp.up_proj.weight": "model-00009-of-00061.safetensors",
         
     | 
| 30 | 
         
            +
                "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00061.safetensors",
         
     | 
| 31 | 
         
            +
                "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00061.safetensors",
         
     | 
| 32 | 
         
            +
                "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00061.safetensors",
         
     | 
| 33 | 
         
            +
                "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00061.safetensors",
         
     | 
| 34 | 
         
            +
                "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00061.safetensors",
         
     | 
| 35 | 
         
            +
                "model.layers.11.input_layernorm.weight": "model-00010-of-00061.safetensors",
         
     | 
| 36 | 
         
            +
                "model.layers.11.mlp.down_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 37 | 
         
            +
                "model.layers.11.mlp.gate_proj.weight": "model-00009-of-00061.safetensors",
         
     | 
| 38 | 
         
            +
                "model.layers.11.mlp.up_proj.weight": "model-00009-of-00061.safetensors",
         
     | 
| 39 | 
         
            +
                "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00061.safetensors",
         
     | 
| 40 | 
         
            +
                "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00061.safetensors",
         
     | 
| 41 | 
         
            +
                "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00061.safetensors",
         
     | 
| 42 | 
         
            +
                "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00061.safetensors",
         
     | 
| 43 | 
         
            +
                "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00061.safetensors",
         
     | 
| 44 | 
         
            +
                "model.layers.12.input_layernorm.weight": "model-00010-of-00061.safetensors",
         
     | 
| 45 | 
         
            +
                "model.layers.12.mlp.down_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 46 | 
         
            +
                "model.layers.12.mlp.gate_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 47 | 
         
            +
                "model.layers.12.mlp.up_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 48 | 
         
            +
                "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00061.safetensors",
         
     | 
| 49 | 
         
            +
                "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 50 | 
         
            +
                "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 51 | 
         
            +
                "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 52 | 
         
            +
                "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 53 | 
         
            +
                "model.layers.13.input_layernorm.weight": "model-00011-of-00061.safetensors",
         
     | 
| 54 | 
         
            +
                "model.layers.13.mlp.down_proj.weight": "model-00011-of-00061.safetensors",
         
     | 
| 55 | 
         
            +
                "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00061.safetensors",
         
     | 
| 56 | 
         
            +
                "model.layers.13.mlp.up_proj.weight": "model-00011-of-00061.safetensors",
         
     | 
| 57 | 
         
            +
                "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00061.safetensors",
         
     | 
| 58 | 
         
            +
                "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 59 | 
         
            +
                "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 60 | 
         
            +
                "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 61 | 
         
            +
                "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00061.safetensors",
         
     | 
| 62 | 
         
            +
                "model.layers.14.input_layernorm.weight": "model-00012-of-00061.safetensors",
         
     | 
| 63 | 
         
            +
                "model.layers.14.mlp.down_proj.weight": "model-00012-of-00061.safetensors",
         
     | 
| 64 | 
         
            +
                "model.layers.14.mlp.gate_proj.weight": "model-00011-of-00061.safetensors",
         
     | 
| 65 | 
         
            +
                "model.layers.14.mlp.up_proj.weight": "model-00012-of-00061.safetensors",
         
     | 
| 66 | 
         
            +
                "model.layers.14.post_attention_layernorm.weight": "model-00012-of-00061.safetensors",
         
     | 
| 67 | 
         
            +
                "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00061.safetensors",
         
     | 
| 68 | 
         
            +
                "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00061.safetensors",
         
     | 
| 69 | 
         
            +
                "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00061.safetensors",
         
     | 
| 70 | 
         
            +
                "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00061.safetensors",
         
     | 
| 71 | 
         
            +
                "model.layers.15.input_layernorm.weight": "model-00013-of-00061.safetensors",
         
     | 
| 72 | 
         
            +
                "model.layers.15.mlp.down_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 73 | 
         
            +
                "model.layers.15.mlp.gate_proj.weight": "model-00012-of-00061.safetensors",
         
     | 
| 74 | 
         
            +
                "model.layers.15.mlp.up_proj.weight": "model-00012-of-00061.safetensors",
         
     | 
| 75 | 
         
            +
                "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00061.safetensors",
         
     | 
| 76 | 
         
            +
                "model.layers.15.self_attn.k_proj.weight": "model-00012-of-00061.safetensors",
         
     | 
| 77 | 
         
            +
                "model.layers.15.self_attn.o_proj.weight": "model-00012-of-00061.safetensors",
         
     | 
| 78 | 
         
            +
                "model.layers.15.self_attn.q_proj.weight": "model-00012-of-00061.safetensors",
         
     | 
| 79 | 
         
            +
                "model.layers.15.self_attn.v_proj.weight": "model-00012-of-00061.safetensors",
         
     | 
| 80 | 
         
            +
                "model.layers.16.input_layernorm.weight": "model-00013-of-00061.safetensors",
         
     | 
| 81 | 
         
            +
                "model.layers.16.mlp.down_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 82 | 
         
            +
                "model.layers.16.mlp.gate_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 83 | 
         
            +
                "model.layers.16.mlp.up_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 84 | 
         
            +
                "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00061.safetensors",
         
     | 
| 85 | 
         
            +
                "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 86 | 
         
            +
                "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 87 | 
         
            +
                "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 88 | 
         
            +
                "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 89 | 
         
            +
                "model.layers.17.input_layernorm.weight": "model-00014-of-00061.safetensors",
         
     | 
| 90 | 
         
            +
                "model.layers.17.mlp.down_proj.weight": "model-00014-of-00061.safetensors",
         
     | 
| 91 | 
         
            +
                "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00061.safetensors",
         
     | 
| 92 | 
         
            +
                "model.layers.17.mlp.up_proj.weight": "model-00014-of-00061.safetensors",
         
     | 
| 93 | 
         
            +
                "model.layers.17.post_attention_layernorm.weight": "model-00014-of-00061.safetensors",
         
     | 
| 94 | 
         
            +
                "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 95 | 
         
            +
                "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 96 | 
         
            +
                "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 97 | 
         
            +
                "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00061.safetensors",
         
     | 
| 98 | 
         
            +
                "model.layers.18.input_layernorm.weight": "model-00015-of-00061.safetensors",
         
     | 
| 99 | 
         
            +
                "model.layers.18.mlp.down_proj.weight": "model-00015-of-00061.safetensors",
         
     | 
| 100 | 
         
            +
                "model.layers.18.mlp.gate_proj.weight": "model-00014-of-00061.safetensors",
         
     | 
| 101 | 
         
            +
                "model.layers.18.mlp.up_proj.weight": "model-00015-of-00061.safetensors",
         
     | 
| 102 | 
         
            +
                "model.layers.18.post_attention_layernorm.weight": "model-00015-of-00061.safetensors",
         
     | 
| 103 | 
         
            +
                "model.layers.18.self_attn.k_proj.weight": "model-00014-of-00061.safetensors",
         
     | 
| 104 | 
         
            +
                "model.layers.18.self_attn.o_proj.weight": "model-00014-of-00061.safetensors",
         
     | 
| 105 | 
         
            +
                "model.layers.18.self_attn.q_proj.weight": "model-00014-of-00061.safetensors",
         
     | 
| 106 | 
         
            +
                "model.layers.18.self_attn.v_proj.weight": "model-00014-of-00061.safetensors",
         
     | 
| 107 | 
         
            +
                "model.layers.19.input_layernorm.weight": "model-00016-of-00061.safetensors",
         
     | 
| 108 | 
         
            +
                "model.layers.19.mlp.down_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 109 | 
         
            +
                "model.layers.19.mlp.gate_proj.weight": "model-00015-of-00061.safetensors",
         
     | 
| 110 | 
         
            +
                "model.layers.19.mlp.up_proj.weight": "model-00015-of-00061.safetensors",
         
     | 
| 111 | 
         
            +
                "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00061.safetensors",
         
     | 
| 112 | 
         
            +
                "model.layers.19.self_attn.k_proj.weight": "model-00015-of-00061.safetensors",
         
     | 
| 113 | 
         
            +
                "model.layers.19.self_attn.o_proj.weight": "model-00015-of-00061.safetensors",
         
     | 
| 114 | 
         
            +
                "model.layers.19.self_attn.q_proj.weight": "model-00015-of-00061.safetensors",
         
     | 
| 115 | 
         
            +
                "model.layers.19.self_attn.v_proj.weight": "model-00015-of-00061.safetensors",
         
     | 
| 116 | 
         
            +
                "model.layers.2.input_layernorm.weight": "model-00003-of-00061.safetensors",
         
     | 
| 117 | 
         
            +
                "model.layers.2.mlp.down_proj.weight": "model-00003-of-00061.safetensors",
         
     | 
| 118 | 
         
            +
                "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00061.safetensors",
         
     | 
| 119 | 
         
            +
                "model.layers.2.mlp.up_proj.weight": "model-00003-of-00061.safetensors",
         
     | 
| 120 | 
         
            +
                "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00061.safetensors",
         
     | 
| 121 | 
         
            +
                "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00061.safetensors",
         
     | 
| 122 | 
         
            +
                "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00061.safetensors",
         
     | 
| 123 | 
         
            +
                "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00061.safetensors",
         
     | 
| 124 | 
         
            +
                "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00061.safetensors",
         
     | 
| 125 | 
         
            +
                "model.layers.20.input_layernorm.weight": "model-00016-of-00061.safetensors",
         
     | 
| 126 | 
         
            +
                "model.layers.20.mlp.down_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 127 | 
         
            +
                "model.layers.20.mlp.gate_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 128 | 
         
            +
                "model.layers.20.mlp.up_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 129 | 
         
            +
                "model.layers.20.post_attention_layernorm.weight": "model-00016-of-00061.safetensors",
         
     | 
| 130 | 
         
            +
                "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 131 | 
         
            +
                "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 132 | 
         
            +
                "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 133 | 
         
            +
                "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 134 | 
         
            +
                "model.layers.21.input_layernorm.weight": "model-00017-of-00061.safetensors",
         
     | 
| 135 | 
         
            +
                "model.layers.21.mlp.down_proj.weight": "model-00017-of-00061.safetensors",
         
     | 
| 136 | 
         
            +
                "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00061.safetensors",
         
     | 
| 137 | 
         
            +
                "model.layers.21.mlp.up_proj.weight": "model-00017-of-00061.safetensors",
         
     | 
| 138 | 
         
            +
                "model.layers.21.post_attention_layernorm.weight": "model-00017-of-00061.safetensors",
         
     | 
| 139 | 
         
            +
                "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 140 | 
         
            +
                "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 141 | 
         
            +
                "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 142 | 
         
            +
                "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00061.safetensors",
         
     | 
| 143 | 
         
            +
                "model.layers.22.input_layernorm.weight": "model-00018-of-00061.safetensors",
         
     | 
| 144 | 
         
            +
                "model.layers.22.mlp.down_proj.weight": "model-00018-of-00061.safetensors",
         
     | 
| 145 | 
         
            +
                "model.layers.22.mlp.gate_proj.weight": "model-00017-of-00061.safetensors",
         
     | 
| 146 | 
         
            +
                "model.layers.22.mlp.up_proj.weight": "model-00018-of-00061.safetensors",
         
     | 
| 147 | 
         
            +
                "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00061.safetensors",
         
     | 
| 148 | 
         
            +
                "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00061.safetensors",
         
     | 
| 149 | 
         
            +
                "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00061.safetensors",
         
     | 
| 150 | 
         
            +
                "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00061.safetensors",
         
     | 
| 151 | 
         
            +
                "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00061.safetensors",
         
     | 
| 152 | 
         
            +
                "model.layers.23.input_layernorm.weight": "model-00019-of-00061.safetensors",
         
     | 
| 153 | 
         
            +
                "model.layers.23.mlp.down_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 154 | 
         
            +
                "model.layers.23.mlp.gate_proj.weight": "model-00018-of-00061.safetensors",
         
     | 
| 155 | 
         
            +
                "model.layers.23.mlp.up_proj.weight": "model-00018-of-00061.safetensors",
         
     | 
| 156 | 
         
            +
                "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00061.safetensors",
         
     | 
| 157 | 
         
            +
                "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00061.safetensors",
         
     | 
| 158 | 
         
            +
                "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00061.safetensors",
         
     | 
| 159 | 
         
            +
                "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00061.safetensors",
         
     | 
| 160 | 
         
            +
                "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00061.safetensors",
         
     | 
| 161 | 
         
            +
                "model.layers.24.input_layernorm.weight": "model-00019-of-00061.safetensors",
         
     | 
| 162 | 
         
            +
                "model.layers.24.mlp.down_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 163 | 
         
            +
                "model.layers.24.mlp.gate_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 164 | 
         
            +
                "model.layers.24.mlp.up_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 165 | 
         
            +
                "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00061.safetensors",
         
     | 
| 166 | 
         
            +
                "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 167 | 
         
            +
                "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 168 | 
         
            +
                "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 169 | 
         
            +
                "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 170 | 
         
            +
                "model.layers.25.input_layernorm.weight": "model-00020-of-00061.safetensors",
         
     | 
| 171 | 
         
            +
                "model.layers.25.mlp.down_proj.weight": "model-00020-of-00061.safetensors",
         
     | 
| 172 | 
         
            +
                "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00061.safetensors",
         
     | 
| 173 | 
         
            +
                "model.layers.25.mlp.up_proj.weight": "model-00020-of-00061.safetensors",
         
     | 
| 174 | 
         
            +
                "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00061.safetensors",
         
     | 
| 175 | 
         
            +
                "model.layers.25.self_attn.k_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 176 | 
         
            +
                "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 177 | 
         
            +
                "model.layers.25.self_attn.q_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 178 | 
         
            +
                "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00061.safetensors",
         
     | 
| 179 | 
         
            +
                "model.layers.26.input_layernorm.weight": "model-00021-of-00061.safetensors",
         
     | 
| 180 | 
         
            +
                "model.layers.26.mlp.down_proj.weight": "model-00021-of-00061.safetensors",
         
     | 
| 181 | 
         
            +
                "model.layers.26.mlp.gate_proj.weight": "model-00020-of-00061.safetensors",
         
     | 
| 182 | 
         
            +
                "model.layers.26.mlp.up_proj.weight": "model-00021-of-00061.safetensors",
         
     | 
| 183 | 
         
            +
                "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00061.safetensors",
         
     | 
| 184 | 
         
            +
                "model.layers.26.self_attn.k_proj.weight": "model-00020-of-00061.safetensors",
         
     | 
| 185 | 
         
            +
                "model.layers.26.self_attn.o_proj.weight": "model-00020-of-00061.safetensors",
         
     | 
| 186 | 
         
            +
                "model.layers.26.self_attn.q_proj.weight": "model-00020-of-00061.safetensors",
         
     | 
| 187 | 
         
            +
                "model.layers.26.self_attn.v_proj.weight": "model-00020-of-00061.safetensors",
         
     | 
| 188 | 
         
            +
                "model.layers.27.input_layernorm.weight": "model-00022-of-00061.safetensors",
         
     | 
| 189 | 
         
            +
                "model.layers.27.mlp.down_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 190 | 
         
            +
                "model.layers.27.mlp.gate_proj.weight": "model-00021-of-00061.safetensors",
         
     | 
| 191 | 
         
            +
                "model.layers.27.mlp.up_proj.weight": "model-00021-of-00061.safetensors",
         
     | 
| 192 | 
         
            +
                "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00061.safetensors",
         
     | 
| 193 | 
         
            +
                "model.layers.27.self_attn.k_proj.weight": "model-00021-of-00061.safetensors",
         
     | 
| 194 | 
         
            +
                "model.layers.27.self_attn.o_proj.weight": "model-00021-of-00061.safetensors",
         
     | 
| 195 | 
         
            +
                "model.layers.27.self_attn.q_proj.weight": "model-00021-of-00061.safetensors",
         
     | 
| 196 | 
         
            +
                "model.layers.27.self_attn.v_proj.weight": "model-00021-of-00061.safetensors",
         
     | 
| 197 | 
         
            +
                "model.layers.28.input_layernorm.weight": "model-00022-of-00061.safetensors",
         
     | 
| 198 | 
         
            +
                "model.layers.28.mlp.down_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 199 | 
         
            +
                "model.layers.28.mlp.gate_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 200 | 
         
            +
                "model.layers.28.mlp.up_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 201 | 
         
            +
                "model.layers.28.post_attention_layernorm.weight": "model-00022-of-00061.safetensors",
         
     | 
| 202 | 
         
            +
                "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 203 | 
         
            +
                "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 204 | 
         
            +
                "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 205 | 
         
            +
                "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 206 | 
         
            +
                "model.layers.29.input_layernorm.weight": "model-00023-of-00061.safetensors",
         
     | 
| 207 | 
         
            +
                "model.layers.29.mlp.down_proj.weight": "model-00023-of-00061.safetensors",
         
     | 
| 208 | 
         
            +
                "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00061.safetensors",
         
     | 
| 209 | 
         
            +
                "model.layers.29.mlp.up_proj.weight": "model-00023-of-00061.safetensors",
         
     | 
| 210 | 
         
            +
                "model.layers.29.post_attention_layernorm.weight": "model-00023-of-00061.safetensors",
         
     | 
| 211 | 
         
            +
                "model.layers.29.self_attn.k_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 212 | 
         
            +
                "model.layers.29.self_attn.o_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 213 | 
         
            +
                "model.layers.29.self_attn.q_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 214 | 
         
            +
                "model.layers.29.self_attn.v_proj.weight": "model-00022-of-00061.safetensors",
         
     | 
| 215 | 
         
            +
                "model.layers.3.input_layernorm.weight": "model-00004-of-00061.safetensors",
         
     | 
| 216 | 
         
            +
                "model.layers.3.mlp.down_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 217 | 
         
            +
                "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00061.safetensors",
         
     | 
| 218 | 
         
            +
                "model.layers.3.mlp.up_proj.weight": "model-00003-of-00061.safetensors",
         
     | 
| 219 | 
         
            +
                "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00061.safetensors",
         
     | 
| 220 | 
         
            +
                "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00061.safetensors",
         
     | 
| 221 | 
         
            +
                "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00061.safetensors",
         
     | 
| 222 | 
         
            +
                "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00061.safetensors",
         
     | 
| 223 | 
         
            +
                "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00061.safetensors",
         
     | 
| 224 | 
         
            +
                "model.layers.30.input_layernorm.weight": "model-00024-of-00061.safetensors",
         
     | 
| 225 | 
         
            +
                "model.layers.30.mlp.down_proj.weight": "model-00024-of-00061.safetensors",
         
     | 
| 226 | 
         
            +
                "model.layers.30.mlp.gate_proj.weight": "model-00023-of-00061.safetensors",
         
     | 
| 227 | 
         
            +
                "model.layers.30.mlp.up_proj.weight": "model-00024-of-00061.safetensors",
         
     | 
| 228 | 
         
            +
                "model.layers.30.post_attention_layernorm.weight": "model-00024-of-00061.safetensors",
         
     | 
| 229 | 
         
            +
                "model.layers.30.self_attn.k_proj.weight": "model-00023-of-00061.safetensors",
         
     | 
| 230 | 
         
            +
                "model.layers.30.self_attn.o_proj.weight": "model-00023-of-00061.safetensors",
         
     | 
| 231 | 
         
            +
                "model.layers.30.self_attn.q_proj.weight": "model-00023-of-00061.safetensors",
         
     | 
| 232 | 
         
            +
                "model.layers.30.self_attn.v_proj.weight": "model-00023-of-00061.safetensors",
         
     | 
| 233 | 
         
            +
                "model.layers.31.input_layernorm.weight": "model-00025-of-00061.safetensors",
         
     | 
| 234 | 
         
            +
                "model.layers.31.mlp.down_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 235 | 
         
            +
                "model.layers.31.mlp.gate_proj.weight": "model-00024-of-00061.safetensors",
         
     | 
| 236 | 
         
            +
                "model.layers.31.mlp.up_proj.weight": "model-00024-of-00061.safetensors",
         
     | 
| 237 | 
         
            +
                "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00061.safetensors",
         
     | 
| 238 | 
         
            +
                "model.layers.31.self_attn.k_proj.weight": "model-00024-of-00061.safetensors",
         
     | 
| 239 | 
         
            +
                "model.layers.31.self_attn.o_proj.weight": "model-00024-of-00061.safetensors",
         
     | 
| 240 | 
         
            +
                "model.layers.31.self_attn.q_proj.weight": "model-00024-of-00061.safetensors",
         
     | 
| 241 | 
         
            +
                "model.layers.31.self_attn.v_proj.weight": "model-00024-of-00061.safetensors",
         
     | 
| 242 | 
         
            +
                "model.layers.32.input_layernorm.weight": "model-00025-of-00061.safetensors",
         
     | 
| 243 | 
         
            +
                "model.layers.32.mlp.down_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 244 | 
         
            +
                "model.layers.32.mlp.gate_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 245 | 
         
            +
                "model.layers.32.mlp.up_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 246 | 
         
            +
                "model.layers.32.post_attention_layernorm.weight": "model-00025-of-00061.safetensors",
         
     | 
| 247 | 
         
            +
                "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 248 | 
         
            +
                "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 249 | 
         
            +
                "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 250 | 
         
            +
                "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 251 | 
         
            +
                "model.layers.33.input_layernorm.weight": "model-00026-of-00061.safetensors",
         
     | 
| 252 | 
         
            +
                "model.layers.33.mlp.down_proj.weight": "model-00026-of-00061.safetensors",
         
     | 
| 253 | 
         
            +
                "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00061.safetensors",
         
     | 
| 254 | 
         
            +
                "model.layers.33.mlp.up_proj.weight": "model-00026-of-00061.safetensors",
         
     | 
| 255 | 
         
            +
                "model.layers.33.post_attention_layernorm.weight": "model-00026-of-00061.safetensors",
         
     | 
| 256 | 
         
            +
                "model.layers.33.self_attn.k_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 257 | 
         
            +
                "model.layers.33.self_attn.o_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 258 | 
         
            +
                "model.layers.33.self_attn.q_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 259 | 
         
            +
                "model.layers.33.self_attn.v_proj.weight": "model-00025-of-00061.safetensors",
         
     | 
| 260 | 
         
            +
                "model.layers.34.input_layernorm.weight": "model-00027-of-00061.safetensors",
         
     | 
| 261 | 
         
            +
                "model.layers.34.mlp.down_proj.weight": "model-00027-of-00061.safetensors",
         
     | 
| 262 | 
         
            +
                "model.layers.34.mlp.gate_proj.weight": "model-00026-of-00061.safetensors",
         
     | 
| 263 | 
         
            +
                "model.layers.34.mlp.up_proj.weight": "model-00027-of-00061.safetensors",
         
     | 
| 264 | 
         
            +
                "model.layers.34.post_attention_layernorm.weight": "model-00027-of-00061.safetensors",
         
     | 
| 265 | 
         
            +
                "model.layers.34.self_attn.k_proj.weight": "model-00026-of-00061.safetensors",
         
     | 
| 266 | 
         
            +
                "model.layers.34.self_attn.o_proj.weight": "model-00026-of-00061.safetensors",
         
     | 
| 267 | 
         
            +
                "model.layers.34.self_attn.q_proj.weight": "model-00026-of-00061.safetensors",
         
     | 
| 268 | 
         
            +
                "model.layers.34.self_attn.v_proj.weight": "model-00026-of-00061.safetensors",
         
     | 
| 269 | 
         
            +
                "model.layers.35.input_layernorm.weight": "model-00028-of-00061.safetensors",
         
     | 
| 270 | 
         
            +
                "model.layers.35.mlp.down_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 271 | 
         
            +
                "model.layers.35.mlp.gate_proj.weight": "model-00027-of-00061.safetensors",
         
     | 
| 272 | 
         
            +
                "model.layers.35.mlp.up_proj.weight": "model-00027-of-00061.safetensors",
         
     | 
| 273 | 
         
            +
                "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00061.safetensors",
         
     | 
| 274 | 
         
            +
                "model.layers.35.self_attn.k_proj.weight": "model-00027-of-00061.safetensors",
         
     | 
| 275 | 
         
            +
                "model.layers.35.self_attn.o_proj.weight": "model-00027-of-00061.safetensors",
         
     | 
| 276 | 
         
            +
                "model.layers.35.self_attn.q_proj.weight": "model-00027-of-00061.safetensors",
         
     | 
| 277 | 
         
            +
                "model.layers.35.self_attn.v_proj.weight": "model-00027-of-00061.safetensors",
         
     | 
| 278 | 
         
            +
                "model.layers.36.input_layernorm.weight": "model-00028-of-00061.safetensors",
         
     | 
| 279 | 
         
            +
                "model.layers.36.mlp.down_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 280 | 
         
            +
                "model.layers.36.mlp.gate_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 281 | 
         
            +
                "model.layers.36.mlp.up_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 282 | 
         
            +
                "model.layers.36.post_attention_layernorm.weight": "model-00028-of-00061.safetensors",
         
     | 
| 283 | 
         
            +
                "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 284 | 
         
            +
                "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 285 | 
         
            +
                "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 286 | 
         
            +
                "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 287 | 
         
            +
                "model.layers.37.input_layernorm.weight": "model-00029-of-00061.safetensors",
         
     | 
| 288 | 
         
            +
                "model.layers.37.mlp.down_proj.weight": "model-00029-of-00061.safetensors",
         
     | 
| 289 | 
         
            +
                "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00061.safetensors",
         
     | 
| 290 | 
         
            +
                "model.layers.37.mlp.up_proj.weight": "model-00029-of-00061.safetensors",
         
     | 
| 291 | 
         
            +
                "model.layers.37.post_attention_layernorm.weight": "model-00029-of-00061.safetensors",
         
     | 
| 292 | 
         
            +
                "model.layers.37.self_attn.k_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 293 | 
         
            +
                "model.layers.37.self_attn.o_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 294 | 
         
            +
                "model.layers.37.self_attn.q_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 295 | 
         
            +
                "model.layers.37.self_attn.v_proj.weight": "model-00028-of-00061.safetensors",
         
     | 
| 296 | 
         
            +
                "model.layers.38.input_layernorm.weight": "model-00030-of-00061.safetensors",
         
     | 
| 297 | 
         
            +
                "model.layers.38.mlp.down_proj.weight": "model-00030-of-00061.safetensors",
         
     | 
| 298 | 
         
            +
                "model.layers.38.mlp.gate_proj.weight": "model-00029-of-00061.safetensors",
         
     | 
| 299 | 
         
            +
                "model.layers.38.mlp.up_proj.weight": "model-00030-of-00061.safetensors",
         
     | 
| 300 | 
         
            +
                "model.layers.38.post_attention_layernorm.weight": "model-00030-of-00061.safetensors",
         
     | 
| 301 | 
         
            +
                "model.layers.38.self_attn.k_proj.weight": "model-00029-of-00061.safetensors",
         
     | 
| 302 | 
         
            +
                "model.layers.38.self_attn.o_proj.weight": "model-00029-of-00061.safetensors",
         
     | 
| 303 | 
         
            +
                "model.layers.38.self_attn.q_proj.weight": "model-00029-of-00061.safetensors",
         
     | 
| 304 | 
         
            +
                "model.layers.38.self_attn.v_proj.weight": "model-00029-of-00061.safetensors",
         
     | 
| 305 | 
         
            +
                "model.layers.39.input_layernorm.weight": "model-00031-of-00061.safetensors",
         
     | 
| 306 | 
         
            +
                "model.layers.39.mlp.down_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 307 | 
         
            +
                "model.layers.39.mlp.gate_proj.weight": "model-00030-of-00061.safetensors",
         
     | 
| 308 | 
         
            +
                "model.layers.39.mlp.up_proj.weight": "model-00030-of-00061.safetensors",
         
     | 
| 309 | 
         
            +
                "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00061.safetensors",
         
     | 
| 310 | 
         
            +
                "model.layers.39.self_attn.k_proj.weight": "model-00030-of-00061.safetensors",
         
     | 
| 311 | 
         
            +
                "model.layers.39.self_attn.o_proj.weight": "model-00030-of-00061.safetensors",
         
     | 
| 312 | 
         
            +
                "model.layers.39.self_attn.q_proj.weight": "model-00030-of-00061.safetensors",
         
     | 
| 313 | 
         
            +
                "model.layers.39.self_attn.v_proj.weight": "model-00030-of-00061.safetensors",
         
     | 
| 314 | 
         
            +
                "model.layers.4.input_layernorm.weight": "model-00004-of-00061.safetensors",
         
     | 
| 315 | 
         
            +
                "model.layers.4.mlp.down_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 316 | 
         
            +
                "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 317 | 
         
            +
                "model.layers.4.mlp.up_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 318 | 
         
            +
                "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00061.safetensors",
         
     | 
| 319 | 
         
            +
                "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 320 | 
         
            +
                "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 321 | 
         
            +
                "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 322 | 
         
            +
                "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 323 | 
         
            +
                "model.layers.40.input_layernorm.weight": "model-00031-of-00061.safetensors",
         
     | 
| 324 | 
         
            +
                "model.layers.40.mlp.down_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 325 | 
         
            +
                "model.layers.40.mlp.gate_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 326 | 
         
            +
                "model.layers.40.mlp.up_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 327 | 
         
            +
                "model.layers.40.post_attention_layernorm.weight": "model-00031-of-00061.safetensors",
         
     | 
| 328 | 
         
            +
                "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 329 | 
         
            +
                "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 330 | 
         
            +
                "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 331 | 
         
            +
                "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 332 | 
         
            +
                "model.layers.41.input_layernorm.weight": "model-00032-of-00061.safetensors",
         
     | 
| 333 | 
         
            +
                "model.layers.41.mlp.down_proj.weight": "model-00032-of-00061.safetensors",
         
     | 
| 334 | 
         
            +
                "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00061.safetensors",
         
     | 
| 335 | 
         
            +
                "model.layers.41.mlp.up_proj.weight": "model-00032-of-00061.safetensors",
         
     | 
| 336 | 
         
            +
                "model.layers.41.post_attention_layernorm.weight": "model-00032-of-00061.safetensors",
         
     | 
| 337 | 
         
            +
                "model.layers.41.self_attn.k_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 338 | 
         
            +
                "model.layers.41.self_attn.o_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 339 | 
         
            +
                "model.layers.41.self_attn.q_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 340 | 
         
            +
                "model.layers.41.self_attn.v_proj.weight": "model-00031-of-00061.safetensors",
         
     | 
| 341 | 
         
            +
                "model.layers.42.input_layernorm.weight": "model-00033-of-00061.safetensors",
         
     | 
| 342 | 
         
            +
                "model.layers.42.mlp.down_proj.weight": "model-00033-of-00061.safetensors",
         
     | 
| 343 | 
         
            +
                "model.layers.42.mlp.gate_proj.weight": "model-00032-of-00061.safetensors",
         
     | 
| 344 | 
         
            +
                "model.layers.42.mlp.up_proj.weight": "model-00033-of-00061.safetensors",
         
     | 
| 345 | 
         
            +
                "model.layers.42.post_attention_layernorm.weight": "model-00033-of-00061.safetensors",
         
     | 
| 346 | 
         
            +
                "model.layers.42.self_attn.k_proj.weight": "model-00032-of-00061.safetensors",
         
     | 
| 347 | 
         
            +
                "model.layers.42.self_attn.o_proj.weight": "model-00032-of-00061.safetensors",
         
     | 
| 348 | 
         
            +
                "model.layers.42.self_attn.q_proj.weight": "model-00032-of-00061.safetensors",
         
     | 
| 349 | 
         
            +
                "model.layers.42.self_attn.v_proj.weight": "model-00032-of-00061.safetensors",
         
     | 
| 350 | 
         
            +
                "model.layers.43.input_layernorm.weight": "model-00034-of-00061.safetensors",
         
     | 
| 351 | 
         
            +
                "model.layers.43.mlp.down_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 352 | 
         
            +
                "model.layers.43.mlp.gate_proj.weight": "model-00033-of-00061.safetensors",
         
     | 
| 353 | 
         
            +
                "model.layers.43.mlp.up_proj.weight": "model-00033-of-00061.safetensors",
         
     | 
| 354 | 
         
            +
                "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00061.safetensors",
         
     | 
| 355 | 
         
            +
                "model.layers.43.self_attn.k_proj.weight": "model-00033-of-00061.safetensors",
         
     | 
| 356 | 
         
            +
                "model.layers.43.self_attn.o_proj.weight": "model-00033-of-00061.safetensors",
         
     | 
| 357 | 
         
            +
                "model.layers.43.self_attn.q_proj.weight": "model-00033-of-00061.safetensors",
         
     | 
| 358 | 
         
            +
                "model.layers.43.self_attn.v_proj.weight": "model-00033-of-00061.safetensors",
         
     | 
| 359 | 
         
            +
                "model.layers.44.input_layernorm.weight": "model-00034-of-00061.safetensors",
         
     | 
| 360 | 
         
            +
                "model.layers.44.mlp.down_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 361 | 
         
            +
                "model.layers.44.mlp.gate_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 362 | 
         
            +
                "model.layers.44.mlp.up_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 363 | 
         
            +
                "model.layers.44.post_attention_layernorm.weight": "model-00034-of-00061.safetensors",
         
     | 
| 364 | 
         
            +
                "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 365 | 
         
            +
                "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 366 | 
         
            +
                "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 367 | 
         
            +
                "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 368 | 
         
            +
                "model.layers.45.input_layernorm.weight": "model-00035-of-00061.safetensors",
         
     | 
| 369 | 
         
            +
                "model.layers.45.mlp.down_proj.weight": "model-00035-of-00061.safetensors",
         
     | 
| 370 | 
         
            +
                "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00061.safetensors",
         
     | 
| 371 | 
         
            +
                "model.layers.45.mlp.up_proj.weight": "model-00035-of-00061.safetensors",
         
     | 
| 372 | 
         
            +
                "model.layers.45.post_attention_layernorm.weight": "model-00035-of-00061.safetensors",
         
     | 
| 373 | 
         
            +
                "model.layers.45.self_attn.k_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 374 | 
         
            +
                "model.layers.45.self_attn.o_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 375 | 
         
            +
                "model.layers.45.self_attn.q_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 376 | 
         
            +
                "model.layers.45.self_attn.v_proj.weight": "model-00034-of-00061.safetensors",
         
     | 
| 377 | 
         
            +
                "model.layers.46.input_layernorm.weight": "model-00036-of-00061.safetensors",
         
     | 
| 378 | 
         
            +
                "model.layers.46.mlp.down_proj.weight": "model-00036-of-00061.safetensors",
         
     | 
| 379 | 
         
            +
                "model.layers.46.mlp.gate_proj.weight": "model-00035-of-00061.safetensors",
         
     | 
| 380 | 
         
            +
                "model.layers.46.mlp.up_proj.weight": "model-00036-of-00061.safetensors",
         
     | 
| 381 | 
         
            +
                "model.layers.46.post_attention_layernorm.weight": "model-00036-of-00061.safetensors",
         
     | 
| 382 | 
         
            +
                "model.layers.46.self_attn.k_proj.weight": "model-00035-of-00061.safetensors",
         
     | 
| 383 | 
         
            +
                "model.layers.46.self_attn.o_proj.weight": "model-00035-of-00061.safetensors",
         
     | 
| 384 | 
         
            +
                "model.layers.46.self_attn.q_proj.weight": "model-00035-of-00061.safetensors",
         
     | 
| 385 | 
         
            +
                "model.layers.46.self_attn.v_proj.weight": "model-00035-of-00061.safetensors",
         
     | 
| 386 | 
         
            +
                "model.layers.47.input_layernorm.weight": "model-00037-of-00061.safetensors",
         
     | 
| 387 | 
         
            +
                "model.layers.47.mlp.down_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 388 | 
         
            +
                "model.layers.47.mlp.gate_proj.weight": "model-00036-of-00061.safetensors",
         
     | 
| 389 | 
         
            +
                "model.layers.47.mlp.up_proj.weight": "model-00036-of-00061.safetensors",
         
     | 
| 390 | 
         
            +
                "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00061.safetensors",
         
     | 
| 391 | 
         
            +
                "model.layers.47.self_attn.k_proj.weight": "model-00036-of-00061.safetensors",
         
     | 
| 392 | 
         
            +
                "model.layers.47.self_attn.o_proj.weight": "model-00036-of-00061.safetensors",
         
     | 
| 393 | 
         
            +
                "model.layers.47.self_attn.q_proj.weight": "model-00036-of-00061.safetensors",
         
     | 
| 394 | 
         
            +
                "model.layers.47.self_attn.v_proj.weight": "model-00036-of-00061.safetensors",
         
     | 
| 395 | 
         
            +
                "model.layers.48.input_layernorm.weight": "model-00037-of-00061.safetensors",
         
     | 
| 396 | 
         
            +
                "model.layers.48.mlp.down_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 397 | 
         
            +
                "model.layers.48.mlp.gate_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 398 | 
         
            +
                "model.layers.48.mlp.up_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 399 | 
         
            +
                "model.layers.48.post_attention_layernorm.weight": "model-00037-of-00061.safetensors",
         
     | 
| 400 | 
         
            +
                "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 401 | 
         
            +
                "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 402 | 
         
            +
                "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 403 | 
         
            +
                "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 404 | 
         
            +
                "model.layers.49.input_layernorm.weight": "model-00038-of-00061.safetensors",
         
     | 
| 405 | 
         
            +
                "model.layers.49.mlp.down_proj.weight": "model-00038-of-00061.safetensors",
         
     | 
| 406 | 
         
            +
                "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00061.safetensors",
         
     | 
| 407 | 
         
            +
                "model.layers.49.mlp.up_proj.weight": "model-00038-of-00061.safetensors",
         
     | 
| 408 | 
         
            +
                "model.layers.49.post_attention_layernorm.weight": "model-00038-of-00061.safetensors",
         
     | 
| 409 | 
         
            +
                "model.layers.49.self_attn.k_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 410 | 
         
            +
                "model.layers.49.self_attn.o_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 411 | 
         
            +
                "model.layers.49.self_attn.q_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 412 | 
         
            +
                "model.layers.49.self_attn.v_proj.weight": "model-00037-of-00061.safetensors",
         
     | 
| 413 | 
         
            +
                "model.layers.5.input_layernorm.weight": "model-00005-of-00061.safetensors",
         
     | 
| 414 | 
         
            +
                "model.layers.5.mlp.down_proj.weight": "model-00005-of-00061.safetensors",
         
     | 
| 415 | 
         
            +
                "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00061.safetensors",
         
     | 
| 416 | 
         
            +
                "model.layers.5.mlp.up_proj.weight": "model-00005-of-00061.safetensors",
         
     | 
| 417 | 
         
            +
                "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00061.safetensors",
         
     | 
| 418 | 
         
            +
                "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 419 | 
         
            +
                "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 420 | 
         
            +
                "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 421 | 
         
            +
                "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00061.safetensors",
         
     | 
| 422 | 
         
            +
                "model.layers.50.input_layernorm.weight": "model-00039-of-00061.safetensors",
         
     | 
| 423 | 
         
            +
                "model.layers.50.mlp.down_proj.weight": "model-00039-of-00061.safetensors",
         
     | 
| 424 | 
         
            +
                "model.layers.50.mlp.gate_proj.weight": "model-00038-of-00061.safetensors",
         
     | 
| 425 | 
         
            +
                "model.layers.50.mlp.up_proj.weight": "model-00039-of-00061.safetensors",
         
     | 
| 426 | 
         
            +
                "model.layers.50.post_attention_layernorm.weight": "model-00039-of-00061.safetensors",
         
     | 
| 427 | 
         
            +
                "model.layers.50.self_attn.k_proj.weight": "model-00038-of-00061.safetensors",
         
     | 
| 428 | 
         
            +
                "model.layers.50.self_attn.o_proj.weight": "model-00038-of-00061.safetensors",
         
     | 
| 429 | 
         
            +
                "model.layers.50.self_attn.q_proj.weight": "model-00038-of-00061.safetensors",
         
     | 
| 430 | 
         
            +
                "model.layers.50.self_attn.v_proj.weight": "model-00038-of-00061.safetensors",
         
     | 
| 431 | 
         
            +
                "model.layers.51.input_layernorm.weight": "model-00040-of-00061.safetensors",
         
     | 
| 432 | 
         
            +
                "model.layers.51.mlp.down_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 433 | 
         
            +
                "model.layers.51.mlp.gate_proj.weight": "model-00039-of-00061.safetensors",
         
     | 
| 434 | 
         
            +
                "model.layers.51.mlp.up_proj.weight": "model-00039-of-00061.safetensors",
         
     | 
| 435 | 
         
            +
                "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00061.safetensors",
         
     | 
| 436 | 
         
            +
                "model.layers.51.self_attn.k_proj.weight": "model-00039-of-00061.safetensors",
         
     | 
| 437 | 
         
            +
                "model.layers.51.self_attn.o_proj.weight": "model-00039-of-00061.safetensors",
         
     | 
| 438 | 
         
            +
                "model.layers.51.self_attn.q_proj.weight": "model-00039-of-00061.safetensors",
         
     | 
| 439 | 
         
            +
                "model.layers.51.self_attn.v_proj.weight": "model-00039-of-00061.safetensors",
         
     | 
| 440 | 
         
            +
                "model.layers.52.input_layernorm.weight": "model-00040-of-00061.safetensors",
         
     | 
| 441 | 
         
            +
                "model.layers.52.mlp.down_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 442 | 
         
            +
                "model.layers.52.mlp.gate_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 443 | 
         
            +
                "model.layers.52.mlp.up_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 444 | 
         
            +
                "model.layers.52.post_attention_layernorm.weight": "model-00040-of-00061.safetensors",
         
     | 
| 445 | 
         
            +
                "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 446 | 
         
            +
                "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 447 | 
         
            +
                "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 448 | 
         
            +
                "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 449 | 
         
            +
                "model.layers.53.input_layernorm.weight": "model-00041-of-00061.safetensors",
         
     | 
| 450 | 
         
            +
                "model.layers.53.mlp.down_proj.weight": "model-00041-of-00061.safetensors",
         
     | 
| 451 | 
         
            +
                "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00061.safetensors",
         
     | 
| 452 | 
         
            +
                "model.layers.53.mlp.up_proj.weight": "model-00041-of-00061.safetensors",
         
     | 
| 453 | 
         
            +
                "model.layers.53.post_attention_layernorm.weight": "model-00041-of-00061.safetensors",
         
     | 
| 454 | 
         
            +
                "model.layers.53.self_attn.k_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 455 | 
         
            +
                "model.layers.53.self_attn.o_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 456 | 
         
            +
                "model.layers.53.self_attn.q_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 457 | 
         
            +
                "model.layers.53.self_attn.v_proj.weight": "model-00040-of-00061.safetensors",
         
     | 
| 458 | 
         
            +
                "model.layers.54.input_layernorm.weight": "model-00042-of-00061.safetensors",
         
     | 
| 459 | 
         
            +
                "model.layers.54.mlp.down_proj.weight": "model-00042-of-00061.safetensors",
         
     | 
| 460 | 
         
            +
                "model.layers.54.mlp.gate_proj.weight": "model-00041-of-00061.safetensors",
         
     | 
| 461 | 
         
            +
                "model.layers.54.mlp.up_proj.weight": "model-00042-of-00061.safetensors",
         
     | 
| 462 | 
         
            +
                "model.layers.54.post_attention_layernorm.weight": "model-00042-of-00061.safetensors",
         
     | 
| 463 | 
         
            +
                "model.layers.54.self_attn.k_proj.weight": "model-00041-of-00061.safetensors",
         
     | 
| 464 | 
         
            +
                "model.layers.54.self_attn.o_proj.weight": "model-00041-of-00061.safetensors",
         
     | 
| 465 | 
         
            +
                "model.layers.54.self_attn.q_proj.weight": "model-00041-of-00061.safetensors",
         
     | 
| 466 | 
         
            +
                "model.layers.54.self_attn.v_proj.weight": "model-00041-of-00061.safetensors",
         
     | 
| 467 | 
         
            +
                "model.layers.55.input_layernorm.weight": "model-00043-of-00061.safetensors",
         
     | 
| 468 | 
         
            +
                "model.layers.55.mlp.down_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 469 | 
         
            +
                "model.layers.55.mlp.gate_proj.weight": "model-00042-of-00061.safetensors",
         
     | 
| 470 | 
         
            +
                "model.layers.55.mlp.up_proj.weight": "model-00042-of-00061.safetensors",
         
     | 
| 471 | 
         
            +
                "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00061.safetensors",
         
     | 
| 472 | 
         
            +
                "model.layers.55.self_attn.k_proj.weight": "model-00042-of-00061.safetensors",
         
     | 
| 473 | 
         
            +
                "model.layers.55.self_attn.o_proj.weight": "model-00042-of-00061.safetensors",
         
     | 
| 474 | 
         
            +
                "model.layers.55.self_attn.q_proj.weight": "model-00042-of-00061.safetensors",
         
     | 
| 475 | 
         
            +
                "model.layers.55.self_attn.v_proj.weight": "model-00042-of-00061.safetensors",
         
     | 
| 476 | 
         
            +
                "model.layers.56.input_layernorm.weight": "model-00043-of-00061.safetensors",
         
     | 
| 477 | 
         
            +
                "model.layers.56.mlp.down_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 478 | 
         
            +
                "model.layers.56.mlp.gate_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 479 | 
         
            +
                "model.layers.56.mlp.up_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 480 | 
         
            +
                "model.layers.56.post_attention_layernorm.weight": "model-00043-of-00061.safetensors",
         
     | 
| 481 | 
         
            +
                "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 482 | 
         
            +
                "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 483 | 
         
            +
                "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 484 | 
         
            +
                "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 485 | 
         
            +
                "model.layers.57.input_layernorm.weight": "model-00044-of-00061.safetensors",
         
     | 
| 486 | 
         
            +
                "model.layers.57.mlp.down_proj.weight": "model-00044-of-00061.safetensors",
         
     | 
| 487 | 
         
            +
                "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00061.safetensors",
         
     | 
| 488 | 
         
            +
                "model.layers.57.mlp.up_proj.weight": "model-00044-of-00061.safetensors",
         
     | 
| 489 | 
         
            +
                "model.layers.57.post_attention_layernorm.weight": "model-00044-of-00061.safetensors",
         
     | 
| 490 | 
         
            +
                "model.layers.57.self_attn.k_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 491 | 
         
            +
                "model.layers.57.self_attn.o_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 492 | 
         
            +
                "model.layers.57.self_attn.q_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 493 | 
         
            +
                "model.layers.57.self_attn.v_proj.weight": "model-00043-of-00061.safetensors",
         
     | 
| 494 | 
         
            +
                "model.layers.58.input_layernorm.weight": "model-00045-of-00061.safetensors",
         
     | 
| 495 | 
         
            +
                "model.layers.58.mlp.down_proj.weight": "model-00045-of-00061.safetensors",
         
     | 
| 496 | 
         
            +
                "model.layers.58.mlp.gate_proj.weight": "model-00044-of-00061.safetensors",
         
     | 
| 497 | 
         
            +
                "model.layers.58.mlp.up_proj.weight": "model-00045-of-00061.safetensors",
         
     | 
| 498 | 
         
            +
                "model.layers.58.post_attention_layernorm.weight": "model-00045-of-00061.safetensors",
         
     | 
| 499 | 
         
            +
                "model.layers.58.self_attn.k_proj.weight": "model-00044-of-00061.safetensors",
         
     | 
| 500 | 
         
            +
                "model.layers.58.self_attn.o_proj.weight": "model-00044-of-00061.safetensors",
         
     | 
| 501 | 
         
            +
                "model.layers.58.self_attn.q_proj.weight": "model-00044-of-00061.safetensors",
         
     | 
| 502 | 
         
            +
                "model.layers.58.self_attn.v_proj.weight": "model-00044-of-00061.safetensors",
         
     | 
| 503 | 
         
            +
                "model.layers.59.input_layernorm.weight": "model-00046-of-00061.safetensors",
         
     | 
| 504 | 
         
            +
                "model.layers.59.mlp.down_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 505 | 
         
            +
                "model.layers.59.mlp.gate_proj.weight": "model-00045-of-00061.safetensors",
         
     | 
| 506 | 
         
            +
                "model.layers.59.mlp.up_proj.weight": "model-00045-of-00061.safetensors",
         
     | 
| 507 | 
         
            +
                "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00061.safetensors",
         
     | 
| 508 | 
         
            +
                "model.layers.59.self_attn.k_proj.weight": "model-00045-of-00061.safetensors",
         
     | 
| 509 | 
         
            +
                "model.layers.59.self_attn.o_proj.weight": "model-00045-of-00061.safetensors",
         
     | 
| 510 | 
         
            +
                "model.layers.59.self_attn.q_proj.weight": "model-00045-of-00061.safetensors",
         
     | 
| 511 | 
         
            +
                "model.layers.59.self_attn.v_proj.weight": "model-00045-of-00061.safetensors",
         
     | 
| 512 | 
         
            +
                "model.layers.6.input_layernorm.weight": "model-00006-of-00061.safetensors",
         
     | 
| 513 | 
         
            +
                "model.layers.6.mlp.down_proj.weight": "model-00006-of-00061.safetensors",
         
     | 
| 514 | 
         
            +
                "model.layers.6.mlp.gate_proj.weight": "model-00005-of-00061.safetensors",
         
     | 
| 515 | 
         
            +
                "model.layers.6.mlp.up_proj.weight": "model-00006-of-00061.safetensors",
         
     | 
| 516 | 
         
            +
                "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00061.safetensors",
         
     | 
| 517 | 
         
            +
                "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00061.safetensors",
         
     | 
| 518 | 
         
            +
                "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00061.safetensors",
         
     | 
| 519 | 
         
            +
                "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00061.safetensors",
         
     | 
| 520 | 
         
            +
                "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00061.safetensors",
         
     | 
| 521 | 
         
            +
                "model.layers.60.input_layernorm.weight": "model-00046-of-00061.safetensors",
         
     | 
| 522 | 
         
            +
                "model.layers.60.mlp.down_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 523 | 
         
            +
                "model.layers.60.mlp.gate_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 524 | 
         
            +
                "model.layers.60.mlp.up_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 525 | 
         
            +
                "model.layers.60.post_attention_layernorm.weight": "model-00046-of-00061.safetensors",
         
     | 
| 526 | 
         
            +
                "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 527 | 
         
            +
                "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 528 | 
         
            +
                "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 529 | 
         
            +
                "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 530 | 
         
            +
                "model.layers.61.input_layernorm.weight": "model-00047-of-00061.safetensors",
         
     | 
| 531 | 
         
            +
                "model.layers.61.mlp.down_proj.weight": "model-00047-of-00061.safetensors",
         
     | 
| 532 | 
         
            +
                "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00061.safetensors",
         
     | 
| 533 | 
         
            +
                "model.layers.61.mlp.up_proj.weight": "model-00047-of-00061.safetensors",
         
     | 
| 534 | 
         
            +
                "model.layers.61.post_attention_layernorm.weight": "model-00047-of-00061.safetensors",
         
     | 
| 535 | 
         
            +
                "model.layers.61.self_attn.k_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 536 | 
         
            +
                "model.layers.61.self_attn.o_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 537 | 
         
            +
                "model.layers.61.self_attn.q_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 538 | 
         
            +
                "model.layers.61.self_attn.v_proj.weight": "model-00046-of-00061.safetensors",
         
     | 
| 539 | 
         
            +
                "model.layers.62.input_layernorm.weight": "model-00048-of-00061.safetensors",
         
     | 
| 540 | 
         
            +
                "model.layers.62.mlp.down_proj.weight": "model-00048-of-00061.safetensors",
         
     | 
| 541 | 
         
            +
                "model.layers.62.mlp.gate_proj.weight": "model-00047-of-00061.safetensors",
         
     | 
| 542 | 
         
            +
                "model.layers.62.mlp.up_proj.weight": "model-00048-of-00061.safetensors",
         
     | 
| 543 | 
         
            +
                "model.layers.62.post_attention_layernorm.weight": "model-00048-of-00061.safetensors",
         
     | 
| 544 | 
         
            +
                "model.layers.62.self_attn.k_proj.weight": "model-00047-of-00061.safetensors",
         
     | 
| 545 | 
         
            +
                "model.layers.62.self_attn.o_proj.weight": "model-00047-of-00061.safetensors",
         
     | 
| 546 | 
         
            +
                "model.layers.62.self_attn.q_proj.weight": "model-00047-of-00061.safetensors",
         
     | 
| 547 | 
         
            +
                "model.layers.62.self_attn.v_proj.weight": "model-00047-of-00061.safetensors",
         
     | 
| 548 | 
         
            +
                "model.layers.63.input_layernorm.weight": "model-00049-of-00061.safetensors",
         
     | 
| 549 | 
         
            +
                "model.layers.63.mlp.down_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 550 | 
         
            +
                "model.layers.63.mlp.gate_proj.weight": "model-00048-of-00061.safetensors",
         
     | 
| 551 | 
         
            +
                "model.layers.63.mlp.up_proj.weight": "model-00048-of-00061.safetensors",
         
     | 
| 552 | 
         
            +
                "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00061.safetensors",
         
     | 
| 553 | 
         
            +
                "model.layers.63.self_attn.k_proj.weight": "model-00048-of-00061.safetensors",
         
     | 
| 554 | 
         
            +
                "model.layers.63.self_attn.o_proj.weight": "model-00048-of-00061.safetensors",
         
     | 
| 555 | 
         
            +
                "model.layers.63.self_attn.q_proj.weight": "model-00048-of-00061.safetensors",
         
     | 
| 556 | 
         
            +
                "model.layers.63.self_attn.v_proj.weight": "model-00048-of-00061.safetensors",
         
     | 
| 557 | 
         
            +
                "model.layers.64.input_layernorm.weight": "model-00049-of-00061.safetensors",
         
     | 
| 558 | 
         
            +
                "model.layers.64.mlp.down_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 559 | 
         
            +
                "model.layers.64.mlp.gate_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 560 | 
         
            +
                "model.layers.64.mlp.up_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 561 | 
         
            +
                "model.layers.64.post_attention_layernorm.weight": "model-00049-of-00061.safetensors",
         
     | 
| 562 | 
         
            +
                "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 563 | 
         
            +
                "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 564 | 
         
            +
                "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 565 | 
         
            +
                "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 566 | 
         
            +
                "model.layers.65.input_layernorm.weight": "model-00050-of-00061.safetensors",
         
     | 
| 567 | 
         
            +
                "model.layers.65.mlp.down_proj.weight": "model-00050-of-00061.safetensors",
         
     | 
| 568 | 
         
            +
                "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00061.safetensors",
         
     | 
| 569 | 
         
            +
                "model.layers.65.mlp.up_proj.weight": "model-00050-of-00061.safetensors",
         
     | 
| 570 | 
         
            +
                "model.layers.65.post_attention_layernorm.weight": "model-00050-of-00061.safetensors",
         
     | 
| 571 | 
         
            +
                "model.layers.65.self_attn.k_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 572 | 
         
            +
                "model.layers.65.self_attn.o_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 573 | 
         
            +
                "model.layers.65.self_attn.q_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 574 | 
         
            +
                "model.layers.65.self_attn.v_proj.weight": "model-00049-of-00061.safetensors",
         
     | 
| 575 | 
         
            +
                "model.layers.66.input_layernorm.weight": "model-00051-of-00061.safetensors",
         
     | 
| 576 | 
         
            +
                "model.layers.66.mlp.down_proj.weight": "model-00051-of-00061.safetensors",
         
     | 
| 577 | 
         
            +
                "model.layers.66.mlp.gate_proj.weight": "model-00050-of-00061.safetensors",
         
     | 
| 578 | 
         
            +
                "model.layers.66.mlp.up_proj.weight": "model-00051-of-00061.safetensors",
         
     | 
| 579 | 
         
            +
                "model.layers.66.post_attention_layernorm.weight": "model-00051-of-00061.safetensors",
         
     | 
| 580 | 
         
            +
                "model.layers.66.self_attn.k_proj.weight": "model-00050-of-00061.safetensors",
         
     | 
| 581 | 
         
            +
                "model.layers.66.self_attn.o_proj.weight": "model-00050-of-00061.safetensors",
         
     | 
| 582 | 
         
            +
                "model.layers.66.self_attn.q_proj.weight": "model-00050-of-00061.safetensors",
         
     | 
| 583 | 
         
            +
                "model.layers.66.self_attn.v_proj.weight": "model-00050-of-00061.safetensors",
         
     | 
| 584 | 
         
            +
                "model.layers.67.input_layernorm.weight": "model-00052-of-00061.safetensors",
         
     | 
| 585 | 
         
            +
                "model.layers.67.mlp.down_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 586 | 
         
            +
                "model.layers.67.mlp.gate_proj.weight": "model-00051-of-00061.safetensors",
         
     | 
| 587 | 
         
            +
                "model.layers.67.mlp.up_proj.weight": "model-00051-of-00061.safetensors",
         
     | 
| 588 | 
         
            +
                "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00061.safetensors",
         
     | 
| 589 | 
         
            +
                "model.layers.67.self_attn.k_proj.weight": "model-00051-of-00061.safetensors",
         
     | 
| 590 | 
         
            +
                "model.layers.67.self_attn.o_proj.weight": "model-00051-of-00061.safetensors",
         
     | 
| 591 | 
         
            +
                "model.layers.67.self_attn.q_proj.weight": "model-00051-of-00061.safetensors",
         
     | 
| 592 | 
         
            +
                "model.layers.67.self_attn.v_proj.weight": "model-00051-of-00061.safetensors",
         
     | 
| 593 | 
         
            +
                "model.layers.68.input_layernorm.weight": "model-00052-of-00061.safetensors",
         
     | 
| 594 | 
         
            +
                "model.layers.68.mlp.down_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 595 | 
         
            +
                "model.layers.68.mlp.gate_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 596 | 
         
            +
                "model.layers.68.mlp.up_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 597 | 
         
            +
                "model.layers.68.post_attention_layernorm.weight": "model-00052-of-00061.safetensors",
         
     | 
| 598 | 
         
            +
                "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 599 | 
         
            +
                "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 600 | 
         
            +
                "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 601 | 
         
            +
                "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 602 | 
         
            +
                "model.layers.69.input_layernorm.weight": "model-00053-of-00061.safetensors",
         
     | 
| 603 | 
         
            +
                "model.layers.69.mlp.down_proj.weight": "model-00053-of-00061.safetensors",
         
     | 
| 604 | 
         
            +
                "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00061.safetensors",
         
     | 
| 605 | 
         
            +
                "model.layers.69.mlp.up_proj.weight": "model-00053-of-00061.safetensors",
         
     | 
| 606 | 
         
            +
                "model.layers.69.post_attention_layernorm.weight": "model-00053-of-00061.safetensors",
         
     | 
| 607 | 
         
            +
                "model.layers.69.self_attn.k_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 608 | 
         
            +
                "model.layers.69.self_attn.o_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 609 | 
         
            +
                "model.layers.69.self_attn.q_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 610 | 
         
            +
                "model.layers.69.self_attn.v_proj.weight": "model-00052-of-00061.safetensors",
         
     | 
| 611 | 
         
            +
                "model.layers.7.input_layernorm.weight": "model-00007-of-00061.safetensors",
         
     | 
| 612 | 
         
            +
                "model.layers.7.mlp.down_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 613 | 
         
            +
                "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00061.safetensors",
         
     | 
| 614 | 
         
            +
                "model.layers.7.mlp.up_proj.weight": "model-00006-of-00061.safetensors",
         
     | 
| 615 | 
         
            +
                "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00061.safetensors",
         
     | 
| 616 | 
         
            +
                "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00061.safetensors",
         
     | 
| 617 | 
         
            +
                "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00061.safetensors",
         
     | 
| 618 | 
         
            +
                "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00061.safetensors",
         
     | 
| 619 | 
         
            +
                "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00061.safetensors",
         
     | 
| 620 | 
         
            +
                "model.layers.70.input_layernorm.weight": "model-00054-of-00061.safetensors",
         
     | 
| 621 | 
         
            +
                "model.layers.70.mlp.down_proj.weight": "model-00054-of-00061.safetensors",
         
     | 
| 622 | 
         
            +
                "model.layers.70.mlp.gate_proj.weight": "model-00053-of-00061.safetensors",
         
     | 
| 623 | 
         
            +
                "model.layers.70.mlp.up_proj.weight": "model-00054-of-00061.safetensors",
         
     | 
| 624 | 
         
            +
                "model.layers.70.post_attention_layernorm.weight": "model-00054-of-00061.safetensors",
         
     | 
| 625 | 
         
            +
                "model.layers.70.self_attn.k_proj.weight": "model-00053-of-00061.safetensors",
         
     | 
| 626 | 
         
            +
                "model.layers.70.self_attn.o_proj.weight": "model-00053-of-00061.safetensors",
         
     | 
| 627 | 
         
            +
                "model.layers.70.self_attn.q_proj.weight": "model-00053-of-00061.safetensors",
         
     | 
| 628 | 
         
            +
                "model.layers.70.self_attn.v_proj.weight": "model-00053-of-00061.safetensors",
         
     | 
| 629 | 
         
            +
                "model.layers.71.input_layernorm.weight": "model-00055-of-00061.safetensors",
         
     | 
| 630 | 
         
            +
                "model.layers.71.mlp.down_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 631 | 
         
            +
                "model.layers.71.mlp.gate_proj.weight": "model-00054-of-00061.safetensors",
         
     | 
| 632 | 
         
            +
                "model.layers.71.mlp.up_proj.weight": "model-00054-of-00061.safetensors",
         
     | 
| 633 | 
         
            +
                "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00061.safetensors",
         
     | 
| 634 | 
         
            +
                "model.layers.71.self_attn.k_proj.weight": "model-00054-of-00061.safetensors",
         
     | 
| 635 | 
         
            +
                "model.layers.71.self_attn.o_proj.weight": "model-00054-of-00061.safetensors",
         
     | 
| 636 | 
         
            +
                "model.layers.71.self_attn.q_proj.weight": "model-00054-of-00061.safetensors",
         
     | 
| 637 | 
         
            +
                "model.layers.71.self_attn.v_proj.weight": "model-00054-of-00061.safetensors",
         
     | 
| 638 | 
         
            +
                "model.layers.72.input_layernorm.weight": "model-00055-of-00061.safetensors",
         
     | 
| 639 | 
         
            +
                "model.layers.72.mlp.down_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 640 | 
         
            +
                "model.layers.72.mlp.gate_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 641 | 
         
            +
                "model.layers.72.mlp.up_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 642 | 
         
            +
                "model.layers.72.post_attention_layernorm.weight": "model-00055-of-00061.safetensors",
         
     | 
| 643 | 
         
            +
                "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 644 | 
         
            +
                "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 645 | 
         
            +
                "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 646 | 
         
            +
                "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 647 | 
         
            +
                "model.layers.73.input_layernorm.weight": "model-00056-of-00061.safetensors",
         
     | 
| 648 | 
         
            +
                "model.layers.73.mlp.down_proj.weight": "model-00056-of-00061.safetensors",
         
     | 
| 649 | 
         
            +
                "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00061.safetensors",
         
     | 
| 650 | 
         
            +
                "model.layers.73.mlp.up_proj.weight": "model-00056-of-00061.safetensors",
         
     | 
| 651 | 
         
            +
                "model.layers.73.post_attention_layernorm.weight": "model-00056-of-00061.safetensors",
         
     | 
| 652 | 
         
            +
                "model.layers.73.self_attn.k_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 653 | 
         
            +
                "model.layers.73.self_attn.o_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 654 | 
         
            +
                "model.layers.73.self_attn.q_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 655 | 
         
            +
                "model.layers.73.self_attn.v_proj.weight": "model-00055-of-00061.safetensors",
         
     | 
| 656 | 
         
            +
                "model.layers.74.input_layernorm.weight": "model-00057-of-00061.safetensors",
         
     | 
| 657 | 
         
            +
                "model.layers.74.mlp.down_proj.weight": "model-00057-of-00061.safetensors",
         
     | 
| 658 | 
         
            +
                "model.layers.74.mlp.gate_proj.weight": "model-00056-of-00061.safetensors",
         
     | 
| 659 | 
         
            +
                "model.layers.74.mlp.up_proj.weight": "model-00057-of-00061.safetensors",
         
     | 
| 660 | 
         
            +
                "model.layers.74.post_attention_layernorm.weight": "model-00057-of-00061.safetensors",
         
     | 
| 661 | 
         
            +
                "model.layers.74.self_attn.k_proj.weight": "model-00056-of-00061.safetensors",
         
     | 
| 662 | 
         
            +
                "model.layers.74.self_attn.o_proj.weight": "model-00056-of-00061.safetensors",
         
     | 
| 663 | 
         
            +
                "model.layers.74.self_attn.q_proj.weight": "model-00056-of-00061.safetensors",
         
     | 
| 664 | 
         
            +
                "model.layers.74.self_attn.v_proj.weight": "model-00056-of-00061.safetensors",
         
     | 
| 665 | 
         
            +
                "model.layers.75.input_layernorm.weight": "model-00058-of-00061.safetensors",
         
     | 
| 666 | 
         
            +
                "model.layers.75.mlp.down_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 667 | 
         
            +
                "model.layers.75.mlp.gate_proj.weight": "model-00057-of-00061.safetensors",
         
     | 
| 668 | 
         
            +
                "model.layers.75.mlp.up_proj.weight": "model-00057-of-00061.safetensors",
         
     | 
| 669 | 
         
            +
                "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00061.safetensors",
         
     | 
| 670 | 
         
            +
                "model.layers.75.self_attn.k_proj.weight": "model-00057-of-00061.safetensors",
         
     | 
| 671 | 
         
            +
                "model.layers.75.self_attn.o_proj.weight": "model-00057-of-00061.safetensors",
         
     | 
| 672 | 
         
            +
                "model.layers.75.self_attn.q_proj.weight": "model-00057-of-00061.safetensors",
         
     | 
| 673 | 
         
            +
                "model.layers.75.self_attn.v_proj.weight": "model-00057-of-00061.safetensors",
         
     | 
| 674 | 
         
            +
                "model.layers.76.input_layernorm.weight": "model-00058-of-00061.safetensors",
         
     | 
| 675 | 
         
            +
                "model.layers.76.mlp.down_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 676 | 
         
            +
                "model.layers.76.mlp.gate_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 677 | 
         
            +
                "model.layers.76.mlp.up_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 678 | 
         
            +
                "model.layers.76.post_attention_layernorm.weight": "model-00058-of-00061.safetensors",
         
     | 
| 679 | 
         
            +
                "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 680 | 
         
            +
                "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 681 | 
         
            +
                "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 682 | 
         
            +
                "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 683 | 
         
            +
                "model.layers.77.input_layernorm.weight": "model-00059-of-00061.safetensors",
         
     | 
| 684 | 
         
            +
                "model.layers.77.mlp.down_proj.weight": "model-00059-of-00061.safetensors",
         
     | 
| 685 | 
         
            +
                "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00061.safetensors",
         
     | 
| 686 | 
         
            +
                "model.layers.77.mlp.up_proj.weight": "model-00059-of-00061.safetensors",
         
     | 
| 687 | 
         
            +
                "model.layers.77.post_attention_layernorm.weight": "model-00059-of-00061.safetensors",
         
     | 
| 688 | 
         
            +
                "model.layers.77.self_attn.k_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 689 | 
         
            +
                "model.layers.77.self_attn.o_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 690 | 
         
            +
                "model.layers.77.self_attn.q_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 691 | 
         
            +
                "model.layers.77.self_attn.v_proj.weight": "model-00058-of-00061.safetensors",
         
     | 
| 692 | 
         
            +
                "model.layers.78.input_layernorm.weight": "model-00060-of-00061.safetensors",
         
     | 
| 693 | 
         
            +
                "model.layers.78.mlp.down_proj.weight": "model-00060-of-00061.safetensors",
         
     | 
| 694 | 
         
            +
                "model.layers.78.mlp.gate_proj.weight": "model-00059-of-00061.safetensors",
         
     | 
| 695 | 
         
            +
                "model.layers.78.mlp.up_proj.weight": "model-00060-of-00061.safetensors",
         
     | 
| 696 | 
         
            +
                "model.layers.78.post_attention_layernorm.weight": "model-00060-of-00061.safetensors",
         
     | 
| 697 | 
         
            +
                "model.layers.78.self_attn.k_proj.weight": "model-00059-of-00061.safetensors",
         
     | 
| 698 | 
         
            +
                "model.layers.78.self_attn.o_proj.weight": "model-00059-of-00061.safetensors",
         
     | 
| 699 | 
         
            +
                "model.layers.78.self_attn.q_proj.weight": "model-00059-of-00061.safetensors",
         
     | 
| 700 | 
         
            +
                "model.layers.78.self_attn.v_proj.weight": "model-00059-of-00061.safetensors",
         
     | 
| 701 | 
         
            +
                "model.layers.79.input_layernorm.weight": "model-00061-of-00061.safetensors",
         
     | 
| 702 | 
         
            +
                "model.layers.79.mlp.down_proj.weight": "model-00061-of-00061.safetensors",
         
     | 
| 703 | 
         
            +
                "model.layers.79.mlp.gate_proj.weight": "model-00060-of-00061.safetensors",
         
     | 
| 704 | 
         
            +
                "model.layers.79.mlp.up_proj.weight": "model-00060-of-00061.safetensors",
         
     | 
| 705 | 
         
            +
                "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00061.safetensors",
         
     | 
| 706 | 
         
            +
                "model.layers.79.self_attn.k_proj.weight": "model-00060-of-00061.safetensors",
         
     | 
| 707 | 
         
            +
                "model.layers.79.self_attn.o_proj.weight": "model-00060-of-00061.safetensors",
         
     | 
| 708 | 
         
            +
                "model.layers.79.self_attn.q_proj.weight": "model-00060-of-00061.safetensors",
         
     | 
| 709 | 
         
            +
                "model.layers.79.self_attn.v_proj.weight": "model-00060-of-00061.safetensors",
         
     | 
| 710 | 
         
            +
                "model.layers.8.input_layernorm.weight": "model-00007-of-00061.safetensors",
         
     | 
| 711 | 
         
            +
                "model.layers.8.mlp.down_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 712 | 
         
            +
                "model.layers.8.mlp.gate_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 713 | 
         
            +
                "model.layers.8.mlp.up_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 714 | 
         
            +
                "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00061.safetensors",
         
     | 
| 715 | 
         
            +
                "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 716 | 
         
            +
                "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 717 | 
         
            +
                "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 718 | 
         
            +
                "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 719 | 
         
            +
                "model.layers.9.input_layernorm.weight": "model-00008-of-00061.safetensors",
         
     | 
| 720 | 
         
            +
                "model.layers.9.mlp.down_proj.weight": "model-00008-of-00061.safetensors",
         
     | 
| 721 | 
         
            +
                "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00061.safetensors",
         
     | 
| 722 | 
         
            +
                "model.layers.9.mlp.up_proj.weight": "model-00008-of-00061.safetensors",
         
     | 
| 723 | 
         
            +
                "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00061.safetensors",
         
     | 
| 724 | 
         
            +
                "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 725 | 
         
            +
                "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 726 | 
         
            +
                "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 727 | 
         
            +
                "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00061.safetensors",
         
     | 
| 728 | 
         
            +
                "model.norm.weight": "model-00061-of-00061.safetensors"
         
     | 
| 729 | 
         
            +
              }
         
     | 
| 730 | 
         
            +
            }
         
     | 
| 
         @@ -0,0 +1,24 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "bos_token": {
         
     | 
| 3 | 
         
            +
                "content": "<s>",
         
     | 
| 4 | 
         
            +
                "lstrip": false,
         
     | 
| 5 | 
         
            +
                "normalized": false,
         
     | 
| 6 | 
         
            +
                "rstrip": false,
         
     | 
| 7 | 
         
            +
                "single_word": false
         
     | 
| 8 | 
         
            +
              },
         
     | 
| 9 | 
         
            +
              "eos_token": {
         
     | 
| 10 | 
         
            +
                "content": "</s>",
         
     | 
| 11 | 
         
            +
                "lstrip": false,
         
     | 
| 12 | 
         
            +
                "normalized": false,
         
     | 
| 13 | 
         
            +
                "rstrip": false,
         
     | 
| 14 | 
         
            +
                "single_word": false
         
     | 
| 15 | 
         
            +
              },
         
     | 
| 16 | 
         
            +
              "pad_token": "</s>",
         
     | 
| 17 | 
         
            +
              "unk_token": {
         
     | 
| 18 | 
         
            +
                "content": "<unk>",
         
     | 
| 19 | 
         
            +
                "lstrip": false,
         
     | 
| 20 | 
         
            +
                "normalized": false,
         
     | 
| 21 | 
         
            +
                "rstrip": false,
         
     | 
| 22 | 
         
            +
                "single_word": false
         
     | 
| 23 | 
         
            +
              }
         
     | 
| 24 | 
         
            +
            }
         
     | 
| 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
| 
         @@ -0,0 +1,40 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "add_bos_token": true,
         
     | 
| 3 | 
         
            +
              "add_eos_token": false,
         
     | 
| 4 | 
         
            +
              "added_tokens_decoder": {
         
     | 
| 5 | 
         
            +
                "0": {
         
     | 
| 6 | 
         
            +
                  "content": "<unk>",
         
     | 
| 7 | 
         
            +
                  "lstrip": false,
         
     | 
| 8 | 
         
            +
                  "normalized": false,
         
     | 
| 9 | 
         
            +
                  "rstrip": false,
         
     | 
| 10 | 
         
            +
                  "single_word": false,
         
     | 
| 11 | 
         
            +
                  "special": true
         
     | 
| 12 | 
         
            +
                },
         
     | 
| 13 | 
         
            +
                "1": {
         
     | 
| 14 | 
         
            +
                  "content": "<s>",
         
     | 
| 15 | 
         
            +
                  "lstrip": false,
         
     | 
| 16 | 
         
            +
                  "normalized": false,
         
     | 
| 17 | 
         
            +
                  "rstrip": false,
         
     | 
| 18 | 
         
            +
                  "single_word": false,
         
     | 
| 19 | 
         
            +
                  "special": true
         
     | 
| 20 | 
         
            +
                },
         
     | 
| 21 | 
         
            +
                "2": {
         
     | 
| 22 | 
         
            +
                  "content": "</s>",
         
     | 
| 23 | 
         
            +
                  "lstrip": false,
         
     | 
| 24 | 
         
            +
                  "normalized": false,
         
     | 
| 25 | 
         
            +
                  "rstrip": false,
         
     | 
| 26 | 
         
            +
                  "single_word": false,
         
     | 
| 27 | 
         
            +
                  "special": true
         
     | 
| 28 | 
         
            +
                }
         
     | 
| 29 | 
         
            +
              },
         
     | 
| 30 | 
         
            +
              "bos_token": "<s>",
         
     | 
| 31 | 
         
            +
              "clean_up_tokenization_spaces": false,
         
     | 
| 32 | 
         
            +
              "eos_token": "</s>",
         
     | 
| 33 | 
         
            +
              "legacy": false,
         
     | 
| 34 | 
         
            +
              "model_max_length": 1000000000000000019884624838656,
         
     | 
| 35 | 
         
            +
              "pad_token": "</s>",
         
     | 
| 36 | 
         
            +
              "sp_model_kwargs": {},
         
     | 
| 37 | 
         
            +
              "tokenizer_class": "LlamaTokenizer",
         
     | 
| 38 | 
         
            +
              "unk_token": "<unk>",
         
     | 
| 39 | 
         
            +
              "use_default_system_prompt": true
         
     | 
| 40 | 
         
            +
            }
         
     | 
| 
         @@ -0,0 +1,29 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "_name_or_path": "sequelbox/SpellBlade",
         
     | 
| 3 | 
         
            +
              "architectures": [
         
     | 
| 4 | 
         
            +
                "LlamaForCausalLM"
         
     | 
| 5 | 
         
            +
              ],
         
     | 
| 6 | 
         
            +
              "attention_bias": false,
         
     | 
| 7 | 
         
            +
              "attention_dropout": 0.0,
         
     | 
| 8 | 
         
            +
              "bos_token_id": 1,
         
     | 
| 9 | 
         
            +
              "eos_token_id": 2,
         
     | 
| 10 | 
         
            +
              "hidden_act": "silu",
         
     | 
| 11 | 
         
            +
              "hidden_size": 8192,
         
     | 
| 12 | 
         
            +
              "initializer_range": 0.02,
         
     | 
| 13 | 
         
            +
              "intermediate_size": 28672,
         
     | 
| 14 | 
         
            +
              "max_position_embeddings": 4096,
         
     | 
| 15 | 
         
            +
              "model_type": "llama",
         
     | 
| 16 | 
         
            +
              "num_attention_heads": 64,
         
     | 
| 17 | 
         
            +
              "num_hidden_layers": 80,
         
     | 
| 18 | 
         
            +
              "num_key_value_heads": 8,
         
     | 
| 19 | 
         
            +
              "pad_token_id": 0,
         
     | 
| 20 | 
         
            +
              "pretraining_tp": 1,
         
     | 
| 21 | 
         
            +
              "rms_norm_eps": 1e-05,
         
     | 
| 22 | 
         
            +
              "rope_scaling": null,
         
     | 
| 23 | 
         
            +
              "rope_theta": 10000.0,
         
     | 
| 24 | 
         
            +
              "tie_word_embeddings": false,
         
     | 
| 25 | 
         
            +
              "torch_dtype": "float32",
         
     | 
| 26 | 
         
            +
              "transformers_version": "4.36.2",
         
     | 
| 27 | 
         
            +
              "use_cache": false,
         
     | 
| 28 | 
         
            +
              "vocab_size": 32000
         
     | 
| 29 | 
         
            +
            }
         
     | 
| 
         @@ -0,0 +1,8 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "_from_model_config": true,
         
     | 
| 3 | 
         
            +
              "bos_token_id": 1,
         
     | 
| 4 | 
         
            +
              "eos_token_id": 2,
         
     | 
| 5 | 
         
            +
              "pad_token_id": 0,
         
     | 
| 6 | 
         
            +
              "transformers_version": "4.36.2",
         
     | 
| 7 | 
         
            +
              "use_cache": false
         
     | 
| 8 | 
         
            +
            }
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:5352d1d26fb1a4187e6655674d3d7b9e65578dd4f88e05e3053008c53acfe878
         
     | 
| 3 | 
         
            +
            size 4806739440
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:4e28499b8d3d1d6db61e3385c3736a1d8faaf891a2fc2ce89b71739f3b23a551
         
     | 
| 3 | 
         
            +
            size 4630578440
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:a69dced67f39f9f6c60816642904fba0e123102682df36a7703c4da3ed2b3478
         
     | 
| 3 | 
         
            +
            size 4362142864
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:c06846f87b3828821532d315409db90421ec463a8fb3057a2e28d5ebff4b368b
         
     | 
| 3 | 
         
            +
            size 4966188864
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:da16ef7dba13ec0c23aa5721c1c39a1793f0aac98a122942afb50bfd33ab0d12
         
     | 
| 3 | 
         
            +
            size 4362142864
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:b7af94011077c5ff5cb6e41b28360c9ddecb491db47380274a61e687cea70001
         
     | 
| 3 | 
         
            +
            size 4362142864
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:54c6b0177452845eef4a53ff47787150687efe0e8ae05cf3bcdb45ae9b85ce95
         
     | 
| 3 | 
         
            +
            size 4966188864
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:b60807d26c02d8c09f84aae1e630e4c298eec45b90bc8a9bb4ff21b9886a65f0
         
     | 
| 3 | 
         
            +
            size 4362142880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:20fafa021662573a48905ec08221dd85ed0181a0088055832fa95997f92cbed6
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:d86186d95a9895d1c2942dedaa390c45b8749fa7dc816461ee45fc08c764b4ec
         
     | 
| 3 | 
         
            +
            size 4966188880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:70fa9d2b4d1922ed02f61b527528d56444692ec3fc136998eb06eeb344194016
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:e696bdf1acd81583bd846d5355bd401876398bc3b1ba8977846a1125dcaf1cc6
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:0eed999502c6eeff2b1a5158b16f7db3d3121c957ec9143eea3552eef8562082
         
     | 
| 3 | 
         
            +
            size 4966188880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:5ed50e80d456ce42bab05404bc5a13f36cd668a8acfd055cbe945647b79bdf69
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:24534c62918587285be9533a4cde711864570ecc425350fbb0278a591f2c6393
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:417705ff101a6516ed3d3a018fe5968109e6505eaf88966233fc3485956a1efa
         
     | 
| 3 | 
         
            +
            size 4966188880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:50ac743e5437c9321360795d533c3767a4454ba373d0ab781970c5b67d871568
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:35bb3eb4e4d049834b41635404827f9df27ab3c996aa9734a25da421c200c2a0
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:a1cedc43136ff86f562bc67c9088d54e9b8e495c0052f31e975dda03d715a0c1
         
     | 
| 3 | 
         
            +
            size 4966188880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:692706f2b9e2c65accd5e8c2cafbb12447f9b97968c30ceb7645de7f70a5111f
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:70f933252a5599e0388aeeaf126f2a3c14f9be83724ed8a08bad491511ec7269
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:454788a5f8ed3ca95d84b6fae013ab9832d517c1714239240e80016d7a1c4d15
         
     | 
| 3 | 
         
            +
            size 4966188880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:4aefb059c46a9998dfc942961bfeedc719e854ed1189c0b4102709acf297e539
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:15de4c62eeee6d8c6c58039c17c9dc1d298b73e93838880c5d8242842b75b05f
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:6f8a8d5a10febcf110b98976ffcec9ca8cd17d1fda8d7046b6726b04728b412c
         
     | 
| 3 | 
         
            +
            size 4966188880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:4afd46ecebb697ade9f2a6d9323a635e43bbb1f3364c91fe3434162eaffdb8af
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:8c7c91daeae73cfdbcb3167604a4e002fb3803e4ed0e21a5c9cb1fde61688835
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:602add813be1febc8a4330796075ad18927aea3c5215dea0f976fd8a87cac61f
         
     | 
| 3 | 
         
            +
            size 4966188880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:e9d2db54aecdf9bdf854cb640be346b4023ef2857abeb720297c59627a7ea8e2
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:c49e0ba95fa93516bb09eb6d6d8cdc1c9c36d9f11ea24d64834e5a81817a41a9
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:74afde4aa694a3f0e7568bfc5c2cc72d14e632fb0b8c1c0b52cc34131588d03e
         
     | 
| 3 | 
         
            +
            size 4966188880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:c553a79c4dc7527158d1ee8c63e6cfdfed7cb544247e14f8a262a869db928ac1
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:e64f4c3e6c63b879ab412b1b10c81411e2a885deea883c4d1cc1392e6ad06ee6
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:3a7c8d6bf49315f4a15792aad78c45ad306f997fc870fa639e82c2411c361362
         
     | 
| 3 | 
         
            +
            size 4966188880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:d57fb6faaecc3f1ce6646bd3c1576712e7e00e82bebc099475710f1a30ed133e
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:2543b012997bc2f260198166c51c328e98edd2e4a764317c43147aa6ee1834d6
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:a2caf2871528d1a8a5871d90208b53312e3b821f09ca8c9aa6ac7ef8e55fc6a4
         
     | 
| 3 | 
         
            +
            size 4966188880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:818c322947e1675269b0e4513cc8cb8168849f46e9ce8c4bec12b57c6a5a3fbe
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:c0f4aae1fc83d6c5a7235e7255b5f86c28b42b1077d1237878b59bcc0ac53b86
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:f5ad4528388d48d8854250724ad0025e05091bb4af040186d0cd11aa508e7a90
         
     | 
| 3 | 
         
            +
            size 4966188880
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:c8bf95940073d1e160e5349dbacafff6aa9b93d56ee7dc17500462c38b913dec
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     | 
| 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:d1f30feb65fd41e48016368c5c12a077ae2c2c01cb2690bf465288c9141eefaa
         
     | 
| 3 | 
         
            +
            size 4362142872
         
     |