diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..cacd7c317b49e2ff5c8009b286ccab8ed385251f --- /dev/null +++ b/README.md @@ -0,0 +1,201 @@ +--- +library_name: transformers +tags: [] +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + +This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + + diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..6465b9290092ca87a98bafa5144002e16d1f00f9 --- /dev/null +++ b/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "meta-llama/Llama-2-70b-hf", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 1, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 4096, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 10000.0, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.37.2", + "use_cache": true, + "vocab_size": 32000 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..55265837bb0725852b4b654901635511f73bc977 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,10 @@ +{ + "bos_token_id": 1, + "do_sample": true, + "eos_token_id": 2, + "max_length": 4096, + "pad_token_id": 0, + "temperature": 0.6, + "top_p": 0.9, + "transformers_version": "4.37.2" +} diff --git a/model-00001-of-00061.safetensors b/model-00001-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fc4685283e4c5acbcee8f22d3c84ef54bcf032d8 --- /dev/null +++ b/model-00001-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d24658ed13c9ff7707a8a8b6640d985729e4e3e9a8e4ebbd296e392a913c9759 +size 4806739440 diff --git a/model-00002-of-00061.safetensors b/model-00002-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..91dcef50aae78dbbd549998e8f7057115a3271cc --- /dev/null +++ b/model-00002-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e0692ef1d8cc33688d0201918ad6c01680c4fd2cc881dc1baa156ff2c1b65df +size 4630578440 diff --git a/model-00003-of-00061.safetensors b/model-00003-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..098490345bbb437c70b1f741d811dc680a8a0b27 --- /dev/null +++ b/model-00003-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f8c4b4386918b267a88b2f519df037f53c7c8aef2cdb9ae2256c0efd2f6e174 +size 4362142864 diff --git a/model-00004-of-00061.safetensors b/model-00004-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..433fd88d7bfdcd1f4e6754196ce50c1cb37114b0 --- /dev/null +++ b/model-00004-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22b2fdfa6318a04ad9b7cfb5a91c2c917af223bca26385f716b8dde898c369d1 +size 4966188864 diff --git a/model-00005-of-00061.safetensors b/model-00005-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5caaba5d10e95ed8a5b078dd6875c4af0231b789 --- /dev/null +++ b/model-00005-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0cd7c614dd38fd3e8a9693ab33756ce702fac75e370beba56720cab473dcff96 +size 4362142864 diff --git a/model-00006-of-00061.safetensors b/model-00006-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..876e627e348bb9ffbd2241d62fdb88e72be985e8 --- /dev/null +++ b/model-00006-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c43ecb4f299f80138e074eac7cdfa2ee5b2f713a5e4c36b37d98a140c92e84b +size 4362142864 diff --git a/model-00007-of-00061.safetensors b/model-00007-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f4ff0567f216ddd2b0427f7520c5a0156d8d3450 --- /dev/null +++ b/model-00007-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9ea01c4a23a9f41ba72886eb3b69e47e25f7b0a190a8df03882c2d2803748f8 +size 4966188864 diff --git a/model-00008-of-00061.safetensors b/model-00008-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fb3c660f6bee06cac99f8f371c4b4554b3b89940 --- /dev/null +++ b/model-00008-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b35ae0b345a244c962f0bf47ff3117c3ea2d03a6fa42fc776e7d71116ef7d70a +size 4362142880 diff --git a/model-00009-of-00061.safetensors b/model-00009-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1a6996797b5ef447313d5240028d4e44192b366d --- /dev/null +++ b/model-00009-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61d3cc6cf7a80abc9015aba9869871acddead21d566604f799e135e17fab6276 +size 4362142872 diff --git a/model-00010-of-00061.safetensors b/model-00010-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c444c0d8437bb1aa8af723bbbdf1a384f0bcb410 --- /dev/null +++ b/model-00010-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5261d70a37247052a9d3900b49e275d96fe0ad9374fc90caa8977d9bebed9680 +size 4966188880 diff --git a/model-00011-of-00061.safetensors b/model-00011-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7ca7a219b078b1513caef647066d2f9bf8bc46e3 --- /dev/null +++ b/model-00011-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b276fc79d14915ba95ba4cc6a77b3cf8b010de5e5f78a4397238c34d3144212d +size 4362142872 diff --git a/model-00012-of-00061.safetensors b/model-00012-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..87f9ba397e93ec2b785b1ee8bb371cf1fd25f121 --- /dev/null +++ b/model-00012-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91043437eec7704507ca6bf82ec8d564fe24474852b06a811a63f1c01c627a0f +size 4362142872 diff --git a/model-00013-of-00061.safetensors b/model-00013-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a3a5e2b2bc31263b73bd5bcb8663881d800a6be5 --- /dev/null +++ b/model-00013-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c90ae72ddb6225693f2f00ad71096643ec7ec13aee7ef7a904aa34a44bb4f733 +size 4966188880 diff --git a/model-00014-of-00061.safetensors b/model-00014-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c2211a1cc3b2d14ba1088f0dab5e0895f6ae2caf --- /dev/null +++ b/model-00014-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7dce4008765103091b1aa60d15351d9da1e61f7f662d3651188ed32c9a17aaea +size 4362142872 diff --git a/model-00015-of-00061.safetensors b/model-00015-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7086665c3115ed0ae4e26b94cca9ce792b223f6e --- /dev/null +++ b/model-00015-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c40d64205c0f403e4cfeabdf7790e0a49419e32ef68193cd931a78dfca82e31 +size 4362142872 diff --git a/model-00016-of-00061.safetensors b/model-00016-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0541fa856cf3cdc00aa23a3bf20cd496ed4ab232 --- /dev/null +++ b/model-00016-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a96b7eb0024c767129dbc9fa7bf1d89473510933d2f90243ce630a733895a582 +size 4966188880 diff --git a/model-00017-of-00061.safetensors b/model-00017-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3af8752573db6ffb1bd66e3e3338dcb5070fe9c2 --- /dev/null +++ b/model-00017-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5f41fdf43aa042ae871acbeaff7d4b0ed6a45a061b00430bd4ab8ed622b5d37 +size 4362142872 diff --git a/model-00018-of-00061.safetensors b/model-00018-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..596fd1b56275b12d148b81daa7352a17827a4efb --- /dev/null +++ b/model-00018-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a27c715bd092dc08f19b9ab63000d8b31a860ca80ef064f21cf0969a988bf1f +size 4362142872 diff --git a/model-00019-of-00061.safetensors b/model-00019-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..674090d74fdbc717e53d6009be92afb577e3037d --- /dev/null +++ b/model-00019-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fd935827e24f928bc171bbc6cd3f0cc2eb7685dde88b42d40c5c5e24af5d988 +size 4966188880 diff --git a/model-00020-of-00061.safetensors b/model-00020-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..02f0e2d28fb903158962f81cb249318b24ba5d37 --- /dev/null +++ b/model-00020-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48362a9a0a50369298b43cfc7f727b797c034748d02622ada3a4d9fc2df15e59 +size 4362142872 diff --git a/model-00021-of-00061.safetensors b/model-00021-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3acd9b753fac77e5a4bf2fc9b2ecc777ad2f888f --- /dev/null +++ b/model-00021-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ac62f35f0affe918278a70ba8acceb4da7e1fa7f9c44b52da43ba3b7e779b88 +size 4362142872 diff --git a/model-00022-of-00061.safetensors b/model-00022-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..91ccb91a454410bb025275de0cfe11427802a5c0 --- /dev/null +++ b/model-00022-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4e621fbba0fa7c5489b5b0414997fb8770ec60639c670987045661b45588e27 +size 4966188880 diff --git a/model-00023-of-00061.safetensors b/model-00023-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9d37d9b9e2011838c396926be294a693bb2c16db --- /dev/null +++ b/model-00023-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df211c400fa8150514396545ff65c81bb19ea4ef2826b488e2c780ba081de67b +size 4362142872 diff --git a/model-00024-of-00061.safetensors b/model-00024-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1c82f9eb2f31121970a20449bfd3eefb79303a92 --- /dev/null +++ b/model-00024-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5df0778e1925f24e58bd75c1b0dd9b8e1318352ec8efe3efbff5423bbe6e2090 +size 4362142872 diff --git a/model-00025-of-00061.safetensors b/model-00025-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..40dec711fc06b51d56a8ee1051b6f3fd1e9522bf --- /dev/null +++ b/model-00025-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64570c5c8fb659a6f8891ceafa9b17a132b66be1ce28637e654b047e72d8503c +size 4966188880 diff --git a/model-00026-of-00061.safetensors b/model-00026-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7cb88519680bcb16360d143d4259b3d8249ed897 --- /dev/null +++ b/model-00026-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d555d1c7009232dcb516a2c40a292bbfb201cfafaf9bedf20e5735fc386d3458 +size 4362142872 diff --git a/model-00027-of-00061.safetensors b/model-00027-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..604c1bd2b5717a1b364c903d4478eaf931628999 --- /dev/null +++ b/model-00027-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f44fd24d29aea57a93a3056cd3f00d4374460d23bfff2b50817c5f6c6719c85e +size 4362142872 diff --git a/model-00028-of-00061.safetensors b/model-00028-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d87a8bbe34d293bb1b36d674ef60f8496c169b0c --- /dev/null +++ b/model-00028-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77d7f23960e9f2f04d798802fed83e034295458473e81865878774277fe19d41 +size 4966188880 diff --git a/model-00029-of-00061.safetensors b/model-00029-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d13a385bf8e021c611f197618b8b04ac718afd38 --- /dev/null +++ b/model-00029-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56866aca5707facf39f3557c57f0eb3f40392ef3b3d84ab83eea1d7efa966b31 +size 4362142872 diff --git a/model-00030-of-00061.safetensors b/model-00030-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..16ef8727a1461df19fb41fc8fd988a248f9349f1 --- /dev/null +++ b/model-00030-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:144113d02f0a98d6636f864cd4f3b6d426b7e9db65c85a600ac284f3dc41b4f0 +size 4362142872 diff --git a/model-00031-of-00061.safetensors b/model-00031-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7251bec61a9aa83b8560026113535fac1d2ed170 --- /dev/null +++ b/model-00031-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:755a0a3833b384834653732e00c3350dc84dedef3c9a676230f7c75c65a2fa31 +size 4966188880 diff --git a/model-00032-of-00061.safetensors b/model-00032-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b1c114316b324b05f552579fa6ae2f5e56dcd5ca --- /dev/null +++ b/model-00032-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b79526a83db463a3921e495099c7bc10654e7f6cd5c66ba50fcc02f18df7c855 +size 4362142872 diff --git a/model-00033-of-00061.safetensors b/model-00033-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..345065e08e916f1f12220f947c8ca32a67f4ccff --- /dev/null +++ b/model-00033-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d00b05182920afd34167d1fe2b3db540f0cf52729d353597b927d7c27e00cdc3 +size 4362142872 diff --git a/model-00034-of-00061.safetensors b/model-00034-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..255e7150a30f184e15d7548c9832011bfac801e5 --- /dev/null +++ b/model-00034-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa00155d49241f4df83c4710894c2e4a21b50b5b006705b1176a9a5e6e1bccc2 +size 4966188880 diff --git a/model-00035-of-00061.safetensors b/model-00035-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7f7cb987bcd62ef8afe395511fbe519e117c749b --- /dev/null +++ b/model-00035-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec7a100aa0a9deb83bf36d6ea5010182fc716a9e1be1a4aa0bb4e75f64e800fd +size 4362142872 diff --git a/model-00036-of-00061.safetensors b/model-00036-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0fadf80cb47cc552d1feea24c84ff5d4fae81a18 --- /dev/null +++ b/model-00036-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9deb9aeb48ad82355d59bb907dd3b7d87d8e60dcec399cf82289e976a38949a4 +size 4362142872 diff --git a/model-00037-of-00061.safetensors b/model-00037-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1ffb7746c6426592df195e6f8a8858d638f4ad46 --- /dev/null +++ b/model-00037-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd0f54e3831b3fd36ceb3d7d6d9256e5fc2bb187b47a52637311843a35c504cd +size 4966188880 diff --git a/model-00038-of-00061.safetensors b/model-00038-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3c03aeca859aaadd18406fefc4acde32634ba2bc --- /dev/null +++ b/model-00038-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8667d99ed2be26507aa5869c721f19a7e85fb6afab85855b4fa06d535fb9429 +size 4362142872 diff --git a/model-00039-of-00061.safetensors b/model-00039-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fcf65cdab6b360206b9cf50a7ae905f51ce7d8b8 --- /dev/null +++ b/model-00039-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a935edb896dcea6d0a8f5472531ead49b7e144976f2d2216ced0f07fbfc362e6 +size 4362142872 diff --git a/model-00040-of-00061.safetensors b/model-00040-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2f0c6e003e0a92470ca73eef3fed35845d9edd6b --- /dev/null +++ b/model-00040-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:991af2c2121aa7206d87c9e1f222df228a8d88d72599f559cccab400f2aee0c2 +size 4966188880 diff --git a/model-00041-of-00061.safetensors b/model-00041-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9887e426a3c7129041b68d513c5b27ef10a4e169 --- /dev/null +++ b/model-00041-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d77a4a6dc5f51a7fc3a59a9704e77c809578f63ad386524e94bc76bf0f0974b3 +size 4362142872 diff --git a/model-00042-of-00061.safetensors b/model-00042-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3cd6feda876e519ffac9e44d7a4dc1c7240fddcd --- /dev/null +++ b/model-00042-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb9ae2e11d7f0f1676dbaf34e3f9c56848314e25ecc5849f6feb55f27e66e855 +size 4362142872 diff --git a/model-00043-of-00061.safetensors b/model-00043-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a459203179af403ca87bf0d407fa5c5837ad5326 --- /dev/null +++ b/model-00043-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5646a0b1c8215f5fd5c6a59526a05aed00f686c0bd005eba139051d791abf0d3 +size 4966188880 diff --git a/model-00044-of-00061.safetensors b/model-00044-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ccc64cb664ac16a8ca7b94d6b952f7a36c345caf --- /dev/null +++ b/model-00044-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9da471ee76441edcd9f1efc071439e4540a09f0f04fb69bc7184ec2c6b0ec79 +size 4362142872 diff --git a/model-00045-of-00061.safetensors b/model-00045-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..72b286c80b6f28ef5addfc77b29622309a745ce4 --- /dev/null +++ b/model-00045-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ed41d5f1ae3737efdd2fe9a0a09d824b94303df54bce628908bcf69e4a53a8c +size 4362142872 diff --git a/model-00046-of-00061.safetensors b/model-00046-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b9391759930ee2e6af5de9588bd4db0a56f1bf42 --- /dev/null +++ b/model-00046-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:683a9c798d76f22640bc715532d891aded93b91edec6597ca99e1a662a186bae +size 4966188880 diff --git a/model-00047-of-00061.safetensors b/model-00047-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..81b7ecee79cad42c55a071207ac2274e4a9d2781 --- /dev/null +++ b/model-00047-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0704410f32e5cc7217a4a2e70716c82d9b436ca6089bf4515b348e6a030e297c +size 4362142872 diff --git a/model-00048-of-00061.safetensors b/model-00048-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..255bfd8203212f2977531bd95464d4708d1196ba --- /dev/null +++ b/model-00048-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90283f82c2ee4af1e59c080c5e9d4c4fab5469b61689b98ec78d193f70bca1eb +size 4362142872 diff --git a/model-00049-of-00061.safetensors b/model-00049-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..db90d027c2d49f3cc7c4e4197d7ee1e236704c19 --- /dev/null +++ b/model-00049-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a44edd6ab0df1338f0277e6b408d2fe47512da42be1c8b4e1e272494653cde5 +size 4966188880 diff --git a/model-00050-of-00061.safetensors b/model-00050-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..860278705821b1c18f59e2e4aa742526b7794cda --- /dev/null +++ b/model-00050-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c02a4d8f11b3d77bce34045d00df75690c5047e605467b367cfd501c8ba5594f +size 4362142872 diff --git a/model-00051-of-00061.safetensors b/model-00051-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0b9be030892f7c2bf3efce47f216f507c7068323 --- /dev/null +++ b/model-00051-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51a1e9aa170fe43083ce1e49856edb11845281891877378589c8a3d28a20f79a +size 4362142872 diff --git a/model-00052-of-00061.safetensors b/model-00052-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ac647a44d2293e30a171554269de205e20eee025 --- /dev/null +++ b/model-00052-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:623fea8fa44715afd32b3835f5f1f3d3d582c62e4726383a8491fb61d7700cbd +size 4966188880 diff --git a/model-00053-of-00061.safetensors b/model-00053-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7855177a8d6acfb46a1f593b91ba80d7ee07121b --- /dev/null +++ b/model-00053-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:445f12c7aedc9f2533202fbba1ab0e210fd607efa449e3a548dcbda61c7336a1 +size 4362142872 diff --git a/model-00054-of-00061.safetensors b/model-00054-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..73feed49ce650829ce7448b7e060bc0447883fa1 --- /dev/null +++ b/model-00054-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ec69a703f3dbd849a705530eb71544c599dfb39f6451e505b02df1112a98e86 +size 4362142872 diff --git a/model-00055-of-00061.safetensors b/model-00055-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b6cf8b05a2aefcf78dc5473030af3db8a153b31c --- /dev/null +++ b/model-00055-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:63c369aefbb6b4131c208f5437d9f3b54e202df8329df14499db6fa9043cf37d +size 4966188880 diff --git a/model-00056-of-00061.safetensors b/model-00056-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5795e8854306ef3079c87e010161830be5e647fc --- /dev/null +++ b/model-00056-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0697fdf578a039bb4f6bcf74c78342e223d3f2358a05e4a56301e3e603e7cfd4 +size 4362142872 diff --git a/model-00057-of-00061.safetensors b/model-00057-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..17192680f6f743f294f42759db127c51cbcd3098 --- /dev/null +++ b/model-00057-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58651e577271f1c1a838969372589fc719969558c10e2ba8dafceea46d837aee +size 4362142872 diff --git a/model-00058-of-00061.safetensors b/model-00058-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..22018bddd66b0627fe7b9a77f2b7b67a4b1400cc --- /dev/null +++ b/model-00058-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4ec05ef7524c7511e143f20b3e20ebf7011fa2f515130609f5c81a63bb8d2fa +size 4966188880 diff --git a/model-00059-of-00061.safetensors b/model-00059-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3fdf1dd4fec5a3c831fc793522e110e814dc0b3c --- /dev/null +++ b/model-00059-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c97104ae34c099c34d261ba93a199c699db80b97aace32844f0470aeb869bd0 +size 4362142872 diff --git a/model-00060-of-00061.safetensors b/model-00060-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0ecbc0d565f967ad680b1e761d80f09bd78e3609 --- /dev/null +++ b/model-00060-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d91ed3235c65733cc332c32ad21506820d895cfe0e0c03672e011954cc4a9ff +size 4362142872 diff --git a/model-00061-of-00061.safetensors b/model-00061-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1d048079c2e3fda364ff3978fadf34e8ceff963a --- /dev/null +++ b/model-00061-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f211436028040e1988b1f088ff96737441b9fd043055cdd5d6de9248fbdaa75 +size 1988198960 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..4156aa9dfeed60097cda66ec60af33ad1607de57 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,730 @@ +{ + "metadata": { + "total_size": 275906592768 + }, + "weight_map": { + "lm_head.weight": "model-00061-of-00061.safetensors", + "model.embed_tokens.weight": "model-00001-of-00061.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00061.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00061.safetensors", + "model.layers.10.input_layernorm.weight": "model-00009-of-00061.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00061.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.11.input_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00061.safetensors", + "model.layers.12.input_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.input_layernorm.weight": "model-00011-of-00061.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00061.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00061.safetensors", + "model.layers.14.input_layernorm.weight": "model-00012-of-00061.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00012-of-00061.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00061.safetensors", + "model.layers.15.input_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00012-of-00061.safetensors", + "model.layers.16.input_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.input_layernorm.weight": "model-00014-of-00061.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00014-of-00061.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00061.safetensors", + "model.layers.18.input_layernorm.weight": "model-00015-of-00061.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00015-of-00061.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00014-of-00061.safetensors", + "model.layers.19.input_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00015-of-00061.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00061.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00061.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00061.safetensors", + "model.layers.20.input_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.input_layernorm.weight": "model-00017-of-00061.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00017-of-00061.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00061.safetensors", + "model.layers.22.input_layernorm.weight": "model-00018-of-00061.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00061.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00061.safetensors", + "model.layers.23.input_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00061.safetensors", + "model.layers.24.input_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.input_layernorm.weight": "model-00020-of-00061.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00061.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00061.safetensors", + "model.layers.26.input_layernorm.weight": "model-00021-of-00061.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00061.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00020-of-00061.safetensors", + "model.layers.27.input_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00021-of-00061.safetensors", + "model.layers.28.input_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.input_layernorm.weight": "model-00023-of-00061.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00023-of-00061.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00022-of-00061.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00061.safetensors", + "model.layers.30.input_layernorm.weight": "model-00024-of-00061.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00024-of-00061.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00023-of-00061.safetensors", + "model.layers.31.input_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00024-of-00061.safetensors", + "model.layers.32.input_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.input_layernorm.weight": "model-00026-of-00061.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00026-of-00061.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00025-of-00061.safetensors", + "model.layers.34.input_layernorm.weight": "model-00027-of-00061.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00027-of-00061.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00026-of-00061.safetensors", + "model.layers.35.input_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00027-of-00061.safetensors", + "model.layers.36.input_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.input_layernorm.weight": "model-00029-of-00061.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00029-of-00061.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00028-of-00061.safetensors", + "model.layers.38.input_layernorm.weight": "model-00030-of-00061.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00030-of-00061.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00029-of-00061.safetensors", + "model.layers.39.input_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00030-of-00061.safetensors", + "model.layers.4.input_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.40.input_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.input_layernorm.weight": "model-00032-of-00061.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00032-of-00061.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00031-of-00061.safetensors", + "model.layers.42.input_layernorm.weight": "model-00033-of-00061.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00033-of-00061.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00032-of-00061.safetensors", + "model.layers.43.input_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00033-of-00061.safetensors", + "model.layers.44.input_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.input_layernorm.weight": "model-00035-of-00061.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00035-of-00061.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00034-of-00061.safetensors", + "model.layers.46.input_layernorm.weight": "model-00036-of-00061.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00036-of-00061.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00035-of-00061.safetensors", + "model.layers.47.input_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00036-of-00061.safetensors", + "model.layers.48.input_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.input_layernorm.weight": "model-00038-of-00061.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00038-of-00061.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00037-of-00061.safetensors", + "model.layers.5.input_layernorm.weight": "model-00005-of-00061.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00061.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00061.safetensors", + "model.layers.50.input_layernorm.weight": "model-00039-of-00061.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00039-of-00061.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00038-of-00061.safetensors", + "model.layers.51.input_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00039-of-00061.safetensors", + "model.layers.52.input_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.input_layernorm.weight": "model-00041-of-00061.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00041-of-00061.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00040-of-00061.safetensors", + "model.layers.54.input_layernorm.weight": "model-00042-of-00061.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00042-of-00061.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00041-of-00061.safetensors", + "model.layers.55.input_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00042-of-00061.safetensors", + "model.layers.56.input_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.input_layernorm.weight": "model-00044-of-00061.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00044-of-00061.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00043-of-00061.safetensors", + "model.layers.58.input_layernorm.weight": "model-00045-of-00061.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00045-of-00061.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00044-of-00061.safetensors", + "model.layers.59.input_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00045-of-00061.safetensors", + "model.layers.6.input_layernorm.weight": "model-00006-of-00061.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00061.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00061.safetensors", + "model.layers.60.input_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.input_layernorm.weight": "model-00047-of-00061.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00047-of-00061.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00046-of-00061.safetensors", + "model.layers.62.input_layernorm.weight": "model-00048-of-00061.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00048-of-00061.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00047-of-00061.safetensors", + "model.layers.63.input_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00048-of-00061.safetensors", + "model.layers.64.input_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.input_layernorm.weight": "model-00050-of-00061.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00050-of-00061.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00049-of-00061.safetensors", + "model.layers.66.input_layernorm.weight": "model-00051-of-00061.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00051-of-00061.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00050-of-00061.safetensors", + "model.layers.67.input_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00051-of-00061.safetensors", + "model.layers.68.input_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.input_layernorm.weight": "model-00053-of-00061.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00053-of-00061.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00052-of-00061.safetensors", + "model.layers.7.input_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00061.safetensors", + "model.layers.70.input_layernorm.weight": "model-00054-of-00061.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00054-of-00061.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00053-of-00061.safetensors", + "model.layers.71.input_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00054-of-00061.safetensors", + "model.layers.72.input_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.input_layernorm.weight": "model-00056-of-00061.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00056-of-00061.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00055-of-00061.safetensors", + "model.layers.74.input_layernorm.weight": "model-00057-of-00061.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00057-of-00061.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00056-of-00061.safetensors", + "model.layers.75.input_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00057-of-00061.safetensors", + "model.layers.76.input_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.input_layernorm.weight": "model-00059-of-00061.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00059-of-00061.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00058-of-00061.safetensors", + "model.layers.78.input_layernorm.weight": "model-00060-of-00061.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00060-of-00061.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00059-of-00061.safetensors", + "model.layers.79.input_layernorm.weight": "model-00061-of-00061.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00061-of-00061.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00061.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00060-of-00061.safetensors", + "model.layers.8.input_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.input_layernorm.weight": "model-00008-of-00061.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00008-of-00061.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00061.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00061.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00061.safetensors", + "model.norm.weight": "model-00061-of-00061.safetensors" + } +}