diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..bc5f30d6632ac0efdc7be2e9095e9e9579af2e33 --- /dev/null +++ b/README.md @@ -0,0 +1,199 @@ +--- +library_name: transformers +tags: [] +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + +This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3300e7fdab2681880604f1da0d8e54d43fa7dcf5 --- /dev/null +++ b/config.json @@ -0,0 +1,36 @@ +{ + "_name_or_path": "nbeerbower/Llama-3.1-Saoirse-70B", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "eos_token_id": 128039, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 8.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.48.0", + "use_cache": true, + "vocab_size": 128256 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..79ad8218b543f71675a827fe6405ab3ca519ee2f --- /dev/null +++ b/generation_config.json @@ -0,0 +1,6 @@ +{ + "_from_model_config": true, + "bos_token_id": 128000, + "eos_token_id": 128039, + "transformers_version": "4.48.0" +} diff --git a/model-00001-of-00082.safetensors b/model-00001-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d062dcf1cd74c50fd56ff49a30f0f1a5c6fcbe90 --- /dev/null +++ b/model-00001-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9ba902863e9afb2f0b35a7b9984b4bceff217d6551f8a7c772005eaf9917bbd +size 2101346448 diff --git a/model-00002-of-00082.safetensors b/model-00002-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..98b3b19224ee71a9b521922d1c1fff99f638e549 --- /dev/null +++ b/model-00002-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b7bf851254bf26c0b501be460f6d639bec29d9b49c48ce6a219e4655c2defdc +size 1879082376 diff --git a/model-00003-of-00082.safetensors b/model-00003-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b92c48e3fb8d8ce5ef8a7370d4c33f57ec8e85b4 --- /dev/null +++ b/model-00003-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a6d0f1f14e96086d63a2b278d3a4a5d47f03d8d7575d14e524935f6eaf76001 +size 1845527704 diff --git a/model-00004-of-00082.safetensors b/model-00004-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7500b255a0358e42987e50fec73867714b274981 --- /dev/null +++ b/model-00004-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c27fa407688595d41a5b87737040f03f0e0ef93ffb0c338dc8e0553ce74869b +size 1711309864 diff --git a/model-00005-of-00082.safetensors b/model-00005-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0a6721112ec63170ee604486f1a7424299535a88 --- /dev/null +++ b/model-00005-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:af35bf8c649816b87d89f6e456118e2d5cf75c279b7a8df9b4c511fb3324a4d9 +size 1711309864 diff --git a/model-00006-of-00082.safetensors b/model-00006-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1f4c9bfddfefdf790595725a1041c2f388a5c3ef --- /dev/null +++ b/model-00006-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:357543ff7c4dbb802bd05deb4831c39d16fb54861eb70760658962f10fdbe7ea +size 1711309864 diff --git a/model-00007-of-00082.safetensors b/model-00007-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..edea4ebd397ec028ac05537b2fc623fcb50385bd --- /dev/null +++ b/model-00007-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58d9c80c56612b49ad95fb1c10eb355fe071f69e1d9ad39a0b1c3757b258f48c +size 1711309864 diff --git a/model-00008-of-00082.safetensors b/model-00008-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a6cb257f5c690e1bd81e84592f36188c5e318df6 --- /dev/null +++ b/model-00008-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3021a1aab7adc45c49e8a4e735ab73a9b1bf4c23df72d701070b7a8d803dacee +size 1711309864 diff --git a/model-00009-of-00082.safetensors b/model-00009-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d07ce586a92c9bb48936c04412706b2f8045c634 --- /dev/null +++ b/model-00009-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05f7460ac6bc9ce57d5c3fe2433ae8ee8b42423d824f13b8e4b23221940ca402 +size 1711309864 diff --git a/model-00010-of-00082.safetensors b/model-00010-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4b1183bb4b73fd1ea3f3005b0ea4fbf43c243791 --- /dev/null +++ b/model-00010-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ceb925c2d80561de34dea5f5075fd704b4e321a625ed7abecb472972968bbc1d +size 1711309864 diff --git a/model-00011-of-00082.safetensors b/model-00011-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b51b69c199feb546e9b1e8a6cba994f0163d6af0 --- /dev/null +++ b/model-00011-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48cb47d13ee801cd9e9907359ffad30a5f9220ff7261e9a8e3e7005e47109712 +size 1711309864 diff --git a/model-00012-of-00082.safetensors b/model-00012-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2092fe778e8334f5efb5fd502f68a59639c6bf01 --- /dev/null +++ b/model-00012-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46b6e0407c4d1fec98e7809f65791b33072b6e2efc447164e9b189a96846cbdd +size 1711309872 diff --git a/model-00013-of-00082.safetensors b/model-00013-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a365b12f249db7bb8266c4f8c36a2aa744143b69 --- /dev/null +++ b/model-00013-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:100d8dcf7d0bb11ef5f2ba351c3aeec2df918a6e1224d4c8b11472cf3e00c51b +size 1711309872 diff --git a/model-00014-of-00082.safetensors b/model-00014-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1268344fd74fb355a18d652a6eb4e4e4377fc847 --- /dev/null +++ b/model-00014-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fcfd56efb2c1277a8874c1f901bcfedc76487b3d92c8dffbf8b4eabfd7f12831 +size 1711309872 diff --git a/model-00015-of-00082.safetensors b/model-00015-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f4594cd4cfd78a4382a4be66ca84f5abc80ba594 --- /dev/null +++ b/model-00015-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fbc2986a06482c9cea91fed68f4baa6c4133668621685cd5c12cbbd342e8914 +size 1711309872 diff --git a/model-00016-of-00082.safetensors b/model-00016-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e9bcf4ad32846b20d39d6ba72c9383e4197bfe8b --- /dev/null +++ b/model-00016-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7aea13849015156173a0d46ba9998071946a03b365254a68d02751338c82608e +size 1711309872 diff --git a/model-00017-of-00082.safetensors b/model-00017-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0f2a94bb295e99df0f1627002b658f8cdbc8e63b --- /dev/null +++ b/model-00017-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0008bd30ee0d2faedcccd7be8896296878b029902fdfc5ad1680585b9d55f2d1 +size 1711309872 diff --git a/model-00018-of-00082.safetensors b/model-00018-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..931491b852258beeb4cf57350713e8e37769db85 --- /dev/null +++ b/model-00018-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e35c6642be2b6f9e42cf1d009eace820ffc3a5f68c68885416df28d6cbca989e +size 1711309872 diff --git a/model-00019-of-00082.safetensors b/model-00019-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d23609cbfba5623aea2770b861bf66f2a65f2be4 --- /dev/null +++ b/model-00019-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd3de92d24c01d21235971c87d9a6b993c6092eea624c8cfb9b8331dfcfb55bd +size 1711309872 diff --git a/model-00020-of-00082.safetensors b/model-00020-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6b6e91628cebd5cecf9436d6b5dce144bd767d6e --- /dev/null +++ b/model-00020-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e5a745b215f2df6e8172992a1841de7f7f678cd8836a0d7ffe51a95a1e223eed +size 1711309872 diff --git a/model-00021-of-00082.safetensors b/model-00021-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0629e643dcdab8d82e2e968319c0032066fd1ebb --- /dev/null +++ b/model-00021-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:591bf751cca4e28778f6c0ded5c7afed6dcab5bacb1bd5b852f57b80f14b50e2 +size 1711309872 diff --git a/model-00022-of-00082.safetensors b/model-00022-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d7284e82be326308dda3a2c2ba6e0dfd4a448ec --- /dev/null +++ b/model-00022-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:904751cc00cf100d0af5f2404e1318c7a89091e1adbc5bd0f68cf5b9ff317d17 +size 1711309872 diff --git a/model-00023-of-00082.safetensors b/model-00023-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5bc2a74f5336656ba892cb96c2f90bdff8ac53a4 --- /dev/null +++ b/model-00023-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d45f1d81fcb73fff3dfd17583e2d6882d61ca0a68b86775de04a1d6b55204d0 +size 1711309872 diff --git a/model-00024-of-00082.safetensors b/model-00024-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8bcd9bb1bd8d1e8d203c891b41d2b1bd8a5d6595 --- /dev/null +++ b/model-00024-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf425036dd1ccb5cdeb5ceafdcd4c94a385a3f690d7d310151d80ec0c4bc8223 +size 1711309872 diff --git a/model-00025-of-00082.safetensors b/model-00025-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7ad6f97185eb09b8a292177dc83f28c643e9f7af --- /dev/null +++ b/model-00025-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:788bb20d06f45bb8006e678da9052373ebdd624076eefeebcc128e3e2cfd128e +size 1711309872 diff --git a/model-00026-of-00082.safetensors b/model-00026-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f0984c4c4888b34a9b114c5fb1c0f9af2baaa04d --- /dev/null +++ b/model-00026-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a796102954bdc83ce1e428fb0757fa26e8d3105809943548073a2c8a9e62df05 +size 1711309872 diff --git a/model-00027-of-00082.safetensors b/model-00027-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..71ec69ec92549f1d2d9f0834ca859ff2106371b8 --- /dev/null +++ b/model-00027-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:753a4fd77913a56253a805bdb045ef641da76e54a91c1528aac1dd7947c7e6f2 +size 1711309872 diff --git a/model-00028-of-00082.safetensors b/model-00028-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b6ade67d81ff72e0e75d05325367c471d4543662 --- /dev/null +++ b/model-00028-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1c92b9cae619cfc288c66d310719efb53ebbfa7d5b517bf8095a4d0374c3561 +size 1711309872 diff --git a/model-00029-of-00082.safetensors b/model-00029-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6e8e47ceadf269ffaf7d4bfccf1f8880b5a1c11d --- /dev/null +++ b/model-00029-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad25199ee237d19722c6927ac074368fab61bed4006dfec19886f50ed0448cd2 +size 1711309872 diff --git a/model-00030-of-00082.safetensors b/model-00030-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3d73cd8a6bfd7c967ccda7529168a86f8af0257d --- /dev/null +++ b/model-00030-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31d01a819c35e94b5d78d58ecdfe3b3a7ee3e326b5cb3ce22ead2aa922c01328 +size 1711309872 diff --git a/model-00031-of-00082.safetensors b/model-00031-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e8f7ccc9ea5baadc62f001205a6bf83911728e78 --- /dev/null +++ b/model-00031-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:241565f06b8a8532993b9c5389256a334181b7226a26b2ac19ae4b8164edd2d5 +size 1711309872 diff --git a/model-00032-of-00082.safetensors b/model-00032-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a2a1abf26131bcb9a4aeda5c5ef7bd595c880d04 --- /dev/null +++ b/model-00032-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00ec86e796b235f519f3112b8d210dac91dc9cc6475c2f28a8f23b32ab32bef9 +size 1711309872 diff --git a/model-00033-of-00082.safetensors b/model-00033-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6e6f229c88187d6e224ae995b4c9ce3ec607fdcf --- /dev/null +++ b/model-00033-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a8941bda53ea459ba77a4851699dc3ad39a61cfb5bfa09a89d66bd176063094 +size 1711309872 diff --git a/model-00034-of-00082.safetensors b/model-00034-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..636674c6abfe1090cc546da8c51232605589bf9e --- /dev/null +++ b/model-00034-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bfbc006e0b6a192d832dca7f92808980cc8da4df2e4057fb6458f408a3ba0f8 +size 1711309872 diff --git a/model-00035-of-00082.safetensors b/model-00035-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..869ef4aae359a99f61800908d589d7ec86c12a9f --- /dev/null +++ b/model-00035-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98bbfcc5571ed4e6808e459c26129dba61d3c3d79b4d4fbd88866560d406d43c +size 1711309872 diff --git a/model-00036-of-00082.safetensors b/model-00036-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9a2323e6fa54506f1aa15445e53b539d64eeadb8 --- /dev/null +++ b/model-00036-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04c480c70e55d0daf85f1e1ccb5ed1074e036c255cbaf853fcf032ae0bc4498d +size 1711309872 diff --git a/model-00037-of-00082.safetensors b/model-00037-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..62b9d7788c2650d24dbffb9b634929bdef1a7bb3 --- /dev/null +++ b/model-00037-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6fb383f53f377d6cadbf5d4cdd2553a1db40852d86274281856cda0a100f943 +size 1711309872 diff --git a/model-00038-of-00082.safetensors b/model-00038-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d4b13ea85d292e50269f1791ceca732dd2bd738e --- /dev/null +++ b/model-00038-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c9f1b043d407bcf88365f36380f583a928a00d18e1d117ed883ae30a9c1f70d +size 1711309872 diff --git a/model-00039-of-00082.safetensors b/model-00039-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0c1d1ed45f26f828790121840673178e4bf54391 --- /dev/null +++ b/model-00039-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a586a646e83ed9e5c633af18336e6c92bdeb25cf0ed67fe192158436ea4d2ee0 +size 1711309872 diff --git a/model-00040-of-00082.safetensors b/model-00040-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..86edc0130d7d6e85d80b5dd4d4266d74c94ad9e2 --- /dev/null +++ b/model-00040-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9af6e1aabbec68383af9aa1dc7092ec0bb15b98ef8c4aa10232b97397a7c9d7b +size 1711309872 diff --git a/model-00041-of-00082.safetensors b/model-00041-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b467b3b5330d1960e917722f9b1920512b78bdc1 --- /dev/null +++ b/model-00041-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1fb070bc3ab058278e3f468d57f047de95017262e8787c2dcbb750bae18df818 +size 1711309872 diff --git a/model-00042-of-00082.safetensors b/model-00042-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cd3b4ad72ff3a385bc3b8a20080b3b10c0e674a9 --- /dev/null +++ b/model-00042-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0359a020969f5ed6f97d53096cea108869b851465f9d0901c2362c2ed66ae854 +size 1711309872 diff --git a/model-00043-of-00082.safetensors b/model-00043-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..20aeaf10983fbe1c668efcf5c9c8efcfc597f36f --- /dev/null +++ b/model-00043-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf71546d3b907fefef34eafe685ae246e434919655a29dd0594e79c05f2586da +size 1711309872 diff --git a/model-00044-of-00082.safetensors b/model-00044-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..29dc11582337cba2dea57e41cd693435204ae58b --- /dev/null +++ b/model-00044-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb6ff5bfb97ddfa9a83c69d97f936690825264946d7806da1ebb1de3630e60fd +size 1711309872 diff --git a/model-00045-of-00082.safetensors b/model-00045-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1616427157d0efaeed344f8d76d817f7ee62882c --- /dev/null +++ b/model-00045-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87b2aeb3252a3bc827d5a24bc5fc2c0a611a2adfe2effcde0ed08b8baf43782d +size 1711309872 diff --git a/model-00046-of-00082.safetensors b/model-00046-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f58d29194b6f3ea8f52797ea8ccd45a97c3316c9 --- /dev/null +++ b/model-00046-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f156e9a11728283185f20e7ee63d1f783829d7c052a2e2a0c99c1831a52c818e +size 1711309872 diff --git a/model-00047-of-00082.safetensors b/model-00047-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..68ea5716aed5c9de119c43b7f48519cc552ca92e --- /dev/null +++ b/model-00047-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73aa76b8b7ebfedd278087414e9531c0848b405bee09083866339f430f4dd201 +size 1711309872 diff --git a/model-00048-of-00082.safetensors b/model-00048-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4830be49e1b6ddcc0510b82949ed89e87693e07d --- /dev/null +++ b/model-00048-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6bc40efdd9800180155bb319e72dab21929a7733d79e546593c8b83f445fd153 +size 1711309872 diff --git a/model-00049-of-00082.safetensors b/model-00049-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cefb640da1e8f25e3c6ee3f823312f102669cb7e --- /dev/null +++ b/model-00049-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bcd1224fa4dfe323e3a6068df9638e5f1b534456cc01af32f4f6beaada499b4 +size 1711309872 diff --git a/model-00050-of-00082.safetensors b/model-00050-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..609f282ba864a6ff60884d556510f92a2c71b488 --- /dev/null +++ b/model-00050-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33a98f0a0f5660a762f7d3d523710168e38a5bceb79b2df33feaeab36bdb38b3 +size 1711309872 diff --git a/model-00051-of-00082.safetensors b/model-00051-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dcedc928cbf05b2489fc4fceaa93b7180ea1de9c --- /dev/null +++ b/model-00051-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b162ac79c57db4e505d41f4ae97f244b163cd56cce4c3926ad9bf1fb1757b0a +size 1711309872 diff --git a/model-00052-of-00082.safetensors b/model-00052-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3e12829e73d47cc1a3a4ef1c964a2eabb266c292 --- /dev/null +++ b/model-00052-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb4894fb39f8f0948862de29dc22044b98c05a7c44008a95b205e9f35caaba89 +size 1711309872 diff --git a/model-00053-of-00082.safetensors b/model-00053-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9d3a134ea37ee950ec55797f133e1a2e25822f79 --- /dev/null +++ b/model-00053-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59f96393cb28dac163cc1d3cfd86d9553ecb2170c32aee2afcfc80308a057fb5 +size 1711309872 diff --git a/model-00054-of-00082.safetensors b/model-00054-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ed89d5704faa62763803b4a93bc3571d3f878268 --- /dev/null +++ b/model-00054-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f522677924ef6f90d97283cf0f5299133e9fb2d18464b3db0776d01e9fc9720 +size 1711309872 diff --git a/model-00055-of-00082.safetensors b/model-00055-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..28bd850df5e6dc4bfc2f4d0e175504260cacec1c --- /dev/null +++ b/model-00055-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8463881925550342d1a63eb9cbebac02c9401ae13c6ef918e9cb46217800dad0 +size 1711309872 diff --git a/model-00056-of-00082.safetensors b/model-00056-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..99c50df8f09a71eddd34dcea42d0da0de30be8f1 --- /dev/null +++ b/model-00056-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7a64f67b355742e8b45a0e69a741c73169f7e8d2032ad188766a5269ab85297 +size 1711309872 diff --git a/model-00057-of-00082.safetensors b/model-00057-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a4e37b9e7557e53a6ee212a3329169ae7c4e96e4 --- /dev/null +++ b/model-00057-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c933d771c3e9ee6200f39877268a41ff2b790eccbff8fef3676dd03d50d96fc9 +size 1711309872 diff --git a/model-00058-of-00082.safetensors b/model-00058-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..70c7147779194dd78920fc4633c53e74cf6c6bda --- /dev/null +++ b/model-00058-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea69803a7ce566c10c282c665d2beff5b2240df8e7af9f54ad094c0963f5e611 +size 1711309872 diff --git a/model-00059-of-00082.safetensors b/model-00059-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6e0b954e8c3b4e2396ba1e5ae703dcf7f797398d --- /dev/null +++ b/model-00059-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebc0f132ae6977d544c7a6dd7cda4ad8d8cc8bdea091321c427f66e8605fe85e +size 1711309872 diff --git a/model-00060-of-00082.safetensors b/model-00060-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1f384ddc43fda3c1f2c96a53ad3b1c536c36c3fa --- /dev/null +++ b/model-00060-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:228628db089dfbbdf23d3023611813376592eb36f69361712fa75949c612095a +size 1711309872 diff --git a/model-00061-of-00082.safetensors b/model-00061-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b4260936619c894c356af6b7048e04124fbea3e9 --- /dev/null +++ b/model-00061-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88d7affcb9581df207380a855dd08bb1d28d88622af28850e7787cfe80f9d8bf +size 1711309872 diff --git a/model-00062-of-00082.safetensors b/model-00062-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7e6a2335a29b676c43d7662ffde1e7975d44f16f --- /dev/null +++ b/model-00062-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5144c484edf8d9fb887d1962a5c6704c4c63ccf9e3bebde106e900dee2627f2 +size 1711309872 diff --git a/model-00063-of-00082.safetensors b/model-00063-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..56b5453e2b7a265cfbdc10d9d65d3c0eca720974 --- /dev/null +++ b/model-00063-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb3a59fae4bf06be0f8ba8b16913fd605ede80090b308bb8ff19eea9fe2f269f +size 1711309872 diff --git a/model-00064-of-00082.safetensors b/model-00064-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1d8beee91d104d180b48b37d1f9ea7ae0b5ade43 --- /dev/null +++ b/model-00064-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b00f4ceb41a64374d4b83a66ebbfe70370f645320e4085a55d267701835fbd55 +size 1711309872 diff --git a/model-00065-of-00082.safetensors b/model-00065-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d09fefc04c6410021b3c0da6d8abe9e59b1db7a --- /dev/null +++ b/model-00065-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5a3cea9b775d98fb4a31909bb94261b3dc016d9aad048c6786e0b5174e4bd15 +size 1711309872 diff --git a/model-00066-of-00082.safetensors b/model-00066-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4e74213e89d3c11e6321ac3e05a5a6eb3de336a2 --- /dev/null +++ b/model-00066-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a388b76ebed14277cc454ba033457bb11f62db95426f17487de9368bcdf94aa1 +size 1711309872 diff --git a/model-00067-of-00082.safetensors b/model-00067-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..482961d793a69145bb6d5c3a8657e5ed6d275f6a --- /dev/null +++ b/model-00067-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85058d2f521ebf4fc28aab7ad735ca6960adf4aa4ac79aa7388a315fb479f6ce +size 1711309872 diff --git a/model-00068-of-00082.safetensors b/model-00068-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cee7cf87ef7502559019ca8349d7a0f447c66dff --- /dev/null +++ b/model-00068-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f55240e692e5567cd7150d17ac90b32ecdb3d87a876cc82d07365d040fbddc1d +size 1711309872 diff --git a/model-00069-of-00082.safetensors b/model-00069-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d2a58626489347d50ea985dd70238e0c22434140 --- /dev/null +++ b/model-00069-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98bbc4740d37d0d8e764269dc85ff74bda20451add9e56729e2be28b277e5b9d +size 1711309872 diff --git a/model-00070-of-00082.safetensors b/model-00070-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..99c537eb850295d936e96319a505ddbd5fbce1c2 --- /dev/null +++ b/model-00070-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1a7e7b85befadc6a99e24145304f0cf7b4d2c83b329ce90ccd0d8b73ef144e3 +size 1711309872 diff --git a/model-00071-of-00082.safetensors b/model-00071-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9e4d9ec702a09d1833931d89d76ab1c1a7d4699b --- /dev/null +++ b/model-00071-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:988d0e757837f91300b29b0d85cdb51f8b9e9cd03e24b2f9361f040e886a063c +size 1711309872 diff --git a/model-00072-of-00082.safetensors b/model-00072-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f85d1a18185198c7695d9637d37df5de46fb70b1 --- /dev/null +++ b/model-00072-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d5221ef00ec93c5226807a26e1f38771f2a060357c02d78abc6327bb50db044 +size 1711309872 diff --git a/model-00073-of-00082.safetensors b/model-00073-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fd4edd55789f9e5ef001cdbfb0e07991ac1bba7d --- /dev/null +++ b/model-00073-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a04582bd28e57f8e82432fa497123031986cf41672bf4bb55d9fb8ad66d9fc57 +size 1711309872 diff --git a/model-00074-of-00082.safetensors b/model-00074-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ccba77c94ac44c5e4e31abd861ad1bf0967d578c --- /dev/null +++ b/model-00074-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e0350f06aff2e05e5d7cbfb9fcf33418e6d1e1262e92d12faf62c5096b12e0b +size 1711309872 diff --git a/model-00075-of-00082.safetensors b/model-00075-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3a51d1cea6db0ab9d4fe140e15285d1a009e5be4 --- /dev/null +++ b/model-00075-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e6c943fc3a00acdec078c64e9e81fdc138a7334704562e744aa73f356410352 +size 1711309872 diff --git a/model-00076-of-00082.safetensors b/model-00076-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..36960379af610c4c3f14e1a8454266b78cb0021b --- /dev/null +++ b/model-00076-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba39ed18164b691342a3ff144e8b8ef7c654787631b3e8d4bf7168cadbba736c +size 1711309872 diff --git a/model-00077-of-00082.safetensors b/model-00077-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..118a6f09d1fa8f8781f170f2c7d50525042dd43a --- /dev/null +++ b/model-00077-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebc39841c929dfbf7545d0963523a3cb13a06ca050f9bd42c1664f26bc6fb4a3 +size 1711309872 diff --git a/model-00078-of-00082.safetensors b/model-00078-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..618f85718c29c1e883bc9537afbaed6651ede148 --- /dev/null +++ b/model-00078-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:932ad6038051415b74c18c86980a65e11600d210831881cd5ff922f7bd7bd1d5 +size 1711309872 diff --git a/model-00079-of-00082.safetensors b/model-00079-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..82ca9d4504c8e8a94e5a0a6564f4253c499b06bc --- /dev/null +++ b/model-00079-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51a7220f355ce4ac0e06355e741582e15e50943179fb4a7c5bf67dc66ade8510 +size 1711309872 diff --git a/model-00080-of-00082.safetensors b/model-00080-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..48361b8e1ae5c135f90ca921b6c1217c1decd462 --- /dev/null +++ b/model-00080-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b1e1403b21469ab044487ade3af9a558b3fe6d80852e7e4a0ce6b0ec677d89e +size 1711309872 diff --git a/model-00081-of-00082.safetensors b/model-00081-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..efa560280cbf905b0ef17ca0d4f4a5a3baad1bc9 --- /dev/null +++ b/model-00081-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64e48077f14c09c1b2e6a876f26e06fdea07957802f6f1d93694775a5be61324 +size 2101346432 diff --git a/model-00082-of-00082.safetensors b/model-00082-of-00082.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f29471362af6a046b996f7cb0c3c2a5e1c22c44c --- /dev/null +++ b/model-00082-of-00082.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:958a4d179273036965eb524c6d96069d50e3b5fca26c9cd151a81381a68e762a +size 1409335984 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..542ed87d1a688540fc1068d20e86d28ebe616060 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,730 @@ +{ + "metadata": { + "total_size": 141107412992 + }, + "weight_map": { + "lm_head.weight": "model-00081-of-00082.safetensors", + "model.embed_tokens.weight": "model-00001-of-00082.safetensors", + "model.layers.0.input_layernorm.weight": "model-00002-of-00082.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.1.input_layernorm.weight": "model-00003-of-00082.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00082.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00082.safetensors", + "model.layers.10.input_layernorm.weight": "model-00012-of-00082.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00012-of-00082.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.11.input_layernorm.weight": "model-00013-of-00082.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00013-of-00082.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00012-of-00082.safetensors", + "model.layers.12.input_layernorm.weight": "model-00014-of-00082.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00014-of-00082.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00013-of-00082.safetensors", + "model.layers.13.input_layernorm.weight": "model-00015-of-00082.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00015-of-00082.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00014-of-00082.safetensors", + "model.layers.14.input_layernorm.weight": "model-00016-of-00082.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00016-of-00082.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00015-of-00082.safetensors", + "model.layers.15.input_layernorm.weight": "model-00017-of-00082.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00017-of-00082.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00016-of-00082.safetensors", + "model.layers.16.input_layernorm.weight": "model-00018-of-00082.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00018-of-00082.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00017-of-00082.safetensors", + "model.layers.17.input_layernorm.weight": "model-00019-of-00082.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00019-of-00082.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00018-of-00082.safetensors", + "model.layers.18.input_layernorm.weight": "model-00020-of-00082.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00020-of-00082.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00019-of-00082.safetensors", + "model.layers.19.input_layernorm.weight": "model-00021-of-00082.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00021-of-00082.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00020-of-00082.safetensors", + "model.layers.2.input_layernorm.weight": "model-00004-of-00082.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00004-of-00082.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00082.safetensors", + "model.layers.20.input_layernorm.weight": "model-00022-of-00082.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00022-of-00082.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00021-of-00082.safetensors", + "model.layers.21.input_layernorm.weight": "model-00023-of-00082.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00023-of-00082.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00022-of-00082.safetensors", + "model.layers.22.input_layernorm.weight": "model-00024-of-00082.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00024-of-00082.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00023-of-00082.safetensors", + "model.layers.23.input_layernorm.weight": "model-00025-of-00082.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00025-of-00082.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00024-of-00082.safetensors", + "model.layers.24.input_layernorm.weight": "model-00026-of-00082.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00026-of-00082.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00025-of-00082.safetensors", + "model.layers.25.input_layernorm.weight": "model-00027-of-00082.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00027-of-00082.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00026-of-00082.safetensors", + "model.layers.26.input_layernorm.weight": "model-00028-of-00082.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00028-of-00082.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00027-of-00082.safetensors", + "model.layers.27.input_layernorm.weight": "model-00029-of-00082.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00029-of-00082.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00028-of-00082.safetensors", + "model.layers.28.input_layernorm.weight": "model-00030-of-00082.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00030-of-00082.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00029-of-00082.safetensors", + "model.layers.29.input_layernorm.weight": "model-00031-of-00082.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00031-of-00082.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00030-of-00082.safetensors", + "model.layers.3.input_layernorm.weight": "model-00005-of-00082.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00005-of-00082.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00082.safetensors", + "model.layers.30.input_layernorm.weight": "model-00032-of-00082.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00032-of-00082.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00031-of-00082.safetensors", + "model.layers.31.input_layernorm.weight": "model-00033-of-00082.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00033-of-00082.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00032-of-00082.safetensors", + "model.layers.32.input_layernorm.weight": "model-00034-of-00082.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00034-of-00082.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00033-of-00082.safetensors", + "model.layers.33.input_layernorm.weight": "model-00035-of-00082.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00035-of-00082.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00034-of-00082.safetensors", + "model.layers.34.input_layernorm.weight": "model-00036-of-00082.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00036-of-00082.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00035-of-00082.safetensors", + "model.layers.35.input_layernorm.weight": "model-00037-of-00082.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00037-of-00082.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00036-of-00082.safetensors", + "model.layers.36.input_layernorm.weight": "model-00038-of-00082.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00038-of-00082.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00037-of-00082.safetensors", + "model.layers.37.input_layernorm.weight": "model-00039-of-00082.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00039-of-00082.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00038-of-00082.safetensors", + "model.layers.38.input_layernorm.weight": "model-00040-of-00082.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00040-of-00082.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00039-of-00082.safetensors", + "model.layers.39.input_layernorm.weight": "model-00041-of-00082.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00041-of-00082.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00040-of-00082.safetensors", + "model.layers.4.input_layernorm.weight": "model-00006-of-00082.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00006-of-00082.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00082.safetensors", + "model.layers.40.input_layernorm.weight": "model-00042-of-00082.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00042-of-00082.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00041-of-00082.safetensors", + "model.layers.41.input_layernorm.weight": "model-00043-of-00082.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00043-of-00082.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00042-of-00082.safetensors", + "model.layers.42.input_layernorm.weight": "model-00044-of-00082.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00044-of-00082.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00043-of-00082.safetensors", + "model.layers.43.input_layernorm.weight": "model-00045-of-00082.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00045-of-00082.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00044-of-00082.safetensors", + "model.layers.44.input_layernorm.weight": "model-00046-of-00082.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00046-of-00082.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00045-of-00082.safetensors", + "model.layers.45.input_layernorm.weight": "model-00047-of-00082.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00047-of-00082.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00046-of-00082.safetensors", + "model.layers.46.input_layernorm.weight": "model-00048-of-00082.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00048-of-00082.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00047-of-00082.safetensors", + "model.layers.47.input_layernorm.weight": "model-00049-of-00082.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00049-of-00082.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00048-of-00082.safetensors", + "model.layers.48.input_layernorm.weight": "model-00050-of-00082.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00050-of-00082.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00049-of-00082.safetensors", + "model.layers.49.input_layernorm.weight": "model-00051-of-00082.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00051-of-00082.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00050-of-00082.safetensors", + "model.layers.5.input_layernorm.weight": "model-00007-of-00082.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00007-of-00082.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00006-of-00082.safetensors", + "model.layers.50.input_layernorm.weight": "model-00052-of-00082.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00052-of-00082.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00051-of-00082.safetensors", + "model.layers.51.input_layernorm.weight": "model-00053-of-00082.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00053-of-00082.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00052-of-00082.safetensors", + "model.layers.52.input_layernorm.weight": "model-00054-of-00082.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00054-of-00082.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00053-of-00082.safetensors", + "model.layers.53.input_layernorm.weight": "model-00055-of-00082.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00055-of-00082.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00054-of-00082.safetensors", + "model.layers.54.input_layernorm.weight": "model-00056-of-00082.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00056-of-00082.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00055-of-00082.safetensors", + "model.layers.55.input_layernorm.weight": "model-00057-of-00082.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00057-of-00082.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00056-of-00082.safetensors", + "model.layers.56.input_layernorm.weight": "model-00058-of-00082.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00058-of-00082.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00057-of-00082.safetensors", + "model.layers.57.input_layernorm.weight": "model-00059-of-00082.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00059-of-00082.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00058-of-00082.safetensors", + "model.layers.58.input_layernorm.weight": "model-00060-of-00082.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00060-of-00082.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00059-of-00082.safetensors", + "model.layers.59.input_layernorm.weight": "model-00061-of-00082.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00061-of-00082.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00060-of-00082.safetensors", + "model.layers.6.input_layernorm.weight": "model-00008-of-00082.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00008-of-00082.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00007-of-00082.safetensors", + "model.layers.60.input_layernorm.weight": "model-00062-of-00082.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00062-of-00082.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00061-of-00082.safetensors", + "model.layers.61.input_layernorm.weight": "model-00063-of-00082.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00063-of-00082.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00062-of-00082.safetensors", + "model.layers.62.input_layernorm.weight": "model-00064-of-00082.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00064-of-00082.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00063-of-00082.safetensors", + "model.layers.63.input_layernorm.weight": "model-00065-of-00082.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00065-of-00082.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00064-of-00082.safetensors", + "model.layers.64.input_layernorm.weight": "model-00066-of-00082.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00066-of-00082.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00065-of-00082.safetensors", + "model.layers.65.input_layernorm.weight": "model-00067-of-00082.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00067-of-00082.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00066-of-00082.safetensors", + "model.layers.66.input_layernorm.weight": "model-00068-of-00082.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00068-of-00082.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00067-of-00082.safetensors", + "model.layers.67.input_layernorm.weight": "model-00069-of-00082.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00069-of-00082.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00068-of-00082.safetensors", + "model.layers.68.input_layernorm.weight": "model-00070-of-00082.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00070-of-00082.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00069-of-00082.safetensors", + "model.layers.69.input_layernorm.weight": "model-00071-of-00082.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00071-of-00082.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00070-of-00082.safetensors", + "model.layers.7.input_layernorm.weight": "model-00009-of-00082.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00009-of-00082.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00008-of-00082.safetensors", + "model.layers.70.input_layernorm.weight": "model-00072-of-00082.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00072-of-00082.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00071-of-00082.safetensors", + "model.layers.71.input_layernorm.weight": "model-00073-of-00082.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00073-of-00082.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00072-of-00082.safetensors", + "model.layers.72.input_layernorm.weight": "model-00074-of-00082.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00074-of-00082.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00073-of-00082.safetensors", + "model.layers.73.input_layernorm.weight": "model-00075-of-00082.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00075-of-00082.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00074-of-00082.safetensors", + "model.layers.74.input_layernorm.weight": "model-00076-of-00082.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00076-of-00082.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00075-of-00082.safetensors", + "model.layers.75.input_layernorm.weight": "model-00077-of-00082.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00077-of-00082.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00076-of-00082.safetensors", + "model.layers.76.input_layernorm.weight": "model-00078-of-00082.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00078-of-00082.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00077-of-00082.safetensors", + "model.layers.77.input_layernorm.weight": "model-00079-of-00082.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00079-of-00082.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00078-of-00082.safetensors", + "model.layers.78.input_layernorm.weight": "model-00080-of-00082.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00080-of-00082.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00079-of-00082.safetensors", + "model.layers.79.input_layernorm.weight": "model-00082-of-00082.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00082-of-00082.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00082-of-00082.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00082-of-00082.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00082-of-00082.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00080-of-00082.safetensors", + "model.layers.8.input_layernorm.weight": "model-00010-of-00082.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00010-of-00082.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00009-of-00082.safetensors", + "model.layers.9.input_layernorm.weight": "model-00011-of-00082.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00011-of-00082.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00011-of-00082.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00010-of-00082.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00010-of-00082.safetensors", + "model.norm.weight": "model-00082-of-00082.safetensors" + } +}