Korbinian Poeppel
commited on
Commit
•
f52774c
1
Parent(s):
d5d8142
Add xLSTM-7b.
Browse files- LICENSE +102 -0
- README.md +29 -3
- config.json +41 -0
- generation_config.json +7 -0
- model-00001-of-00006.safetensors +3 -0
- model-00002-of-00006.safetensors +3 -0
- model-00003-of-00006.safetensors +3 -0
- model-00004-of-00006.safetensors +3 -0
- model-00005-of-00006.safetensors +3 -0
- model-00006-of-00006.safetensors +3 -0
- model.safetensors.index.json +490 -0
LICENSE
CHANGED
@@ -0,0 +1,102 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
NXAI COMMUNITY LICENSE AGREEMENT
|
2 |
+
|
3 |
+
Preamble 1
|
4 |
+
We are proud to present the NXAI xLSTM 7B model and software, demonstrating the strength of next-generation RNN-based large language models, delivering high-quality performance and fast inference speeds. While xLSTM 7B is freely available for open research and development, we believe that organizations significantly benefiting from our technology should contribute back. Our goal is to support research, small and medium-sized enterprises (SMEs), and open innovation, while ensuring that large enterprises who incorporate xLSTM 7B into commercial products or services fairly compensate the creators for their research and development efforts.
|
5 |
+
Linz, December 12, 2024.
|
6 |
+
|
7 |
+
Preamble 2
|
8 |
+
The NXAI COMMUNITY LICENSE AGREEMENT is based on the META LLAMA 3 COMMUNITY LICENSE AGREEMENT and contains some modifications, especially Section 2, “Additional Commercial Terms” is different.
|
9 |
+
|
10 |
+
“Agreement” means the terms and conditions for use, reproduction, distribution and modification of the NXAI Materials set forth herein.
|
11 |
+
“Documentation” means the specifications, manuals and documentation accompanying NXAI Materials distributed by NXAI at https://github.com/NX-AI/.
|
12 |
+
“Licensee” or “you” means you, or your employer or any other person or entity (if you are entering into this Agreement on such person or entity’s behalf), of the age required under applicable laws, rules or regulations to provide legal consent and that has legal authority to bind your employer or such other person or entity if you are entering in this Agreement on their behalf.
|
13 |
+
“NXAI Materials” means, collectively, NXAI’s proprietary large language models, algorithms and any Software, including machine-learning model code, trained model weights, inference-enabling code, training-enabling code, fine-tuning enabling code and all other work of NXAI in the field of neural networks, Documentation (and any portion thereof) made available under this Agreement.
|
14 |
+
“NXAI” or “we” means NXAI GmbH, Linz, Austria.
|
15 |
+
|
16 |
+
By using or distributing any portion or element of the NXAI Materials, you agree to be bound by this Agreement.
|
17 |
+
|
18 |
+
1. License Rights and Redistribution.
|
19 |
+
|
20 |
+
a. Grant of Rights. You are granted a non-exclusive, worldwide, non-transferable and royalty-free limited license under NXAI’s intellectual property embodied in the NXAI Materials to use, reproduce, distribute, copy, create derivative works of, and make modifications to the NXAI Materials.
|
21 |
+
|
22 |
+
b. Redistribution and Use.
|
23 |
+
|
24 |
+
i. If you distribute or make available the NXAI Materials (or any derivative works thereof), or a product or service that uses any of them, including another AI model, you shall (A) provide a copy of this Agreement with any such NXAI Materials; and (B) prominently display “Built with technology from NXAI” on a related website, user interface, blogpost, about page, or product documentation.
|
25 |
+
|
26 |
+
ii. If you receive NXAI Materials, or any derivative works thereof, from a Licensee as part of an integrated end user product, then Section 2 of this Agreement will not apply to you.
|
27 |
+
|
28 |
+
iii. You must retain in all copies of the NXAI Materials that you distribute the following attribution notice within a “Notice” text file distributed as a part of such copies: “This product includes materials developed at NXAI that are licensed under the NXAI Community License, Copyright © NXAI GmbH, All Rights Reserved.”
|
29 |
+
|
30 |
+
2. Additional Commercial Terms. If (a) the Licensee, on a consolidated basis (including parent, subsidiaries, and affiliates), exceeds the annual revenue of one hundred million Euros (€100,000,000) or more, and (b) the Licensee incorporates NXAI Material, in whole or in part, into a Commercial Product or Service, then the Licensee must obtain a commercial license from NXAI, which NXAI may grant to you in its sole discretion, and you are not authorized to exercise any of the rights under this Agreement unless or until NXAI otherwise expressly grants you such rights
|
31 |
+
|
32 |
+
3. Disclaimer of Warranty. UNLESS REQUIRED BY APPLICABLE LAW, THE NXAI MATERIALS AND ANY OUTPUT AND RESULTS THEREFROM ARE PROVIDED ON AN “AS IS” BASIS, WITHOUT WARRANTIES OF ANY KIND, AND NXAI DISCLAIMS ALL WARRANTIES OF ANY KIND, BOTH EXPRESS AND IMPLIED, INCLUDING, WITHOUT LIMITATION, ANY WARRANTIES OF TITLE, NON-INFRINGEMENT, MERCHANTABILITY, OR FITNESS FOR A PARTICULAR PURPOSE. YOU ARE SOLELY RESPONSIBLE FOR DETERMINING THE APPROPRIATENESS OF USING OR REDISTRIBUTING THE NXAI MATERIALS AND ASSUME ANY RISKS ASSOCIATED WITH YOUR USE OF THE NXAI MATERIALS AND ANY OUTPUT AND RESULTS.
|
33 |
+
|
34 |
+
4. Limitation of Liability. IN NO EVENT WILL NXAI OR ITS AFFILIATES BE LIABLE UNDER ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, TORT, NEGLIGENCE, PRODUCTS LIABILITY, OR OTHERWISE, ARISING OUT OF THIS AGREEMENT, FOR ANY LOST PROFITS OR ANY INDIRECT, SPECIAL, CONSEQUENTIAL, INCIDENTAL, EXEMPLARY OR PUNITIVE DAMAGES, EVEN IF NXAI OR ITS AFFILIATES HAVE BEEN ADVISED OF THE POSSIBILITY OF ANY OF THE FOREGOING.
|
35 |
+
|
36 |
+
5. Intellectual Property.
|
37 |
+
|
38 |
+
a. No trademark licenses are granted under this Agreement, and in connection with the NXAI Materials, neither NXAI nor Licensee may use any name or mark owned by or associated with the other or any of its affiliates, except as required for reasonable and customary use in describing and redistributing the NXAI Materials or as set forth in this Section 5(a). NXAI hereby grants you a license to use “NXAI” (the “Mark”) solely as required to comply with the last sentence of Section 1.b.i. All goodwill arising out of your use of the Mark will insure to the benefit of NXAI.
|
39 |
+
|
40 |
+
b. Subject to NXAI’s ownership of NXAI Materials and derivatives made by or for NXAI, with respect to any derivative works and modifications of the NXAI Materials that are made by you, as between you and NXAI, you are and will be the owner of such derivative works and modifications.
|
41 |
+
|
42 |
+
c. If you institute litigation or other proceedings against NXAI or any entity (including a cross-claim or counterclaim in a lawsuit) alleging that the NXAI Materials or models released by NXAI outputs or results, or any portion of any of the foregoing, constitutes infringement of intellectual property or other rights owned or licensable by you, then any licenses granted to you under this Agreement shall terminate as of the date such litigation or claim is filed or instituted. You will indemnify and hold harmless NXAI from and against any claim by any third party arising out of or related to your use or distribution of the NXAI Materials.
|
43 |
+
|
44 |
+
6. Term and Termination. The term of this Agreement will commence upon your acceptance of this Agreement or access to the NXAI Materials and will continue in full force and effect until terminated in accordance with the terms and conditions herein. NXAI may terminate this Agreement if you are in breach of any term or condition of this Agreement. Upon termination of this Agreement, you shall delete and cease use of the NXAI Materials. Sections 3, 4 and 7 shall survive the termination of this Agreement.
|
45 |
+
|
46 |
+
7. Governing Law and Jurisdiction. This Agreement shall be governed by and construed in accordance with the laws of the Republic of Austria, without regard to its conflict of laws principles. The courts located in Linz, Austria shall have exclusive jurisdiction over any disputes arising out of or in connection with this Agreement.
|
47 |
+
|
48 |
+
====================================================================================================
|
49 |
+
|
50 |
+
This product includes software licensed under the MIT License:
|
51 |
+
|
52 |
+
MIT License
|
53 |
+
|
54 |
+
Permission is hereby granted, free of charge, to any person obtaining
|
55 |
+
a copy of this software and associated documentation files
|
56 |
+
(the "Software"), to deal in the Software without restriction,
|
57 |
+
including without limitation the rights to use, copy, modify, merge,
|
58 |
+
publish, distribute, sublicense, and/or sell copies of the Software,
|
59 |
+
and to permit persons to whom the Software is furnished to do so,
|
60 |
+
subject to the following conditions:
|
61 |
+
|
62 |
+
The above copyright notice and this permission notice shall be
|
63 |
+
included in all copies or substantial portions of the Software.
|
64 |
+
|
65 |
+
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
66 |
+
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
67 |
+
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
|
68 |
+
IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY
|
69 |
+
CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
|
70 |
+
TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
|
71 |
+
SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
|
72 |
+
|
73 |
+
====================================================================================================
|
74 |
+
|
75 |
+
This product includes software licensed under the BSD-3-Clause License.
|
76 |
+
|
77 |
+
BSD 3-Clause License
|
78 |
+
|
79 |
+
Redistribution and use in source and binary forms, with or without
|
80 |
+
modification, are permitted provided that the following conditions are met:
|
81 |
+
|
82 |
+
* Redistributions of source code must retain the above copyright notice, this
|
83 |
+
list of conditions and the following disclaimer.
|
84 |
+
|
85 |
+
* Redistributions in binary form must reproduce the above copyright notice,
|
86 |
+
this list of conditions and the following disclaimer in the documentation
|
87 |
+
and/or other materials provided with the distribution.
|
88 |
+
|
89 |
+
* Neither the name of the copyright holder nor the names of its
|
90 |
+
contributors may be used to endorse or promote products derived from
|
91 |
+
this software without specific prior written permission.
|
92 |
+
|
93 |
+
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
94 |
+
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
95 |
+
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
96 |
+
DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE
|
97 |
+
FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
98 |
+
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
99 |
+
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
100 |
+
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
101 |
+
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
102 |
+
OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
README.md
CHANGED
@@ -1,5 +1,31 @@
|
|
1 |
---
|
2 |
-
license:
|
3 |
-
license_name: nxai-community-license
|
4 |
-
license_link: LICENSE
|
5 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
license: NXAI Community License
|
|
|
|
|
3 |
---
|
4 |
+
|
5 |
+
# xLSTM goes 7B
|
6 |
+
This xLSTM was pre-trained on the DCLM and selected high-quality data for in a total of approx. 2.3 T tokens using the `xlstm-jax` framework.
|
7 |
+
|
8 |
+
|
9 |
+
## How to use it
|
10 |
+
First, install `xlstm`, which now uses the `mlstm_kernels` package for triton kernels:
|
11 |
+
```bash
|
12 |
+
pip install xlstm
|
13 |
+
```
|
14 |
+
|
15 |
+
For now, install the transformers repositiory fork from NX-AI (until it is merged):
|
16 |
+
```bash
|
17 |
+
pip install 'transformers @ git+ssh://git@github.com/NX-AI/transformers.git@integrate_xlstm'
|
18 |
+
```
|
19 |
+
|
20 |
+
Use this model as:
|
21 |
+
```python
|
22 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer
|
23 |
+
|
24 |
+
xlstm = AutoModelForCausalLM.from_pretrained("NX-AI/xLSTM-7b", device_map="auto")
|
25 |
+
|
26 |
+
# this is a fork of EleutherAI/gpt
|
27 |
+
tokenizers = AutoTokenizer.from_pretrained("NX-AI/xLSTM-7b")
|
28 |
+
|
29 |
+
xlstm(tokenizer("Hello xLSTM, how are you doing?"))
|
30 |
+
```
|
31 |
+
|
config.json
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/nfs-gpu/xlstm/converted_model_checkpoints/dclm_mLSTMv1_7B_ctx8192_sep_finetune_2024-11-29T17:03:51_0_550000",
|
3 |
+
"add_embedding_dropout": false,
|
4 |
+
"add_forward_backend_padding": false,
|
5 |
+
"add_out_norm": true,
|
6 |
+
"add_post_blocks_norm": true,
|
7 |
+
"add_post_norm": false,
|
8 |
+
"add_qk_norm": false,
|
9 |
+
"architectures": [
|
10 |
+
"xLSTMForCausalLM"
|
11 |
+
],
|
12 |
+
"bos_token_id": 0,
|
13 |
+
"cell_norm_eps": 1e-06,
|
14 |
+
"embedding_dim": 4096,
|
15 |
+
"eos_token_id": 2,
|
16 |
+
"ffn_proj_factor": 2.667,
|
17 |
+
"ffn_round_up_to_multiple_of": 64,
|
18 |
+
"force_bos_token_insert": true,
|
19 |
+
"forward_backend_name": "chunkwise--triton_limit_chunk",
|
20 |
+
"gate_soft_cap": 15.0,
|
21 |
+
"head_dim": 512,
|
22 |
+
"igate_bias_init_range": -10.0,
|
23 |
+
"mlstm_round_up_to_multiple_of": 64,
|
24 |
+
"model_type": "xlstm",
|
25 |
+
"norm_eps": 1e-06,
|
26 |
+
"norm_reduction_force_float32": true,
|
27 |
+
"num_blocks": 32,
|
28 |
+
"num_heads": 8,
|
29 |
+
"output_logit_soft_cap": 30.0,
|
30 |
+
"pad_token_id": 1,
|
31 |
+
"qk_dim_factor": 0.5,
|
32 |
+
"return_last_states": true,
|
33 |
+
"step_backend_name": "triton_fused",
|
34 |
+
"tie_word_embeddings": false,
|
35 |
+
"torch_dtype": "float32",
|
36 |
+
"transformers_version": "4.47.0.dev0",
|
37 |
+
"use_bias": false,
|
38 |
+
"use_cache": true,
|
39 |
+
"v_dim_factor": 1.0,
|
40 |
+
"vocab_size": 50304
|
41 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 0,
|
4 |
+
"eos_token_id": 2,
|
5 |
+
"pad_token_id": 1,
|
6 |
+
"transformers_version": "4.47.0.dev0"
|
7 |
+
}
|
model-00001-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2cc9c6ee0a75ec687cf18bad71908421e69b76ea383dcf72fad8c00177bca1f5
|
3 |
+
size 4991755784
|
model-00002-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8210891859f0f60c60df05effc7b43b4d11490c5c6ef889450ba827edb5dffae
|
3 |
+
size 4974522128
|
model-00003-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2902deaf4d990869e37f3ab4dc002f5f8e8ab85142aca7af82143569ec74f272
|
3 |
+
size 4840008560
|
model-00004-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:29cfabad6d13965d248bce3786c7b3a9eed21b0b8e7ce0e7fb561720cf03b6fc
|
3 |
+
size 4840008560
|
model-00005-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0f6a0f4cf4a1c7a11cb073854048813b6f3277f1b2ef6be26f3bb9a35878a40c
|
3 |
+
size 4840008560
|
model-00006-of-00006.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ed2b95082a8a3e0a9444d7bc219ba1eb49009863768ec029768de3383d56f44a
|
3 |
+
size 2975453352
|
model.safetensors.index.json
ADDED
@@ -0,0 +1,490 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"metadata": {
|
3 |
+
"total_size": 27461699584
|
4 |
+
},
|
5 |
+
"weight_map": {
|
6 |
+
"backbone.blocks.0.ffn.proj_down.weight": "model-00001-of-00006.safetensors",
|
7 |
+
"backbone.blocks.0.ffn.proj_up.weight": "model-00001-of-00006.safetensors",
|
8 |
+
"backbone.blocks.0.ffn.proj_up_gate.weight": "model-00001-of-00006.safetensors",
|
9 |
+
"backbone.blocks.0.mlstm_layer.fgate_preact.bias": "model-00001-of-00006.safetensors",
|
10 |
+
"backbone.blocks.0.mlstm_layer.fgate_preact.weight": "model-00001-of-00006.safetensors",
|
11 |
+
"backbone.blocks.0.mlstm_layer.igate_preact.bias": "model-00001-of-00006.safetensors",
|
12 |
+
"backbone.blocks.0.mlstm_layer.igate_preact.weight": "model-00001-of-00006.safetensors",
|
13 |
+
"backbone.blocks.0.mlstm_layer.k.weight": "model-00001-of-00006.safetensors",
|
14 |
+
"backbone.blocks.0.mlstm_layer.multihead_norm.weight": "model-00001-of-00006.safetensors",
|
15 |
+
"backbone.blocks.0.mlstm_layer.ogate_preact.weight": "model-00001-of-00006.safetensors",
|
16 |
+
"backbone.blocks.0.mlstm_layer.out_proj.weight": "model-00001-of-00006.safetensors",
|
17 |
+
"backbone.blocks.0.mlstm_layer.q.weight": "model-00001-of-00006.safetensors",
|
18 |
+
"backbone.blocks.0.mlstm_layer.v.weight": "model-00001-of-00006.safetensors",
|
19 |
+
"backbone.blocks.0.norm_ffn.weight": "model-00001-of-00006.safetensors",
|
20 |
+
"backbone.blocks.0.norm_mlstm.weight": "model-00001-of-00006.safetensors",
|
21 |
+
"backbone.blocks.1.ffn.proj_down.weight": "model-00001-of-00006.safetensors",
|
22 |
+
"backbone.blocks.1.ffn.proj_up.weight": "model-00001-of-00006.safetensors",
|
23 |
+
"backbone.blocks.1.ffn.proj_up_gate.weight": "model-00001-of-00006.safetensors",
|
24 |
+
"backbone.blocks.1.mlstm_layer.fgate_preact.bias": "model-00001-of-00006.safetensors",
|
25 |
+
"backbone.blocks.1.mlstm_layer.fgate_preact.weight": "model-00001-of-00006.safetensors",
|
26 |
+
"backbone.blocks.1.mlstm_layer.igate_preact.bias": "model-00001-of-00006.safetensors",
|
27 |
+
"backbone.blocks.1.mlstm_layer.igate_preact.weight": "model-00001-of-00006.safetensors",
|
28 |
+
"backbone.blocks.1.mlstm_layer.k.weight": "model-00001-of-00006.safetensors",
|
29 |
+
"backbone.blocks.1.mlstm_layer.multihead_norm.weight": "model-00001-of-00006.safetensors",
|
30 |
+
"backbone.blocks.1.mlstm_layer.ogate_preact.weight": "model-00001-of-00006.safetensors",
|
31 |
+
"backbone.blocks.1.mlstm_layer.out_proj.weight": "model-00001-of-00006.safetensors",
|
32 |
+
"backbone.blocks.1.mlstm_layer.q.weight": "model-00001-of-00006.safetensors",
|
33 |
+
"backbone.blocks.1.mlstm_layer.v.weight": "model-00001-of-00006.safetensors",
|
34 |
+
"backbone.blocks.1.norm_ffn.weight": "model-00001-of-00006.safetensors",
|
35 |
+
"backbone.blocks.1.norm_mlstm.weight": "model-00001-of-00006.safetensors",
|
36 |
+
"backbone.blocks.10.ffn.proj_down.weight": "model-00002-of-00006.safetensors",
|
37 |
+
"backbone.blocks.10.ffn.proj_up.weight": "model-00002-of-00006.safetensors",
|
38 |
+
"backbone.blocks.10.ffn.proj_up_gate.weight": "model-00002-of-00006.safetensors",
|
39 |
+
"backbone.blocks.10.mlstm_layer.fgate_preact.bias": "model-00002-of-00006.safetensors",
|
40 |
+
"backbone.blocks.10.mlstm_layer.fgate_preact.weight": "model-00002-of-00006.safetensors",
|
41 |
+
"backbone.blocks.10.mlstm_layer.igate_preact.bias": "model-00002-of-00006.safetensors",
|
42 |
+
"backbone.blocks.10.mlstm_layer.igate_preact.weight": "model-00002-of-00006.safetensors",
|
43 |
+
"backbone.blocks.10.mlstm_layer.k.weight": "model-00002-of-00006.safetensors",
|
44 |
+
"backbone.blocks.10.mlstm_layer.multihead_norm.weight": "model-00002-of-00006.safetensors",
|
45 |
+
"backbone.blocks.10.mlstm_layer.ogate_preact.weight": "model-00002-of-00006.safetensors",
|
46 |
+
"backbone.blocks.10.mlstm_layer.out_proj.weight": "model-00002-of-00006.safetensors",
|
47 |
+
"backbone.blocks.10.mlstm_layer.q.weight": "model-00002-of-00006.safetensors",
|
48 |
+
"backbone.blocks.10.mlstm_layer.v.weight": "model-00002-of-00006.safetensors",
|
49 |
+
"backbone.blocks.10.norm_ffn.weight": "model-00002-of-00006.safetensors",
|
50 |
+
"backbone.blocks.10.norm_mlstm.weight": "model-00002-of-00006.safetensors",
|
51 |
+
"backbone.blocks.11.ffn.proj_down.weight": "model-00003-of-00006.safetensors",
|
52 |
+
"backbone.blocks.11.ffn.proj_up.weight": "model-00003-of-00006.safetensors",
|
53 |
+
"backbone.blocks.11.ffn.proj_up_gate.weight": "model-00003-of-00006.safetensors",
|
54 |
+
"backbone.blocks.11.mlstm_layer.fgate_preact.bias": "model-00002-of-00006.safetensors",
|
55 |
+
"backbone.blocks.11.mlstm_layer.fgate_preact.weight": "model-00002-of-00006.safetensors",
|
56 |
+
"backbone.blocks.11.mlstm_layer.igate_preact.bias": "model-00002-of-00006.safetensors",
|
57 |
+
"backbone.blocks.11.mlstm_layer.igate_preact.weight": "model-00002-of-00006.safetensors",
|
58 |
+
"backbone.blocks.11.mlstm_layer.k.weight": "model-00002-of-00006.safetensors",
|
59 |
+
"backbone.blocks.11.mlstm_layer.multihead_norm.weight": "model-00002-of-00006.safetensors",
|
60 |
+
"backbone.blocks.11.mlstm_layer.ogate_preact.weight": "model-00002-of-00006.safetensors",
|
61 |
+
"backbone.blocks.11.mlstm_layer.out_proj.weight": "model-00002-of-00006.safetensors",
|
62 |
+
"backbone.blocks.11.mlstm_layer.q.weight": "model-00002-of-00006.safetensors",
|
63 |
+
"backbone.blocks.11.mlstm_layer.v.weight": "model-00002-of-00006.safetensors",
|
64 |
+
"backbone.blocks.11.norm_ffn.weight": "model-00002-of-00006.safetensors",
|
65 |
+
"backbone.blocks.11.norm_mlstm.weight": "model-00002-of-00006.safetensors",
|
66 |
+
"backbone.blocks.12.ffn.proj_down.weight": "model-00003-of-00006.safetensors",
|
67 |
+
"backbone.blocks.12.ffn.proj_up.weight": "model-00003-of-00006.safetensors",
|
68 |
+
"backbone.blocks.12.ffn.proj_up_gate.weight": "model-00003-of-00006.safetensors",
|
69 |
+
"backbone.blocks.12.mlstm_layer.fgate_preact.bias": "model-00003-of-00006.safetensors",
|
70 |
+
"backbone.blocks.12.mlstm_layer.fgate_preact.weight": "model-00003-of-00006.safetensors",
|
71 |
+
"backbone.blocks.12.mlstm_layer.igate_preact.bias": "model-00003-of-00006.safetensors",
|
72 |
+
"backbone.blocks.12.mlstm_layer.igate_preact.weight": "model-00003-of-00006.safetensors",
|
73 |
+
"backbone.blocks.12.mlstm_layer.k.weight": "model-00003-of-00006.safetensors",
|
74 |
+
"backbone.blocks.12.mlstm_layer.multihead_norm.weight": "model-00003-of-00006.safetensors",
|
75 |
+
"backbone.blocks.12.mlstm_layer.ogate_preact.weight": "model-00003-of-00006.safetensors",
|
76 |
+
"backbone.blocks.12.mlstm_layer.out_proj.weight": "model-00003-of-00006.safetensors",
|
77 |
+
"backbone.blocks.12.mlstm_layer.q.weight": "model-00003-of-00006.safetensors",
|
78 |
+
"backbone.blocks.12.mlstm_layer.v.weight": "model-00003-of-00006.safetensors",
|
79 |
+
"backbone.blocks.12.norm_ffn.weight": "model-00003-of-00006.safetensors",
|
80 |
+
"backbone.blocks.12.norm_mlstm.weight": "model-00003-of-00006.safetensors",
|
81 |
+
"backbone.blocks.13.ffn.proj_down.weight": "model-00003-of-00006.safetensors",
|
82 |
+
"backbone.blocks.13.ffn.proj_up.weight": "model-00003-of-00006.safetensors",
|
83 |
+
"backbone.blocks.13.ffn.proj_up_gate.weight": "model-00003-of-00006.safetensors",
|
84 |
+
"backbone.blocks.13.mlstm_layer.fgate_preact.bias": "model-00003-of-00006.safetensors",
|
85 |
+
"backbone.blocks.13.mlstm_layer.fgate_preact.weight": "model-00003-of-00006.safetensors",
|
86 |
+
"backbone.blocks.13.mlstm_layer.igate_preact.bias": "model-00003-of-00006.safetensors",
|
87 |
+
"backbone.blocks.13.mlstm_layer.igate_preact.weight": "model-00003-of-00006.safetensors",
|
88 |
+
"backbone.blocks.13.mlstm_layer.k.weight": "model-00003-of-00006.safetensors",
|
89 |
+
"backbone.blocks.13.mlstm_layer.multihead_norm.weight": "model-00003-of-00006.safetensors",
|
90 |
+
"backbone.blocks.13.mlstm_layer.ogate_preact.weight": "model-00003-of-00006.safetensors",
|
91 |
+
"backbone.blocks.13.mlstm_layer.out_proj.weight": "model-00003-of-00006.safetensors",
|
92 |
+
"backbone.blocks.13.mlstm_layer.q.weight": "model-00003-of-00006.safetensors",
|
93 |
+
"backbone.blocks.13.mlstm_layer.v.weight": "model-00003-of-00006.safetensors",
|
94 |
+
"backbone.blocks.13.norm_ffn.weight": "model-00003-of-00006.safetensors",
|
95 |
+
"backbone.blocks.13.norm_mlstm.weight": "model-00003-of-00006.safetensors",
|
96 |
+
"backbone.blocks.14.ffn.proj_down.weight": "model-00003-of-00006.safetensors",
|
97 |
+
"backbone.blocks.14.ffn.proj_up.weight": "model-00003-of-00006.safetensors",
|
98 |
+
"backbone.blocks.14.ffn.proj_up_gate.weight": "model-00003-of-00006.safetensors",
|
99 |
+
"backbone.blocks.14.mlstm_layer.fgate_preact.bias": "model-00003-of-00006.safetensors",
|
100 |
+
"backbone.blocks.14.mlstm_layer.fgate_preact.weight": "model-00003-of-00006.safetensors",
|
101 |
+
"backbone.blocks.14.mlstm_layer.igate_preact.bias": "model-00003-of-00006.safetensors",
|
102 |
+
"backbone.blocks.14.mlstm_layer.igate_preact.weight": "model-00003-of-00006.safetensors",
|
103 |
+
"backbone.blocks.14.mlstm_layer.k.weight": "model-00003-of-00006.safetensors",
|
104 |
+
"backbone.blocks.14.mlstm_layer.multihead_norm.weight": "model-00003-of-00006.safetensors",
|
105 |
+
"backbone.blocks.14.mlstm_layer.ogate_preact.weight": "model-00003-of-00006.safetensors",
|
106 |
+
"backbone.blocks.14.mlstm_layer.out_proj.weight": "model-00003-of-00006.safetensors",
|
107 |
+
"backbone.blocks.14.mlstm_layer.q.weight": "model-00003-of-00006.safetensors",
|
108 |
+
"backbone.blocks.14.mlstm_layer.v.weight": "model-00003-of-00006.safetensors",
|
109 |
+
"backbone.blocks.14.norm_ffn.weight": "model-00003-of-00006.safetensors",
|
110 |
+
"backbone.blocks.14.norm_mlstm.weight": "model-00003-of-00006.safetensors",
|
111 |
+
"backbone.blocks.15.ffn.proj_down.weight": "model-00003-of-00006.safetensors",
|
112 |
+
"backbone.blocks.15.ffn.proj_up.weight": "model-00003-of-00006.safetensors",
|
113 |
+
"backbone.blocks.15.ffn.proj_up_gate.weight": "model-00003-of-00006.safetensors",
|
114 |
+
"backbone.blocks.15.mlstm_layer.fgate_preact.bias": "model-00003-of-00006.safetensors",
|
115 |
+
"backbone.blocks.15.mlstm_layer.fgate_preact.weight": "model-00003-of-00006.safetensors",
|
116 |
+
"backbone.blocks.15.mlstm_layer.igate_preact.bias": "model-00003-of-00006.safetensors",
|
117 |
+
"backbone.blocks.15.mlstm_layer.igate_preact.weight": "model-00003-of-00006.safetensors",
|
118 |
+
"backbone.blocks.15.mlstm_layer.k.weight": "model-00003-of-00006.safetensors",
|
119 |
+
"backbone.blocks.15.mlstm_layer.multihead_norm.weight": "model-00003-of-00006.safetensors",
|
120 |
+
"backbone.blocks.15.mlstm_layer.ogate_preact.weight": "model-00003-of-00006.safetensors",
|
121 |
+
"backbone.blocks.15.mlstm_layer.out_proj.weight": "model-00003-of-00006.safetensors",
|
122 |
+
"backbone.blocks.15.mlstm_layer.q.weight": "model-00003-of-00006.safetensors",
|
123 |
+
"backbone.blocks.15.mlstm_layer.v.weight": "model-00003-of-00006.safetensors",
|
124 |
+
"backbone.blocks.15.norm_ffn.weight": "model-00003-of-00006.safetensors",
|
125 |
+
"backbone.blocks.15.norm_mlstm.weight": "model-00003-of-00006.safetensors",
|
126 |
+
"backbone.blocks.16.ffn.proj_down.weight": "model-00003-of-00006.safetensors",
|
127 |
+
"backbone.blocks.16.ffn.proj_up.weight": "model-00003-of-00006.safetensors",
|
128 |
+
"backbone.blocks.16.ffn.proj_up_gate.weight": "model-00003-of-00006.safetensors",
|
129 |
+
"backbone.blocks.16.mlstm_layer.fgate_preact.bias": "model-00003-of-00006.safetensors",
|
130 |
+
"backbone.blocks.16.mlstm_layer.fgate_preact.weight": "model-00003-of-00006.safetensors",
|
131 |
+
"backbone.blocks.16.mlstm_layer.igate_preact.bias": "model-00003-of-00006.safetensors",
|
132 |
+
"backbone.blocks.16.mlstm_layer.igate_preact.weight": "model-00003-of-00006.safetensors",
|
133 |
+
"backbone.blocks.16.mlstm_layer.k.weight": "model-00003-of-00006.safetensors",
|
134 |
+
"backbone.blocks.16.mlstm_layer.multihead_norm.weight": "model-00003-of-00006.safetensors",
|
135 |
+
"backbone.blocks.16.mlstm_layer.ogate_preact.weight": "model-00003-of-00006.safetensors",
|
136 |
+
"backbone.blocks.16.mlstm_layer.out_proj.weight": "model-00003-of-00006.safetensors",
|
137 |
+
"backbone.blocks.16.mlstm_layer.q.weight": "model-00003-of-00006.safetensors",
|
138 |
+
"backbone.blocks.16.mlstm_layer.v.weight": "model-00003-of-00006.safetensors",
|
139 |
+
"backbone.blocks.16.norm_ffn.weight": "model-00003-of-00006.safetensors",
|
140 |
+
"backbone.blocks.16.norm_mlstm.weight": "model-00003-of-00006.safetensors",
|
141 |
+
"backbone.blocks.17.ffn.proj_down.weight": "model-00004-of-00006.safetensors",
|
142 |
+
"backbone.blocks.17.ffn.proj_up.weight": "model-00004-of-00006.safetensors",
|
143 |
+
"backbone.blocks.17.ffn.proj_up_gate.weight": "model-00004-of-00006.safetensors",
|
144 |
+
"backbone.blocks.17.mlstm_layer.fgate_preact.bias": "model-00003-of-00006.safetensors",
|
145 |
+
"backbone.blocks.17.mlstm_layer.fgate_preact.weight": "model-00003-of-00006.safetensors",
|
146 |
+
"backbone.blocks.17.mlstm_layer.igate_preact.bias": "model-00003-of-00006.safetensors",
|
147 |
+
"backbone.blocks.17.mlstm_layer.igate_preact.weight": "model-00003-of-00006.safetensors",
|
148 |
+
"backbone.blocks.17.mlstm_layer.k.weight": "model-00003-of-00006.safetensors",
|
149 |
+
"backbone.blocks.17.mlstm_layer.multihead_norm.weight": "model-00003-of-00006.safetensors",
|
150 |
+
"backbone.blocks.17.mlstm_layer.ogate_preact.weight": "model-00003-of-00006.safetensors",
|
151 |
+
"backbone.blocks.17.mlstm_layer.out_proj.weight": "model-00003-of-00006.safetensors",
|
152 |
+
"backbone.blocks.17.mlstm_layer.q.weight": "model-00003-of-00006.safetensors",
|
153 |
+
"backbone.blocks.17.mlstm_layer.v.weight": "model-00003-of-00006.safetensors",
|
154 |
+
"backbone.blocks.17.norm_ffn.weight": "model-00003-of-00006.safetensors",
|
155 |
+
"backbone.blocks.17.norm_mlstm.weight": "model-00003-of-00006.safetensors",
|
156 |
+
"backbone.blocks.18.ffn.proj_down.weight": "model-00004-of-00006.safetensors",
|
157 |
+
"backbone.blocks.18.ffn.proj_up.weight": "model-00004-of-00006.safetensors",
|
158 |
+
"backbone.blocks.18.ffn.proj_up_gate.weight": "model-00004-of-00006.safetensors",
|
159 |
+
"backbone.blocks.18.mlstm_layer.fgate_preact.bias": "model-00004-of-00006.safetensors",
|
160 |
+
"backbone.blocks.18.mlstm_layer.fgate_preact.weight": "model-00004-of-00006.safetensors",
|
161 |
+
"backbone.blocks.18.mlstm_layer.igate_preact.bias": "model-00004-of-00006.safetensors",
|
162 |
+
"backbone.blocks.18.mlstm_layer.igate_preact.weight": "model-00004-of-00006.safetensors",
|
163 |
+
"backbone.blocks.18.mlstm_layer.k.weight": "model-00004-of-00006.safetensors",
|
164 |
+
"backbone.blocks.18.mlstm_layer.multihead_norm.weight": "model-00004-of-00006.safetensors",
|
165 |
+
"backbone.blocks.18.mlstm_layer.ogate_preact.weight": "model-00004-of-00006.safetensors",
|
166 |
+
"backbone.blocks.18.mlstm_layer.out_proj.weight": "model-00004-of-00006.safetensors",
|
167 |
+
"backbone.blocks.18.mlstm_layer.q.weight": "model-00004-of-00006.safetensors",
|
168 |
+
"backbone.blocks.18.mlstm_layer.v.weight": "model-00004-of-00006.safetensors",
|
169 |
+
"backbone.blocks.18.norm_ffn.weight": "model-00004-of-00006.safetensors",
|
170 |
+
"backbone.blocks.18.norm_mlstm.weight": "model-00004-of-00006.safetensors",
|
171 |
+
"backbone.blocks.19.ffn.proj_down.weight": "model-00004-of-00006.safetensors",
|
172 |
+
"backbone.blocks.19.ffn.proj_up.weight": "model-00004-of-00006.safetensors",
|
173 |
+
"backbone.blocks.19.ffn.proj_up_gate.weight": "model-00004-of-00006.safetensors",
|
174 |
+
"backbone.blocks.19.mlstm_layer.fgate_preact.bias": "model-00004-of-00006.safetensors",
|
175 |
+
"backbone.blocks.19.mlstm_layer.fgate_preact.weight": "model-00004-of-00006.safetensors",
|
176 |
+
"backbone.blocks.19.mlstm_layer.igate_preact.bias": "model-00004-of-00006.safetensors",
|
177 |
+
"backbone.blocks.19.mlstm_layer.igate_preact.weight": "model-00004-of-00006.safetensors",
|
178 |
+
"backbone.blocks.19.mlstm_layer.k.weight": "model-00004-of-00006.safetensors",
|
179 |
+
"backbone.blocks.19.mlstm_layer.multihead_norm.weight": "model-00004-of-00006.safetensors",
|
180 |
+
"backbone.blocks.19.mlstm_layer.ogate_preact.weight": "model-00004-of-00006.safetensors",
|
181 |
+
"backbone.blocks.19.mlstm_layer.out_proj.weight": "model-00004-of-00006.safetensors",
|
182 |
+
"backbone.blocks.19.mlstm_layer.q.weight": "model-00004-of-00006.safetensors",
|
183 |
+
"backbone.blocks.19.mlstm_layer.v.weight": "model-00004-of-00006.safetensors",
|
184 |
+
"backbone.blocks.19.norm_ffn.weight": "model-00004-of-00006.safetensors",
|
185 |
+
"backbone.blocks.19.norm_mlstm.weight": "model-00004-of-00006.safetensors",
|
186 |
+
"backbone.blocks.2.ffn.proj_down.weight": "model-00001-of-00006.safetensors",
|
187 |
+
"backbone.blocks.2.ffn.proj_up.weight": "model-00001-of-00006.safetensors",
|
188 |
+
"backbone.blocks.2.ffn.proj_up_gate.weight": "model-00001-of-00006.safetensors",
|
189 |
+
"backbone.blocks.2.mlstm_layer.fgate_preact.bias": "model-00001-of-00006.safetensors",
|
190 |
+
"backbone.blocks.2.mlstm_layer.fgate_preact.weight": "model-00001-of-00006.safetensors",
|
191 |
+
"backbone.blocks.2.mlstm_layer.igate_preact.bias": "model-00001-of-00006.safetensors",
|
192 |
+
"backbone.blocks.2.mlstm_layer.igate_preact.weight": "model-00001-of-00006.safetensors",
|
193 |
+
"backbone.blocks.2.mlstm_layer.k.weight": "model-00001-of-00006.safetensors",
|
194 |
+
"backbone.blocks.2.mlstm_layer.multihead_norm.weight": "model-00001-of-00006.safetensors",
|
195 |
+
"backbone.blocks.2.mlstm_layer.ogate_preact.weight": "model-00001-of-00006.safetensors",
|
196 |
+
"backbone.blocks.2.mlstm_layer.out_proj.weight": "model-00001-of-00006.safetensors",
|
197 |
+
"backbone.blocks.2.mlstm_layer.q.weight": "model-00001-of-00006.safetensors",
|
198 |
+
"backbone.blocks.2.mlstm_layer.v.weight": "model-00001-of-00006.safetensors",
|
199 |
+
"backbone.blocks.2.norm_ffn.weight": "model-00001-of-00006.safetensors",
|
200 |
+
"backbone.blocks.2.norm_mlstm.weight": "model-00001-of-00006.safetensors",
|
201 |
+
"backbone.blocks.20.ffn.proj_down.weight": "model-00004-of-00006.safetensors",
|
202 |
+
"backbone.blocks.20.ffn.proj_up.weight": "model-00004-of-00006.safetensors",
|
203 |
+
"backbone.blocks.20.ffn.proj_up_gate.weight": "model-00004-of-00006.safetensors",
|
204 |
+
"backbone.blocks.20.mlstm_layer.fgate_preact.bias": "model-00004-of-00006.safetensors",
|
205 |
+
"backbone.blocks.20.mlstm_layer.fgate_preact.weight": "model-00004-of-00006.safetensors",
|
206 |
+
"backbone.blocks.20.mlstm_layer.igate_preact.bias": "model-00004-of-00006.safetensors",
|
207 |
+
"backbone.blocks.20.mlstm_layer.igate_preact.weight": "model-00004-of-00006.safetensors",
|
208 |
+
"backbone.blocks.20.mlstm_layer.k.weight": "model-00004-of-00006.safetensors",
|
209 |
+
"backbone.blocks.20.mlstm_layer.multihead_norm.weight": "model-00004-of-00006.safetensors",
|
210 |
+
"backbone.blocks.20.mlstm_layer.ogate_preact.weight": "model-00004-of-00006.safetensors",
|
211 |
+
"backbone.blocks.20.mlstm_layer.out_proj.weight": "model-00004-of-00006.safetensors",
|
212 |
+
"backbone.blocks.20.mlstm_layer.q.weight": "model-00004-of-00006.safetensors",
|
213 |
+
"backbone.blocks.20.mlstm_layer.v.weight": "model-00004-of-00006.safetensors",
|
214 |
+
"backbone.blocks.20.norm_ffn.weight": "model-00004-of-00006.safetensors",
|
215 |
+
"backbone.blocks.20.norm_mlstm.weight": "model-00004-of-00006.safetensors",
|
216 |
+
"backbone.blocks.21.ffn.proj_down.weight": "model-00004-of-00006.safetensors",
|
217 |
+
"backbone.blocks.21.ffn.proj_up.weight": "model-00004-of-00006.safetensors",
|
218 |
+
"backbone.blocks.21.ffn.proj_up_gate.weight": "model-00004-of-00006.safetensors",
|
219 |
+
"backbone.blocks.21.mlstm_layer.fgate_preact.bias": "model-00004-of-00006.safetensors",
|
220 |
+
"backbone.blocks.21.mlstm_layer.fgate_preact.weight": "model-00004-of-00006.safetensors",
|
221 |
+
"backbone.blocks.21.mlstm_layer.igate_preact.bias": "model-00004-of-00006.safetensors",
|
222 |
+
"backbone.blocks.21.mlstm_layer.igate_preact.weight": "model-00004-of-00006.safetensors",
|
223 |
+
"backbone.blocks.21.mlstm_layer.k.weight": "model-00004-of-00006.safetensors",
|
224 |
+
"backbone.blocks.21.mlstm_layer.multihead_norm.weight": "model-00004-of-00006.safetensors",
|
225 |
+
"backbone.blocks.21.mlstm_layer.ogate_preact.weight": "model-00004-of-00006.safetensors",
|
226 |
+
"backbone.blocks.21.mlstm_layer.out_proj.weight": "model-00004-of-00006.safetensors",
|
227 |
+
"backbone.blocks.21.mlstm_layer.q.weight": "model-00004-of-00006.safetensors",
|
228 |
+
"backbone.blocks.21.mlstm_layer.v.weight": "model-00004-of-00006.safetensors",
|
229 |
+
"backbone.blocks.21.norm_ffn.weight": "model-00004-of-00006.safetensors",
|
230 |
+
"backbone.blocks.21.norm_mlstm.weight": "model-00004-of-00006.safetensors",
|
231 |
+
"backbone.blocks.22.ffn.proj_down.weight": "model-00004-of-00006.safetensors",
|
232 |
+
"backbone.blocks.22.ffn.proj_up.weight": "model-00004-of-00006.safetensors",
|
233 |
+
"backbone.blocks.22.ffn.proj_up_gate.weight": "model-00004-of-00006.safetensors",
|
234 |
+
"backbone.blocks.22.mlstm_layer.fgate_preact.bias": "model-00004-of-00006.safetensors",
|
235 |
+
"backbone.blocks.22.mlstm_layer.fgate_preact.weight": "model-00004-of-00006.safetensors",
|
236 |
+
"backbone.blocks.22.mlstm_layer.igate_preact.bias": "model-00004-of-00006.safetensors",
|
237 |
+
"backbone.blocks.22.mlstm_layer.igate_preact.weight": "model-00004-of-00006.safetensors",
|
238 |
+
"backbone.blocks.22.mlstm_layer.k.weight": "model-00004-of-00006.safetensors",
|
239 |
+
"backbone.blocks.22.mlstm_layer.multihead_norm.weight": "model-00004-of-00006.safetensors",
|
240 |
+
"backbone.blocks.22.mlstm_layer.ogate_preact.weight": "model-00004-of-00006.safetensors",
|
241 |
+
"backbone.blocks.22.mlstm_layer.out_proj.weight": "model-00004-of-00006.safetensors",
|
242 |
+
"backbone.blocks.22.mlstm_layer.q.weight": "model-00004-of-00006.safetensors",
|
243 |
+
"backbone.blocks.22.mlstm_layer.v.weight": "model-00004-of-00006.safetensors",
|
244 |
+
"backbone.blocks.22.norm_ffn.weight": "model-00004-of-00006.safetensors",
|
245 |
+
"backbone.blocks.22.norm_mlstm.weight": "model-00004-of-00006.safetensors",
|
246 |
+
"backbone.blocks.23.ffn.proj_down.weight": "model-00005-of-00006.safetensors",
|
247 |
+
"backbone.blocks.23.ffn.proj_up.weight": "model-00005-of-00006.safetensors",
|
248 |
+
"backbone.blocks.23.ffn.proj_up_gate.weight": "model-00005-of-00006.safetensors",
|
249 |
+
"backbone.blocks.23.mlstm_layer.fgate_preact.bias": "model-00004-of-00006.safetensors",
|
250 |
+
"backbone.blocks.23.mlstm_layer.fgate_preact.weight": "model-00004-of-00006.safetensors",
|
251 |
+
"backbone.blocks.23.mlstm_layer.igate_preact.bias": "model-00004-of-00006.safetensors",
|
252 |
+
"backbone.blocks.23.mlstm_layer.igate_preact.weight": "model-00004-of-00006.safetensors",
|
253 |
+
"backbone.blocks.23.mlstm_layer.k.weight": "model-00004-of-00006.safetensors",
|
254 |
+
"backbone.blocks.23.mlstm_layer.multihead_norm.weight": "model-00004-of-00006.safetensors",
|
255 |
+
"backbone.blocks.23.mlstm_layer.ogate_preact.weight": "model-00004-of-00006.safetensors",
|
256 |
+
"backbone.blocks.23.mlstm_layer.out_proj.weight": "model-00004-of-00006.safetensors",
|
257 |
+
"backbone.blocks.23.mlstm_layer.q.weight": "model-00004-of-00006.safetensors",
|
258 |
+
"backbone.blocks.23.mlstm_layer.v.weight": "model-00004-of-00006.safetensors",
|
259 |
+
"backbone.blocks.23.norm_ffn.weight": "model-00004-of-00006.safetensors",
|
260 |
+
"backbone.blocks.23.norm_mlstm.weight": "model-00004-of-00006.safetensors",
|
261 |
+
"backbone.blocks.24.ffn.proj_down.weight": "model-00005-of-00006.safetensors",
|
262 |
+
"backbone.blocks.24.ffn.proj_up.weight": "model-00005-of-00006.safetensors",
|
263 |
+
"backbone.blocks.24.ffn.proj_up_gate.weight": "model-00005-of-00006.safetensors",
|
264 |
+
"backbone.blocks.24.mlstm_layer.fgate_preact.bias": "model-00005-of-00006.safetensors",
|
265 |
+
"backbone.blocks.24.mlstm_layer.fgate_preact.weight": "model-00005-of-00006.safetensors",
|
266 |
+
"backbone.blocks.24.mlstm_layer.igate_preact.bias": "model-00005-of-00006.safetensors",
|
267 |
+
"backbone.blocks.24.mlstm_layer.igate_preact.weight": "model-00005-of-00006.safetensors",
|
268 |
+
"backbone.blocks.24.mlstm_layer.k.weight": "model-00005-of-00006.safetensors",
|
269 |
+
"backbone.blocks.24.mlstm_layer.multihead_norm.weight": "model-00005-of-00006.safetensors",
|
270 |
+
"backbone.blocks.24.mlstm_layer.ogate_preact.weight": "model-00005-of-00006.safetensors",
|
271 |
+
"backbone.blocks.24.mlstm_layer.out_proj.weight": "model-00005-of-00006.safetensors",
|
272 |
+
"backbone.blocks.24.mlstm_layer.q.weight": "model-00005-of-00006.safetensors",
|
273 |
+
"backbone.blocks.24.mlstm_layer.v.weight": "model-00005-of-00006.safetensors",
|
274 |
+
"backbone.blocks.24.norm_ffn.weight": "model-00005-of-00006.safetensors",
|
275 |
+
"backbone.blocks.24.norm_mlstm.weight": "model-00005-of-00006.safetensors",
|
276 |
+
"backbone.blocks.25.ffn.proj_down.weight": "model-00005-of-00006.safetensors",
|
277 |
+
"backbone.blocks.25.ffn.proj_up.weight": "model-00005-of-00006.safetensors",
|
278 |
+
"backbone.blocks.25.ffn.proj_up_gate.weight": "model-00005-of-00006.safetensors",
|
279 |
+
"backbone.blocks.25.mlstm_layer.fgate_preact.bias": "model-00005-of-00006.safetensors",
|
280 |
+
"backbone.blocks.25.mlstm_layer.fgate_preact.weight": "model-00005-of-00006.safetensors",
|
281 |
+
"backbone.blocks.25.mlstm_layer.igate_preact.bias": "model-00005-of-00006.safetensors",
|
282 |
+
"backbone.blocks.25.mlstm_layer.igate_preact.weight": "model-00005-of-00006.safetensors",
|
283 |
+
"backbone.blocks.25.mlstm_layer.k.weight": "model-00005-of-00006.safetensors",
|
284 |
+
"backbone.blocks.25.mlstm_layer.multihead_norm.weight": "model-00005-of-00006.safetensors",
|
285 |
+
"backbone.blocks.25.mlstm_layer.ogate_preact.weight": "model-00005-of-00006.safetensors",
|
286 |
+
"backbone.blocks.25.mlstm_layer.out_proj.weight": "model-00005-of-00006.safetensors",
|
287 |
+
"backbone.blocks.25.mlstm_layer.q.weight": "model-00005-of-00006.safetensors",
|
288 |
+
"backbone.blocks.25.mlstm_layer.v.weight": "model-00005-of-00006.safetensors",
|
289 |
+
"backbone.blocks.25.norm_ffn.weight": "model-00005-of-00006.safetensors",
|
290 |
+
"backbone.blocks.25.norm_mlstm.weight": "model-00005-of-00006.safetensors",
|
291 |
+
"backbone.blocks.26.ffn.proj_down.weight": "model-00005-of-00006.safetensors",
|
292 |
+
"backbone.blocks.26.ffn.proj_up.weight": "model-00005-of-00006.safetensors",
|
293 |
+
"backbone.blocks.26.ffn.proj_up_gate.weight": "model-00005-of-00006.safetensors",
|
294 |
+
"backbone.blocks.26.mlstm_layer.fgate_preact.bias": "model-00005-of-00006.safetensors",
|
295 |
+
"backbone.blocks.26.mlstm_layer.fgate_preact.weight": "model-00005-of-00006.safetensors",
|
296 |
+
"backbone.blocks.26.mlstm_layer.igate_preact.bias": "model-00005-of-00006.safetensors",
|
297 |
+
"backbone.blocks.26.mlstm_layer.igate_preact.weight": "model-00005-of-00006.safetensors",
|
298 |
+
"backbone.blocks.26.mlstm_layer.k.weight": "model-00005-of-00006.safetensors",
|
299 |
+
"backbone.blocks.26.mlstm_layer.multihead_norm.weight": "model-00005-of-00006.safetensors",
|
300 |
+
"backbone.blocks.26.mlstm_layer.ogate_preact.weight": "model-00005-of-00006.safetensors",
|
301 |
+
"backbone.blocks.26.mlstm_layer.out_proj.weight": "model-00005-of-00006.safetensors",
|
302 |
+
"backbone.blocks.26.mlstm_layer.q.weight": "model-00005-of-00006.safetensors",
|
303 |
+
"backbone.blocks.26.mlstm_layer.v.weight": "model-00005-of-00006.safetensors",
|
304 |
+
"backbone.blocks.26.norm_ffn.weight": "model-00005-of-00006.safetensors",
|
305 |
+
"backbone.blocks.26.norm_mlstm.weight": "model-00005-of-00006.safetensors",
|
306 |
+
"backbone.blocks.27.ffn.proj_down.weight": "model-00005-of-00006.safetensors",
|
307 |
+
"backbone.blocks.27.ffn.proj_up.weight": "model-00005-of-00006.safetensors",
|
308 |
+
"backbone.blocks.27.ffn.proj_up_gate.weight": "model-00005-of-00006.safetensors",
|
309 |
+
"backbone.blocks.27.mlstm_layer.fgate_preact.bias": "model-00005-of-00006.safetensors",
|
310 |
+
"backbone.blocks.27.mlstm_layer.fgate_preact.weight": "model-00005-of-00006.safetensors",
|
311 |
+
"backbone.blocks.27.mlstm_layer.igate_preact.bias": "model-00005-of-00006.safetensors",
|
312 |
+
"backbone.blocks.27.mlstm_layer.igate_preact.weight": "model-00005-of-00006.safetensors",
|
313 |
+
"backbone.blocks.27.mlstm_layer.k.weight": "model-00005-of-00006.safetensors",
|
314 |
+
"backbone.blocks.27.mlstm_layer.multihead_norm.weight": "model-00005-of-00006.safetensors",
|
315 |
+
"backbone.blocks.27.mlstm_layer.ogate_preact.weight": "model-00005-of-00006.safetensors",
|
316 |
+
"backbone.blocks.27.mlstm_layer.out_proj.weight": "model-00005-of-00006.safetensors",
|
317 |
+
"backbone.blocks.27.mlstm_layer.q.weight": "model-00005-of-00006.safetensors",
|
318 |
+
"backbone.blocks.27.mlstm_layer.v.weight": "model-00005-of-00006.safetensors",
|
319 |
+
"backbone.blocks.27.norm_ffn.weight": "model-00005-of-00006.safetensors",
|
320 |
+
"backbone.blocks.27.norm_mlstm.weight": "model-00005-of-00006.safetensors",
|
321 |
+
"backbone.blocks.28.ffn.proj_down.weight": "model-00005-of-00006.safetensors",
|
322 |
+
"backbone.blocks.28.ffn.proj_up.weight": "model-00005-of-00006.safetensors",
|
323 |
+
"backbone.blocks.28.ffn.proj_up_gate.weight": "model-00005-of-00006.safetensors",
|
324 |
+
"backbone.blocks.28.mlstm_layer.fgate_preact.bias": "model-00005-of-00006.safetensors",
|
325 |
+
"backbone.blocks.28.mlstm_layer.fgate_preact.weight": "model-00005-of-00006.safetensors",
|
326 |
+
"backbone.blocks.28.mlstm_layer.igate_preact.bias": "model-00005-of-00006.safetensors",
|
327 |
+
"backbone.blocks.28.mlstm_layer.igate_preact.weight": "model-00005-of-00006.safetensors",
|
328 |
+
"backbone.blocks.28.mlstm_layer.k.weight": "model-00005-of-00006.safetensors",
|
329 |
+
"backbone.blocks.28.mlstm_layer.multihead_norm.weight": "model-00005-of-00006.safetensors",
|
330 |
+
"backbone.blocks.28.mlstm_layer.ogate_preact.weight": "model-00005-of-00006.safetensors",
|
331 |
+
"backbone.blocks.28.mlstm_layer.out_proj.weight": "model-00005-of-00006.safetensors",
|
332 |
+
"backbone.blocks.28.mlstm_layer.q.weight": "model-00005-of-00006.safetensors",
|
333 |
+
"backbone.blocks.28.mlstm_layer.v.weight": "model-00005-of-00006.safetensors",
|
334 |
+
"backbone.blocks.28.norm_ffn.weight": "model-00005-of-00006.safetensors",
|
335 |
+
"backbone.blocks.28.norm_mlstm.weight": "model-00005-of-00006.safetensors",
|
336 |
+
"backbone.blocks.29.ffn.proj_down.weight": "model-00006-of-00006.safetensors",
|
337 |
+
"backbone.blocks.29.ffn.proj_up.weight": "model-00006-of-00006.safetensors",
|
338 |
+
"backbone.blocks.29.ffn.proj_up_gate.weight": "model-00006-of-00006.safetensors",
|
339 |
+
"backbone.blocks.29.mlstm_layer.fgate_preact.bias": "model-00005-of-00006.safetensors",
|
340 |
+
"backbone.blocks.29.mlstm_layer.fgate_preact.weight": "model-00005-of-00006.safetensors",
|
341 |
+
"backbone.blocks.29.mlstm_layer.igate_preact.bias": "model-00005-of-00006.safetensors",
|
342 |
+
"backbone.blocks.29.mlstm_layer.igate_preact.weight": "model-00005-of-00006.safetensors",
|
343 |
+
"backbone.blocks.29.mlstm_layer.k.weight": "model-00005-of-00006.safetensors",
|
344 |
+
"backbone.blocks.29.mlstm_layer.multihead_norm.weight": "model-00005-of-00006.safetensors",
|
345 |
+
"backbone.blocks.29.mlstm_layer.ogate_preact.weight": "model-00005-of-00006.safetensors",
|
346 |
+
"backbone.blocks.29.mlstm_layer.out_proj.weight": "model-00005-of-00006.safetensors",
|
347 |
+
"backbone.blocks.29.mlstm_layer.q.weight": "model-00005-of-00006.safetensors",
|
348 |
+
"backbone.blocks.29.mlstm_layer.v.weight": "model-00005-of-00006.safetensors",
|
349 |
+
"backbone.blocks.29.norm_ffn.weight": "model-00005-of-00006.safetensors",
|
350 |
+
"backbone.blocks.29.norm_mlstm.weight": "model-00005-of-00006.safetensors",
|
351 |
+
"backbone.blocks.3.ffn.proj_down.weight": "model-00001-of-00006.safetensors",
|
352 |
+
"backbone.blocks.3.ffn.proj_up.weight": "model-00001-of-00006.safetensors",
|
353 |
+
"backbone.blocks.3.ffn.proj_up_gate.weight": "model-00001-of-00006.safetensors",
|
354 |
+
"backbone.blocks.3.mlstm_layer.fgate_preact.bias": "model-00001-of-00006.safetensors",
|
355 |
+
"backbone.blocks.3.mlstm_layer.fgate_preact.weight": "model-00001-of-00006.safetensors",
|
356 |
+
"backbone.blocks.3.mlstm_layer.igate_preact.bias": "model-00001-of-00006.safetensors",
|
357 |
+
"backbone.blocks.3.mlstm_layer.igate_preact.weight": "model-00001-of-00006.safetensors",
|
358 |
+
"backbone.blocks.3.mlstm_layer.k.weight": "model-00001-of-00006.safetensors",
|
359 |
+
"backbone.blocks.3.mlstm_layer.multihead_norm.weight": "model-00001-of-00006.safetensors",
|
360 |
+
"backbone.blocks.3.mlstm_layer.ogate_preact.weight": "model-00001-of-00006.safetensors",
|
361 |
+
"backbone.blocks.3.mlstm_layer.out_proj.weight": "model-00001-of-00006.safetensors",
|
362 |
+
"backbone.blocks.3.mlstm_layer.q.weight": "model-00001-of-00006.safetensors",
|
363 |
+
"backbone.blocks.3.mlstm_layer.v.weight": "model-00001-of-00006.safetensors",
|
364 |
+
"backbone.blocks.3.norm_ffn.weight": "model-00001-of-00006.safetensors",
|
365 |
+
"backbone.blocks.3.norm_mlstm.weight": "model-00001-of-00006.safetensors",
|
366 |
+
"backbone.blocks.30.ffn.proj_down.weight": "model-00006-of-00006.safetensors",
|
367 |
+
"backbone.blocks.30.ffn.proj_up.weight": "model-00006-of-00006.safetensors",
|
368 |
+
"backbone.blocks.30.ffn.proj_up_gate.weight": "model-00006-of-00006.safetensors",
|
369 |
+
"backbone.blocks.30.mlstm_layer.fgate_preact.bias": "model-00006-of-00006.safetensors",
|
370 |
+
"backbone.blocks.30.mlstm_layer.fgate_preact.weight": "model-00006-of-00006.safetensors",
|
371 |
+
"backbone.blocks.30.mlstm_layer.igate_preact.bias": "model-00006-of-00006.safetensors",
|
372 |
+
"backbone.blocks.30.mlstm_layer.igate_preact.weight": "model-00006-of-00006.safetensors",
|
373 |
+
"backbone.blocks.30.mlstm_layer.k.weight": "model-00006-of-00006.safetensors",
|
374 |
+
"backbone.blocks.30.mlstm_layer.multihead_norm.weight": "model-00006-of-00006.safetensors",
|
375 |
+
"backbone.blocks.30.mlstm_layer.ogate_preact.weight": "model-00006-of-00006.safetensors",
|
376 |
+
"backbone.blocks.30.mlstm_layer.out_proj.weight": "model-00006-of-00006.safetensors",
|
377 |
+
"backbone.blocks.30.mlstm_layer.q.weight": "model-00006-of-00006.safetensors",
|
378 |
+
"backbone.blocks.30.mlstm_layer.v.weight": "model-00006-of-00006.safetensors",
|
379 |
+
"backbone.blocks.30.norm_ffn.weight": "model-00006-of-00006.safetensors",
|
380 |
+
"backbone.blocks.30.norm_mlstm.weight": "model-00006-of-00006.safetensors",
|
381 |
+
"backbone.blocks.31.ffn.proj_down.weight": "model-00006-of-00006.safetensors",
|
382 |
+
"backbone.blocks.31.ffn.proj_up.weight": "model-00006-of-00006.safetensors",
|
383 |
+
"backbone.blocks.31.ffn.proj_up_gate.weight": "model-00006-of-00006.safetensors",
|
384 |
+
"backbone.blocks.31.mlstm_layer.fgate_preact.bias": "model-00006-of-00006.safetensors",
|
385 |
+
"backbone.blocks.31.mlstm_layer.fgate_preact.weight": "model-00006-of-00006.safetensors",
|
386 |
+
"backbone.blocks.31.mlstm_layer.igate_preact.bias": "model-00006-of-00006.safetensors",
|
387 |
+
"backbone.blocks.31.mlstm_layer.igate_preact.weight": "model-00006-of-00006.safetensors",
|
388 |
+
"backbone.blocks.31.mlstm_layer.k.weight": "model-00006-of-00006.safetensors",
|
389 |
+
"backbone.blocks.31.mlstm_layer.multihead_norm.weight": "model-00006-of-00006.safetensors",
|
390 |
+
"backbone.blocks.31.mlstm_layer.ogate_preact.weight": "model-00006-of-00006.safetensors",
|
391 |
+
"backbone.blocks.31.mlstm_layer.out_proj.weight": "model-00006-of-00006.safetensors",
|
392 |
+
"backbone.blocks.31.mlstm_layer.q.weight": "model-00006-of-00006.safetensors",
|
393 |
+
"backbone.blocks.31.mlstm_layer.v.weight": "model-00006-of-00006.safetensors",
|
394 |
+
"backbone.blocks.31.norm_ffn.weight": "model-00006-of-00006.safetensors",
|
395 |
+
"backbone.blocks.31.norm_mlstm.weight": "model-00006-of-00006.safetensors",
|
396 |
+
"backbone.blocks.4.ffn.proj_down.weight": "model-00001-of-00006.safetensors",
|
397 |
+
"backbone.blocks.4.ffn.proj_up.weight": "model-00001-of-00006.safetensors",
|
398 |
+
"backbone.blocks.4.ffn.proj_up_gate.weight": "model-00001-of-00006.safetensors",
|
399 |
+
"backbone.blocks.4.mlstm_layer.fgate_preact.bias": "model-00001-of-00006.safetensors",
|
400 |
+
"backbone.blocks.4.mlstm_layer.fgate_preact.weight": "model-00001-of-00006.safetensors",
|
401 |
+
"backbone.blocks.4.mlstm_layer.igate_preact.bias": "model-00001-of-00006.safetensors",
|
402 |
+
"backbone.blocks.4.mlstm_layer.igate_preact.weight": "model-00001-of-00006.safetensors",
|
403 |
+
"backbone.blocks.4.mlstm_layer.k.weight": "model-00001-of-00006.safetensors",
|
404 |
+
"backbone.blocks.4.mlstm_layer.multihead_norm.weight": "model-00001-of-00006.safetensors",
|
405 |
+
"backbone.blocks.4.mlstm_layer.ogate_preact.weight": "model-00001-of-00006.safetensors",
|
406 |
+
"backbone.blocks.4.mlstm_layer.out_proj.weight": "model-00001-of-00006.safetensors",
|
407 |
+
"backbone.blocks.4.mlstm_layer.q.weight": "model-00001-of-00006.safetensors",
|
408 |
+
"backbone.blocks.4.mlstm_layer.v.weight": "model-00001-of-00006.safetensors",
|
409 |
+
"backbone.blocks.4.norm_ffn.weight": "model-00001-of-00006.safetensors",
|
410 |
+
"backbone.blocks.4.norm_mlstm.weight": "model-00001-of-00006.safetensors",
|
411 |
+
"backbone.blocks.5.ffn.proj_down.weight": "model-00002-of-00006.safetensors",
|
412 |
+
"backbone.blocks.5.ffn.proj_up.weight": "model-00002-of-00006.safetensors",
|
413 |
+
"backbone.blocks.5.ffn.proj_up_gate.weight": "model-00002-of-00006.safetensors",
|
414 |
+
"backbone.blocks.5.mlstm_layer.fgate_preact.bias": "model-00002-of-00006.safetensors",
|
415 |
+
"backbone.blocks.5.mlstm_layer.fgate_preact.weight": "model-00002-of-00006.safetensors",
|
416 |
+
"backbone.blocks.5.mlstm_layer.igate_preact.bias": "model-00002-of-00006.safetensors",
|
417 |
+
"backbone.blocks.5.mlstm_layer.igate_preact.weight": "model-00002-of-00006.safetensors",
|
418 |
+
"backbone.blocks.5.mlstm_layer.k.weight": "model-00001-of-00006.safetensors",
|
419 |
+
"backbone.blocks.5.mlstm_layer.multihead_norm.weight": "model-00002-of-00006.safetensors",
|
420 |
+
"backbone.blocks.5.mlstm_layer.ogate_preact.weight": "model-00002-of-00006.safetensors",
|
421 |
+
"backbone.blocks.5.mlstm_layer.out_proj.weight": "model-00002-of-00006.safetensors",
|
422 |
+
"backbone.blocks.5.mlstm_layer.q.weight": "model-00001-of-00006.safetensors",
|
423 |
+
"backbone.blocks.5.mlstm_layer.v.weight": "model-00001-of-00006.safetensors",
|
424 |
+
"backbone.blocks.5.norm_ffn.weight": "model-00002-of-00006.safetensors",
|
425 |
+
"backbone.blocks.5.norm_mlstm.weight": "model-00001-of-00006.safetensors",
|
426 |
+
"backbone.blocks.6.ffn.proj_down.weight": "model-00002-of-00006.safetensors",
|
427 |
+
"backbone.blocks.6.ffn.proj_up.weight": "model-00002-of-00006.safetensors",
|
428 |
+
"backbone.blocks.6.ffn.proj_up_gate.weight": "model-00002-of-00006.safetensors",
|
429 |
+
"backbone.blocks.6.mlstm_layer.fgate_preact.bias": "model-00002-of-00006.safetensors",
|
430 |
+
"backbone.blocks.6.mlstm_layer.fgate_preact.weight": "model-00002-of-00006.safetensors",
|
431 |
+
"backbone.blocks.6.mlstm_layer.igate_preact.bias": "model-00002-of-00006.safetensors",
|
432 |
+
"backbone.blocks.6.mlstm_layer.igate_preact.weight": "model-00002-of-00006.safetensors",
|
433 |
+
"backbone.blocks.6.mlstm_layer.k.weight": "model-00002-of-00006.safetensors",
|
434 |
+
"backbone.blocks.6.mlstm_layer.multihead_norm.weight": "model-00002-of-00006.safetensors",
|
435 |
+
"backbone.blocks.6.mlstm_layer.ogate_preact.weight": "model-00002-of-00006.safetensors",
|
436 |
+
"backbone.blocks.6.mlstm_layer.out_proj.weight": "model-00002-of-00006.safetensors",
|
437 |
+
"backbone.blocks.6.mlstm_layer.q.weight": "model-00002-of-00006.safetensors",
|
438 |
+
"backbone.blocks.6.mlstm_layer.v.weight": "model-00002-of-00006.safetensors",
|
439 |
+
"backbone.blocks.6.norm_ffn.weight": "model-00002-of-00006.safetensors",
|
440 |
+
"backbone.blocks.6.norm_mlstm.weight": "model-00002-of-00006.safetensors",
|
441 |
+
"backbone.blocks.7.ffn.proj_down.weight": "model-00002-of-00006.safetensors",
|
442 |
+
"backbone.blocks.7.ffn.proj_up.weight": "model-00002-of-00006.safetensors",
|
443 |
+
"backbone.blocks.7.ffn.proj_up_gate.weight": "model-00002-of-00006.safetensors",
|
444 |
+
"backbone.blocks.7.mlstm_layer.fgate_preact.bias": "model-00002-of-00006.safetensors",
|
445 |
+
"backbone.blocks.7.mlstm_layer.fgate_preact.weight": "model-00002-of-00006.safetensors",
|
446 |
+
"backbone.blocks.7.mlstm_layer.igate_preact.bias": "model-00002-of-00006.safetensors",
|
447 |
+
"backbone.blocks.7.mlstm_layer.igate_preact.weight": "model-00002-of-00006.safetensors",
|
448 |
+
"backbone.blocks.7.mlstm_layer.k.weight": "model-00002-of-00006.safetensors",
|
449 |
+
"backbone.blocks.7.mlstm_layer.multihead_norm.weight": "model-00002-of-00006.safetensors",
|
450 |
+
"backbone.blocks.7.mlstm_layer.ogate_preact.weight": "model-00002-of-00006.safetensors",
|
451 |
+
"backbone.blocks.7.mlstm_layer.out_proj.weight": "model-00002-of-00006.safetensors",
|
452 |
+
"backbone.blocks.7.mlstm_layer.q.weight": "model-00002-of-00006.safetensors",
|
453 |
+
"backbone.blocks.7.mlstm_layer.v.weight": "model-00002-of-00006.safetensors",
|
454 |
+
"backbone.blocks.7.norm_ffn.weight": "model-00002-of-00006.safetensors",
|
455 |
+
"backbone.blocks.7.norm_mlstm.weight": "model-00002-of-00006.safetensors",
|
456 |
+
"backbone.blocks.8.ffn.proj_down.weight": "model-00002-of-00006.safetensors",
|
457 |
+
"backbone.blocks.8.ffn.proj_up.weight": "model-00002-of-00006.safetensors",
|
458 |
+
"backbone.blocks.8.ffn.proj_up_gate.weight": "model-00002-of-00006.safetensors",
|
459 |
+
"backbone.blocks.8.mlstm_layer.fgate_preact.bias": "model-00002-of-00006.safetensors",
|
460 |
+
"backbone.blocks.8.mlstm_layer.fgate_preact.weight": "model-00002-of-00006.safetensors",
|
461 |
+
"backbone.blocks.8.mlstm_layer.igate_preact.bias": "model-00002-of-00006.safetensors",
|
462 |
+
"backbone.blocks.8.mlstm_layer.igate_preact.weight": "model-00002-of-00006.safetensors",
|
463 |
+
"backbone.blocks.8.mlstm_layer.k.weight": "model-00002-of-00006.safetensors",
|
464 |
+
"backbone.blocks.8.mlstm_layer.multihead_norm.weight": "model-00002-of-00006.safetensors",
|
465 |
+
"backbone.blocks.8.mlstm_layer.ogate_preact.weight": "model-00002-of-00006.safetensors",
|
466 |
+
"backbone.blocks.8.mlstm_layer.out_proj.weight": "model-00002-of-00006.safetensors",
|
467 |
+
"backbone.blocks.8.mlstm_layer.q.weight": "model-00002-of-00006.safetensors",
|
468 |
+
"backbone.blocks.8.mlstm_layer.v.weight": "model-00002-of-00006.safetensors",
|
469 |
+
"backbone.blocks.8.norm_ffn.weight": "model-00002-of-00006.safetensors",
|
470 |
+
"backbone.blocks.8.norm_mlstm.weight": "model-00002-of-00006.safetensors",
|
471 |
+
"backbone.blocks.9.ffn.proj_down.weight": "model-00002-of-00006.safetensors",
|
472 |
+
"backbone.blocks.9.ffn.proj_up.weight": "model-00002-of-00006.safetensors",
|
473 |
+
"backbone.blocks.9.ffn.proj_up_gate.weight": "model-00002-of-00006.safetensors",
|
474 |
+
"backbone.blocks.9.mlstm_layer.fgate_preact.bias": "model-00002-of-00006.safetensors",
|
475 |
+
"backbone.blocks.9.mlstm_layer.fgate_preact.weight": "model-00002-of-00006.safetensors",
|
476 |
+
"backbone.blocks.9.mlstm_layer.igate_preact.bias": "model-00002-of-00006.safetensors",
|
477 |
+
"backbone.blocks.9.mlstm_layer.igate_preact.weight": "model-00002-of-00006.safetensors",
|
478 |
+
"backbone.blocks.9.mlstm_layer.k.weight": "model-00002-of-00006.safetensors",
|
479 |
+
"backbone.blocks.9.mlstm_layer.multihead_norm.weight": "model-00002-of-00006.safetensors",
|
480 |
+
"backbone.blocks.9.mlstm_layer.ogate_preact.weight": "model-00002-of-00006.safetensors",
|
481 |
+
"backbone.blocks.9.mlstm_layer.out_proj.weight": "model-00002-of-00006.safetensors",
|
482 |
+
"backbone.blocks.9.mlstm_layer.q.weight": "model-00002-of-00006.safetensors",
|
483 |
+
"backbone.blocks.9.mlstm_layer.v.weight": "model-00002-of-00006.safetensors",
|
484 |
+
"backbone.blocks.9.norm_ffn.weight": "model-00002-of-00006.safetensors",
|
485 |
+
"backbone.blocks.9.norm_mlstm.weight": "model-00002-of-00006.safetensors",
|
486 |
+
"backbone.embeddings.weight": "model-00001-of-00006.safetensors",
|
487 |
+
"backbone.out_norm.weight": "model-00006-of-00006.safetensors",
|
488 |
+
"lm_head.weight": "model-00006-of-00006.safetensors"
|
489 |
+
}
|
490 |
+
}
|