MinistryofDigitalAffairs commited on
Commit
524372d
·
1 Parent(s): 03aec59

Uploaded PLLuM model.

Browse files
Files changed (49) hide show
  1. .gitattributes +1 -0
  2. README.md +232 -0
  3. apache-2.0.md +449 -0
  4. config.json +29 -0
  5. generation_config.json +6 -0
  6. model-00001-of-00039.safetensors +3 -0
  7. model-00002-of-00039.safetensors +3 -0
  8. model-00003-of-00039.safetensors +3 -0
  9. model-00004-of-00039.safetensors +3 -0
  10. model-00005-of-00039.safetensors +3 -0
  11. model-00006-of-00039.safetensors +3 -0
  12. model-00007-of-00039.safetensors +3 -0
  13. model-00008-of-00039.safetensors +3 -0
  14. model-00009-of-00039.safetensors +3 -0
  15. model-00010-of-00039.safetensors +3 -0
  16. model-00011-of-00039.safetensors +3 -0
  17. model-00012-of-00039.safetensors +3 -0
  18. model-00013-of-00039.safetensors +3 -0
  19. model-00014-of-00039.safetensors +3 -0
  20. model-00015-of-00039.safetensors +3 -0
  21. model-00016-of-00039.safetensors +3 -0
  22. model-00017-of-00039.safetensors +3 -0
  23. model-00018-of-00039.safetensors +3 -0
  24. model-00019-of-00039.safetensors +3 -0
  25. model-00020-of-00039.safetensors +3 -0
  26. model-00021-of-00039.safetensors +3 -0
  27. model-00022-of-00039.safetensors +3 -0
  28. model-00023-of-00039.safetensors +3 -0
  29. model-00024-of-00039.safetensors +3 -0
  30. model-00025-of-00039.safetensors +3 -0
  31. model-00026-of-00039.safetensors +3 -0
  32. model-00027-of-00039.safetensors +3 -0
  33. model-00028-of-00039.safetensors +3 -0
  34. model-00029-of-00039.safetensors +3 -0
  35. model-00030-of-00039.safetensors +3 -0
  36. model-00031-of-00039.safetensors +3 -0
  37. model-00032-of-00039.safetensors +3 -0
  38. model-00033-of-00039.safetensors +3 -0
  39. model-00034-of-00039.safetensors +3 -0
  40. model-00035-of-00039.safetensors +3 -0
  41. model-00036-of-00039.safetensors +3 -0
  42. model-00037-of-00039.safetensors +3 -0
  43. model-00038-of-00039.safetensors +3 -0
  44. model-00039-of-00039.safetensors +3 -0
  45. model.safetensors.index.json +1002 -0
  46. special_tokens_map.json +5 -0
  47. tokenizer.json +3 -0
  48. tokenizer.model +3 -0
  49. tokenizer_config.json +42 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,232 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+ <p align="center">
5
+ <img src="https://pllum.org.pl/_nuxt/PLLuM_logo_RGB_color.DXNEc-VR.png">
6
+ </p>
7
+
8
+ # PLLuM: A Family of Polish Large Language Models
9
+
10
+ ## Overview
11
+ PLLuM is a family of large language models (LLMs) specialized in Polish and other Slavic/Baltic languages, with additional English data incorporated for broader generalization. Developed through an extensive collaboration with various data providers, PLLuM models are built on high-quality text corpora and refined through instruction tuning, preference learning, and advanced alignment techniques. These models are intended to generate contextually coherent text, offer assistance in various tasks (e.g., question answering, summarization), and serve as a foundation for specialized applications such as domain-specific intelligent assistants.
12
+
13
+ ### Key Highlights
14
+ - **Extensive Data Collection**
15
+ We gathered large-scale, high-quality text data in Polish (around 150B tokens after cleaning and deduplication) and additional text in Slavic, Baltic, and English languages. Part of these tokens (28B) can be used in fully open-source models, including for commercial use (in compliance with relevant legal regulations).
16
+
17
+ - **Organic Instruction Dataset**
18
+ We curated the largest Polish collection of manually created “organic instructions” (~40k prompt-response pairs, including ~3.5k multi-turn dialogs). This human-authored instruction set is based on an extensive typology of human-model interactions and it covers a range of subtle aspects of supervised fine-tuning (SFT) that might be overlooked with automated approaches (including large scale distillation of 'strong LLMs'). It was also designed to mitigate negative linguistic transfer from non-Polish textual data used in the pre-training phase.
19
+
20
+ - **Polish Preference Corpus**
21
+ We created the first Polish-language preference corpus, featuring prompts and multiple model responses manually assessed by a demographically diverse team of annotators. This dataset teaches the model not only correctness (factual and linguistic) but also balance and safety—especially for potentially controversial or adversarial topics.
22
+
23
+ - **Evaluation Benchmarks**
24
+ We developed custom benchmarks to evaluate our models on tasks relevant to Polish public administration, where PLLuM achieved top scores among all tested models. In broader Polish-language tasks, PLLuM models also attain state-of-the-art results.
25
+
26
+ ## Model Description
27
+
28
+ Below is a summary of the main PLLuM models, including their licenses, bases, and parameter sizes. All model names link to a specific Hugging Face resources, while the base models and licenses link to their respective sources or license references. Note that all *-nc-* models are intended to non-commercial use.
29
+
30
+ | Model Name | Params | License | Based On |
31
+ |-------------------------------------------------------|----------------------|---------------------------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------------------------|
32
+ | [Llama-PLLuM-8B-base](https://huggingface.co/CYFRAGOVPL/Llama-PLLuM-8B-base) | 8B | [Llama 3.1](https://huggingface.co/meta-llama/Llama-3.1-8B/blob/main/LICENSE) | [Llama3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B) |
33
+ | [Llama-PLLuM-8B-instruct](https://huggingface.co/CYFRAGOVPL/Llama-PLLuM-8B-instruct) | 8B | [Llama 3.1](https://huggingface.co/meta-llama/Llama-3.1-8B/blob/main/LICENSE) | [Llama3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B) |
34
+ | [Llama-PLLuM-8B-chat](https://huggingface.co/CYFRAGOVPL/Llama-PLLuM-8B-chat) | 8B | [Llama 3.1](https://huggingface.co/meta-llama/Llama-3.1-8B/blob/main/LICENSE) | [Llama3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B) |
35
+ | [PLLuM-12B-base](https://huggingface.co/CYFRAGOVPL/PLLuM-12B-base) | 12B | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.txt) | [Mistral-Nemo-Base-2407](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407) |
36
+ | [PLLuM-12B-instruct](https://huggingface.co/CYFRAGOVPL/PLLuM-12B-instruct) | 12B | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.txt) | [Mistral-Nemo-Base-2407](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407) |
37
+ | [PLLuM-12B-chat](https://huggingface.co/CYFRAGOVPL/PLLuM-12B-chat) | 12B | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.txt) | [Mistral-Nemo-Base-2407](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407) |
38
+ | [PLLuM-12B-nc-base](https://huggingface.co/CYFRAGOVPL/PLLuM-12B-nc-base) | 12B | [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/legalcode.txt) | [Mistral-Nemo-Base-2407](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407) |
39
+ | [PLLuM-12B-nc-instruct](https://huggingface.co/CYFRAGOVPL/PLLuM-12B-nc-instruct) | 12B | [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/legalcode.txt) | [Mistral-Nemo-Base-2407](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407) |
40
+ | [PLLuM-12B-nc-chat](https://huggingface.co/CYFRAGOVPL/PLLuM-12B-nc-chat) | 12B | [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/legalcode.txt) | [Mistral-Nemo-Base-2407](https://huggingface.co/mistralai/Mistral-Nemo-Base-2407) |
41
+ | [PLLuM-8x7B-base](https://huggingface.co/CYFRAGOVPL/PLLuM-8x7B-base) | 8×7B | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.txt) | [Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) |
42
+ | [PLLuM-8x7B-instruct](https://huggingface.co/CYFRAGOVPL/PLLuM-8x7B-instruct) | 8×7B | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.txt) | [Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) |
43
+ | [PLLuM-8x7B-chat](https://huggingface.co/CYFRAGOVPL/PLLuM-8x7B-chat) | 8×7B | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0.txt) | [Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) |
44
+ | [PLLuM-8x7B-nc-base](https://huggingface.co/CYFRAGOVPL/PLLuM-8x7B-nc-base) | 8×7B | [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/legalcode.txt) | [Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) |
45
+ | [PLLuM-8x7B-nc-instruct](https://huggingface.co/CYFRAGOVPL/PLLuM-8x7B-nc-instruct) | 8×7B | [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/legalcode.txt) | [Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) |
46
+ | [PLLuM-8x7B-nc-chat](https://huggingface.co/CYFRAGOVPL/PLLuM-8x7B-nc-chat) | 8×7B | [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/legalcode.txt) | [Mixtral-8x7B-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-v0.1) |
47
+ | [Llama-PLLuM-70B-base](https://huggingface.co/CYFRAGOVPL/Llama-PLLuM-70B-base) | 70B | [Llama 3.1](https://huggingface.co/meta-llama/Llama-3.1-70B/blob/main/LICENSE) | [Llama-3.1-70B](https://huggingface.co/meta-llama/Llama-3.1-70B) |
48
+ | [Llama-PLLuM-70B-instruct](https://huggingface.co/CYFRAGOVPL/Llama-PLLuM-70B-instruct) | 70B | [Llama 3.1](https://huggingface.co/meta-llama/Llama-3.1-70B/blob/main/LICENSE) | [Llama-3.1-70B](https://huggingface.co/meta-llama/Llama-3.1-70B) |
49
+ | [Llama-PLLuM-70B-chat](https://huggingface.co/CYFRAGOVPL/Llama-PLLuM-70B-chat) | 70B | [Llama 3.1](https://huggingface.co/meta-llama/Llama-3.1-70B/blob/main/LICENSE) | [Llama-3.1-70B](https://huggingface.co/meta-llama/Llama-3.1-70B) |
50
+
51
+ ### Model Development
52
+ - **Pretraining**: All models were pretrained or continued-pretrained on large-scale Polish corpora (up to 150B tokens) plus a range of additional Slavic/Baltic and English texts.
53
+ - **Instruction Fine-Tuning**: We refined the models on manually curated Polish “organic instructions,” converted instructions from premium Polish corpora, and synthetic instructions generated by strong LLMs.
54
+ - **Alignment and Preference Learning**: Manually annotated preference data taught the models to produce safer, balanced, and contextually appropriate responses, even in adversarial or sensitive cases.
55
+ - **Domain-Specific Adaptations**: Specialized RAG-based (Retrieval Augmented Generation) models were developed for tasks like public administration, demonstrating strong performance in complex information retrieval and question answering.
56
+
57
+ ## Intended Use Cases
58
+ - **General Language Tasks**: Text generation, summarization, question answering, etc.
59
+ - **Domain-Specific Assistants**: Especially effective for Polish public administration and legal or bureaucratic topics where domain-aware retrieval is required.
60
+ - **Research & Development**: Building blocks for downstream AI applications in academic or industrial settings, where a strong command of the Polish language is essential.
61
+
62
+ ## How to Use
63
+ Each PLLuM model can be loaded via the Hugging Face Transformers library (or compatible frameworks). For RAG-based scenarios, pair the model with a relevant vector store or document retrieval system.
64
+
65
+ Below are some recommended steps and code snippets:
66
+
67
+ ### 1. Installation
68
+ Make sure you have the latest versions of `transformers` and `torch` (or another compatible deep learning framework) installed:
69
+ ```bash
70
+ pip install transformers accelerate torch
71
+ ```
72
+
73
+ ### 2. Loading the Model
74
+ Use the following example to load one of the PLLuM models:
75
+ ```python
76
+ from transformers import AutoTokenizer, AutoModelForCausalLM
77
+
78
+ model_name = "CYFRAGOVPL/PLLuM-12B-chat" # Replace with the PLLuM model name of your choice
79
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
80
+ model = AutoModelForCausalLM.from_pretrained(model_name)
81
+ ```
82
+
83
+ ### 3. Using bfloat16 (BF16)
84
+ If your hardware (e.g., newer GPUs) supports bfloat16, you can reduce memory usage and potentially speed up inference:
85
+
86
+ ```python
87
+ import torch
88
+ from transformers import AutoTokenizer, AutoModelForCausalLM
89
+
90
+ model_name = "CYFRAGOVPL/PLLuM-12B-chat"
91
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
92
+
93
+ # Load model in bfloat16 precision
94
+ model = AutoModelForCausalLM.from_pretrained(
95
+ model_name,
96
+ torch_dtype=torch.bfloat16,
97
+ device_map="auto" # automatically places model layers on available devices
98
+ )
99
+ ```
100
+
101
+ ### 4. Generating an Example Text
102
+ ```python
103
+
104
+ prompt = "Napisz krótki wiersz o wiośnie." # EN:"Write a short poem about spring."
105
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
106
+
107
+ outputs = model.generate(
108
+ **inputs,
109
+ max_new_tokens=50,
110
+ do_sample=True,
111
+ top_k=50,
112
+ top_p=0.9,
113
+ temperature=0.7
114
+ )
115
+
116
+ generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
117
+ print(generated_text)
118
+ ```
119
+
120
+ ### 5. Expected Output
121
+ Below is a sample (hypothetical) output for the prompt above:
122
+
123
+ ```css
124
+ Przykładowy wiersz o tematyce wiosennej:
125
+
126
+ Wiosna, wiosna, wiosna, ach to ty!
127
+ Kwiecień plecień wciąż przeplata,
128
+ trochę zimy, trochę lata.
129
+ A ja nie mogę się już doczekać,
130
+ kiedy w kalendarzu ujrzę maj.
131
+ Wtedy wszystko wkoło rozkwita,
132
+ a ptaki tak pięknie śpiewają.
133
+ Wiosno, wiosno, czekam z utęsknieniem,
134
+ zrób mi tę przyjemność i przyjdź wreszcie, proszę!
135
+ ```
136
+ Your results may vary depending on model parameters (e.g., temperature, top_k, top_p), hardware, and other settings.
137
+
138
+
139
+
140
+
141
+ ## Training Procedure
142
+ - **Datasets**: ~150B tokens from Polish and multilingual sources, with ~28B tokens available for fully open-source commercial use.
143
+ - **Hyperparameters**: Vary based on model size, typically including Adam or AdamW optimizers, a range of batch sizes, and carefully tuned learning rates.
144
+ - **Hardware & Duration**: Training using [Bem2](https://man.e-science.pl/pl/kdm/bem2) HPC (up to 300xH100 GPUs). Each model’s training time depends on parameter size and hardware configuration (~8 to ~25 days on multi-GPU cluster for 8B–70B sizes).
145
+
146
+ ## Evaluation and Benchmarks
147
+ - **Public Administration**: PLLuM models demonstrated top-tier performance in specialized tasks relevant to government services.
148
+ - **Polish Language Tasks**: Across a variety of internal benchmarks and standard corpora, PLLuM consistently outperforms other models in accuracy, coherence, and safety metrics.
149
+ - **Custom Tests**: A unique preference corpus and alignment tests ensure robust, safe, and contextually accurate responses.
150
+
151
+ ## Limitations and Bias
152
+ - **Potential Hallucinations**: Like other LLMs, PLLuM may occasionally produce factually incorrect or fabricated content.
153
+ - **Sensitivity & Bias**: While extensive preference learning has been done, biases might still emerge, especially in controversial or subjective topics.
154
+ - **Context Length**: Very long context tasks may challenge certain models, depending on memory constraints.
155
+
156
+ ## Ethical Considerations
157
+ PLLuM models are designed for constructive and responsible usage. Users should exercise caution when deploying them in production scenarios, especially for sensitive or regulated domains. Despite efforts to minimize harmful outputs, there is always a risk of generating offensive, biased, or inappropriate text. Human oversight and due diligence are advised.
158
+
159
+ ## Citation
160
+ If you use PLLuM models or any part of this repository in your research or deployment, please cite as follows (BibTeX):
161
+ ```
162
+ @unpublished{pllum2025,
163
+ title={PLLuM: A Family of Polish Large Language Models},
164
+ author={PLLuM Consortium},
165
+ year={2025}
166
+ }
167
+ ```
168
+
169
+ ## License
170
+ Different models within the PLLuM family are published under various licenses (Apache 2.0, CC-BY-NC-4.0, or Llama 3.1 license). Check each model’s entry in the table above for details.
171
+
172
+ ## Creators & Consortium
173
+
174
+ The PLLuM project is a unique collaboration between leading Polish scientific institutions and experts from various fields, working together to create a groundbreaking Polish language model. This research partnership combines diverse competencies and passions, forming a robust foundation for advancing AI in Poland.
175
+
176
+ <table style="border: none; border-collapse: collapse;">
177
+ <tr>
178
+ <td align="center" valign="middle" style="border: none;">
179
+ <a href="https://pwr.edu.pl/">
180
+ <img src="https://pllum.org.pl/_nuxt/pwr.D1_x0B58.png" alt="pwr.D1_x0B58.png" width="100">
181
+ </a>
182
+ <br><strong>Politechnika Wrocławska</strong><br><em>– Project Leader</em>
183
+ </td>
184
+ <td align="center" valign="middle" style="border: none;">
185
+ <a href="https://www.nask.pl/">
186
+ <img src="https://pllum.org.pl/_nuxt/nask.Bz8rmSzR.png" alt="nask.Bz8rmSzR.png" width="100">
187
+ </a>
188
+ <br><strong>NASK PIB</strong>
189
+ </td>
190
+ <td align="center" valign="middle" style="border: none;">
191
+ <a href="https://www.ipipan.waw.pl/">
192
+ <img src="https://clarin.biz/_nuxt/img/ipipan.294d39c.png" alt="ipipan.294d39c.png" width="100">
193
+ </a>
194
+ <br><strong>Instytut Podstaw Informatyki PAN</strong>
195
+ </td>
196
+ </tr>
197
+ <tr>
198
+ <td align="center" valign="middle" style="border: none;">
199
+ <a href="https://opi.org.pl/">
200
+ <img src="https://pllum.org.pl/_nuxt/opi.CF-COwcC.png" alt="opi.CF-COwcC.png" width="100">
201
+ </a>
202
+ <br><strong>Ośrodek Przetwarzania Informacji PIB</strong>
203
+ </td>
204
+ <td align="center" valign="middle" style="border: none;">
205
+ <a href="https://www.uni.lodz.pl/">
206
+ <img src="https://pllum.org.pl/_nuxt/ul.aTSgr_W6.png" alt="ul.aTSgr_W6.png" width="100">
207
+ </a>
208
+ <br><strong>Uniwersytet Łódzki</strong>
209
+ </td>
210
+ <td align="center" valign="middle" style="border: none;">
211
+ <a href="https://ispan.waw.pl/default/">
212
+ <img src="https://pllum.org.pl/_nuxt/is.Dqb94VRb.png" alt="is.Dqb94VRb.png" width="100">
213
+ </a>
214
+ <br><strong>Instytut Slawistyki PAN</strong>
215
+ </td>
216
+ </tr>
217
+ </table>
218
+
219
+
220
+
221
+ ## Contact and Support
222
+ For questions or contributions, please reach out via: <pllum@e-science.pl>
223
+
224
+ We welcome feedback, collaboration, and further exploration of PLLuM models!
225
+
226
+
227
+ ## Acknowledgements
228
+
229
+ Project financed by the Minister of Digital Affairs under the targeted subsidy No. 1/WI/DBiI/2023: *“Responsible development of the open large language model PLLuM (Polish Large Language Model) to support breakthrough technologies in the public and economic sector, including an open, Polish-language intelligent assistant for petitioners.”*
230
+
231
+ **Funding Amount:** 14,504,392.00 PLN
232
+ **Contract Signing Date:** 2024-01-22
apache-2.0.md ADDED
@@ -0,0 +1,449 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <!doctype html>
2
+ <html class="">
3
+ <head>
4
+ <meta charset="utf-8" />
5
+ <meta name="viewport" content="width=device-width, initial-scale=1.0, user-scalable=no" />
6
+ <meta name="description" content="We’re on a journey to advance and democratize artificial intelligence through open source and open science." />
7
+ <meta property="fb:app_id" content="1321688464574422" />
8
+ <meta name="twitter:card" content="summary_large_image" />
9
+ <meta name="twitter:site" content="@huggingface" />
10
+ <meta name="twitter:image" content="https://cdn-thumbnails.huggingface.co/social-thumbnails/datasets/choosealicense/licenses.png" />
11
+ <meta property="og:title" content="markdown/apache-2.0.md · choosealicense/licenses at main" />
12
+ <meta property="og:type" content="website" />
13
+ <meta property="og:url" content="https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md" />
14
+ <meta property="og:image" content="https://cdn-thumbnails.huggingface.co/social-thumbnails/datasets/choosealicense/licenses.png" />
15
+
16
+ <link rel="stylesheet" href="/front/build/kube-f68c51b/style.css" />
17
+
18
+ <link rel="preconnect" href="https://fonts.gstatic.com" />
19
+ <link
20
+ href="https://fonts.googleapis.com/css2?family=Source+Sans+Pro:ital,wght@0,200;0,300;0,400;0,600;0,700;0,900;1,200;1,300;1,400;1,600;1,700;1,900&display=swap"
21
+ rel="stylesheet"
22
+ />
23
+ <link
24
+ href="https://fonts.googleapis.com/css2?family=IBM+Plex+Mono:wght@400;600;700&display=swap"
25
+ rel="stylesheet"
26
+ />
27
+
28
+ <link
29
+ rel="preload"
30
+ href="https://cdnjs.cloudflare.com/ajax/libs/KaTeX/0.12.0/katex.min.css"
31
+ as="style"
32
+ onload="this.onload=null;this.rel='stylesheet'"
33
+ />
34
+ <noscript>
35
+ <link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/KaTeX/0.12.0/katex.min.css" />
36
+ </noscript>
37
+
38
+ <script>const guestTheme = document.cookie.match(/theme=(\w+)/)?.[1]; document.documentElement.classList.toggle('dark', guestTheme === 'dark' || ( (!guestTheme || guestTheme === 'system') && window.matchMedia('(prefers-color-scheme: dark)').matches));</script>
39
+ <link rel="canonical" href="https://huggingface.co/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md">
40
+
41
+ <title>markdown/apache-2.0.md · choosealicense/licenses at main</title>
42
+
43
+ <script
44
+ defer
45
+ data-domain="huggingface.co"
46
+ event-loggedIn="false"
47
+ src="/js/script.pageview-props.js"
48
+ ></script>
49
+ <script>
50
+ window.plausible =
51
+ window.plausible ||
52
+ function () {
53
+ (window.plausible.q = window.plausible.q || []).push(arguments);
54
+ };
55
+ </script>
56
+ <script>
57
+ window.hubConfig = {"features":{"signupDisabled":false},"sshGitUrl":"git@hf.co","moonHttpUrl":"https:\/\/huggingface.co","captchaApiKey":"bd5f2066-93dc-4bdd-a64b-a24646ca3859","captchaDisabledOnSignup":true,"datasetViewerPublicUrl":"https:\/\/datasets-server.huggingface.co","stripePublicKey":"pk_live_x2tdjFXBCvXo2FFmMybezpeM00J6gPCAAc","environment":"production","userAgent":"HuggingFace (production)","spacesIframeDomain":"hf.space","spacesApiUrl":"https:\/\/api.hf.space","docSearchKey":"ece5e02e57300e17d152c08056145326e90c4bff3dd07d7d1ae40cf1c8d39cb6","logoDev":{"apiUrl":"https:\/\/img.logo.dev\/","apiKey":"pk_UHS2HZOeRnaSOdDp7jbd5w"}};
58
+ </script>
59
+ <script type="text/javascript" src="https://de5282c3ca0c.edge.sdk.awswaf.com/de5282c3ca0c/526cf06acb0d/challenge.js" defer></script>
60
+ </head>
61
+ <body class="flex flex-col min-h-dvh bg-white dark:bg-gray-950 text-black ViewerBlobPage">
62
+ <div class="flex min-h-dvh flex-col">
63
+ <div class="SVELTE_HYDRATER contents" data-target="MainHeader" data-props="{&quot;classNames&quot;:&quot;&quot;,&quot;isWide&quot;:false,&quot;isZh&quot;:false,&quot;isPro&quot;:false}"><header class="border-b border-gray-100 "><div class="w-full px-4 container flex h-16 items-center"><div class="flex flex-1 items-center"><a class="mr-5 flex flex-none items-center lg:mr-6" href="/"><img alt="Hugging Face's logo" class="w-7 md:mr-2" src="/front/assets/huggingface_logo-noborder.svg">
64
+ <span class="hidden whitespace-nowrap text-lg font-bold md:block">Hugging Face</span></a>
65
+ <div class="relative flex-1 lg:max-w-sm mr-2 sm:mr-4 md:mr-3 xl:mr-6"><input autocomplete="off" class="w-full dark:bg-gray-950 pl-8 form-input-alt h-9 pr-3 focus:shadow-xl " name="" placeholder="Search models, datasets, users..." spellcheck="false" type="text" value="">
66
+ <svg class="absolute left-2.5 text-gray-400 top-1/2 transform -translate-y-1/2" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32"><path d="M30 28.59L22.45 21A11 11 0 1 0 21 22.45L28.59 30zM5 14a9 9 0 1 1 9 9a9 9 0 0 1-9-9z" fill="currentColor"></path></svg>
67
+ </div>
68
+ <div class="flex flex-none items-center justify-center p-0.5 place-self-stretch lg:hidden"><button class="relative z-40 flex h-6 w-8 items-center justify-center" type="button"><svg width="1em" height="1em" viewBox="0 0 10 10" class="text-xl" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" preserveAspectRatio="xMidYMid meet" fill="currentColor"><path fill-rule="evenodd" clip-rule="evenodd" d="M1.65039 2.9999C1.65039 2.8066 1.80709 2.6499 2.00039 2.6499H8.00039C8.19369 2.6499 8.35039 2.8066 8.35039 2.9999C8.35039 3.1932 8.19369 3.3499 8.00039 3.3499H2.00039C1.80709 3.3499 1.65039 3.1932 1.65039 2.9999ZM1.65039 4.9999C1.65039 4.8066 1.80709 4.6499 2.00039 4.6499H8.00039C8.19369 4.6499 8.35039 4.8066 8.35039 4.9999C8.35039 5.1932 8.19369 5.3499 8.00039 5.3499H2.00039C1.80709 5.3499 1.65039 5.1932 1.65039 4.9999ZM2.00039 6.6499C1.80709 6.6499 1.65039 6.8066 1.65039 6.9999C1.65039 7.1932 1.80709 7.3499 2.00039 7.3499H8.00039C8.19369 7.3499 8.35039 7.1932 8.35039 6.9999C8.35039 6.8066 8.19369 6.6499 8.00039 6.6499H2.00039Z"></path></svg>
69
+ </button>
70
+
71
+ </div></div>
72
+ <nav aria-label="Main" class="ml-auto hidden lg:block"><ul class="flex items-center space-x-1.5 2xl:space-x-2"><li class="hover:text-indigo-700"><a class="group flex items-center px-2 py-0.5 dark:text-gray-300 dark:hover:text-gray-100" href="/models"><svg class="mr-1.5 text-gray-400 group-hover:text-indigo-500" style="" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 24 24"><path class="uim-quaternary" d="M20.23 7.24L12 12L3.77 7.24a1.98 1.98 0 0 1 .7-.71L11 2.76c.62-.35 1.38-.35 2 0l6.53 3.77c.29.173.531.418.7.71z" opacity=".25" fill="currentColor"></path><path class="uim-tertiary" d="M12 12v9.5a2.09 2.09 0 0 1-.91-.21L4.5 17.48a2.003 2.003 0 0 1-1-1.73v-7.5a2.06 2.06 0 0 1 .27-1.01L12 12z" opacity=".5" fill="currentColor"></path><path class="uim-primary" d="M20.5 8.25v7.5a2.003 2.003 0 0 1-1 1.73l-6.62 3.82c-.275.13-.576.198-.88.2V12l8.23-4.76c.175.308.268.656.27 1.01z" fill="currentColor"></path></svg>
73
+ Models</a>
74
+ </li><li class="hover:text-red-700"><a class="group flex items-center px-2 py-0.5 dark:text-gray-300 dark:hover:text-gray-100" href="/datasets"><svg class="mr-1.5 text-gray-400 group-hover:text-red-500" style="" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 25 25"><ellipse cx="12.5" cy="5" fill="currentColor" fill-opacity="0.25" rx="7.5" ry="2"></ellipse><path d="M12.5 15C16.6421 15 20 14.1046 20 13V20C20 21.1046 16.6421 22 12.5 22C8.35786 22 5 21.1046 5 20V13C5 14.1046 8.35786 15 12.5 15Z" fill="currentColor" opacity="0.5"></path><path d="M12.5 7C16.6421 7 20 6.10457 20 5V11.5C20 12.6046 16.6421 13.5 12.5 13.5C8.35786 13.5 5 12.6046 5 11.5V5C5 6.10457 8.35786 7 12.5 7Z" fill="currentColor" opacity="0.5"></path><path d="M5.23628 12C5.08204 12.1598 5 12.8273 5 13C5 14.1046 8.35786 15 12.5 15C16.6421 15 20 14.1046 20 13C20 12.8273 19.918 12.1598 19.7637 12C18.9311 12.8626 15.9947 13.5 12.5 13.5C9.0053 13.5 6.06886 12.8626 5.23628 12Z" fill="currentColor"></path></svg>
75
+ Datasets</a>
76
+ </li><li class="hover:text-blue-700"><a class="group flex items-center px-2 py-0.5 dark:text-gray-300 dark:hover:text-gray-100" href="/spaces"><svg class="mr-1.5 text-gray-400 group-hover:text-blue-500" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" viewBox="0 0 25 25"><path opacity=".5" d="M6.016 14.674v4.31h4.31v-4.31h-4.31ZM14.674 14.674v4.31h4.31v-4.31h-4.31ZM6.016 6.016v4.31h4.31v-4.31h-4.31Z" fill="currentColor"></path><path opacity=".75" fill-rule="evenodd" clip-rule="evenodd" d="M3 4.914C3 3.857 3.857 3 4.914 3h6.514c.884 0 1.628.6 1.848 1.414a5.171 5.171 0 0 1 7.31 7.31c.815.22 1.414.964 1.414 1.848v6.514A1.914 1.914 0 0 1 20.086 22H4.914A1.914 1.914 0 0 1 3 20.086V4.914Zm3.016 1.102v4.31h4.31v-4.31h-4.31Zm0 12.968v-4.31h4.31v4.31h-4.31Zm8.658 0v-4.31h4.31v4.31h-4.31Zm0-10.813a2.155 2.155 0 1 1 4.31 0 2.155 2.155 0 0 1-4.31 0Z" fill="currentColor"></path><path opacity=".25" d="M16.829 6.016a2.155 2.155 0 1 0 0 4.31 2.155 2.155 0 0 0 0-4.31Z" fill="currentColor"></path></svg>
77
+ Spaces</a>
78
+ </li><li class="hover:text-yellow-700 max-xl:hidden"><a class="group flex items-center px-2 py-0.5 dark:text-gray-300 dark:hover:text-gray-100" href="/posts"><svg class="mr-1.5 text-gray-400 group-hover:text-yellow-500 !text-yellow-500" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" viewBox="0 0 12 12" preserveAspectRatio="xMidYMid meet"><path fill="currentColor" fill-rule="evenodd" d="M3.73 2.4A4.25 4.25 0 1 1 6 10.26H2.17l-.13-.02a.43.43 0 0 1-.3-.43l.01-.06a.43.43 0 0 1 .12-.22l.84-.84A4.26 4.26 0 0 1 3.73 2.4Z" clip-rule="evenodd"></path></svg>
79
+ Posts</a>
80
+ </li><li class="hover:text-yellow-700"><a class="group flex items-center px-2 py-0.5 dark:text-gray-300 dark:hover:text-gray-100" href="/docs"><svg xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" role="img" class="mr-1.5 text-gray-400 group-hover:text-yellow-500" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32"><path opacity="0.5" d="M20.9022 5.10334L10.8012 10.8791L7.76318 9.11193C8.07741 8.56791 8.5256 8.11332 9.06512 7.7914L15.9336 3.73907C17.0868 3.08811 18.5002 3.26422 19.6534 3.91519L19.3859 3.73911C19.9253 4.06087 20.5879 4.56025 20.9022 5.10334Z" fill="currentColor"></path><path d="M10.7999 10.8792V28.5483C10.2136 28.5475 9.63494 28.4139 9.10745 28.1578C8.5429 27.8312 8.074 27.3621 7.74761 26.7975C7.42122 26.2327 7.24878 25.5923 7.24756 24.9402V10.9908C7.25062 10.3319 7.42358 9.68487 7.74973 9.1123L10.7999 10.8792Z" fill="currentColor" fill-opacity="0.75"></path><path fill-rule="evenodd" clip-rule="evenodd" d="M21.3368 10.8499V6.918C21.3331 6.25959 21.16 5.61234 20.8346 5.03949L10.7971 10.8727L10.8046 10.874L21.3368 10.8499Z" fill="currentColor"></path><path opacity="0.5" d="M21.7937 10.8488L10.7825 10.8741V28.5486L21.7937 28.5234C23.3344 28.5234 24.5835 27.2743 24.5835 25.7335V13.6387C24.5835 12.0979 23.4365 11.1233 21.7937 10.8488Z" fill="currentColor"></path></svg>
81
+ Docs</a>
82
+ </li><li class="hover:text-green-700"><a class="group flex items-center px-2 py-0.5 dark:text-gray-300 dark:hover:text-gray-100" href="/enterprise"><svg class="mr-1.5 text-gray-400 group-hover:text-green-500" xmlns="http://www.w3.org/2000/svg" fill="none" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 33 27"><path fill="currentColor" fill-rule="evenodd" d="M13.5.7a8.7 8.7 0 0 0-7.7 5.7L1 20.6c-1 3.1.9 5.7 4.1 5.7h15c3.3 0 6.8-2.6 7.8-5.7l4.6-14.2c1-3.1-.8-5.7-4-5.7h-15Zm1.1 5.7L9.8 20.3h9.8l1-3.1h-5.8l.8-2.5h4.8l1.1-3h-4.8l.8-2.3H23l1-3h-9.5Z" clip-rule="evenodd"></path></svg>
83
+ Enterprise</a>
84
+ </li>
85
+
86
+ <li><a class="group flex items-center px-2 py-0.5 dark:text-gray-300 dark:hover:text-gray-100" href="/pricing">Pricing
87
+ </a></li>
88
+
89
+ <li><div class="relative group">
90
+ <button class="px-2 py-0.5 hover:text-gray-500 dark:hover:text-gray-600 flex items-center " type="button">
91
+ <svg class=" text-gray-500 w-5 group-hover:text-gray-400 dark:text-gray-300 dark:group-hover:text-gray-100" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" viewBox="0 0 32 18" preserveAspectRatio="xMidYMid meet"><path fill-rule="evenodd" clip-rule="evenodd" d="M14.4504 3.30221C14.4504 2.836 14.8284 2.45807 15.2946 2.45807H28.4933C28.9595 2.45807 29.3374 2.836 29.3374 3.30221C29.3374 3.76842 28.9595 4.14635 28.4933 4.14635H15.2946C14.8284 4.14635 14.4504 3.76842 14.4504 3.30221Z" fill="currentColor"></path><path fill-rule="evenodd" clip-rule="evenodd" d="M14.4504 9.00002C14.4504 8.53382 14.8284 8.15588 15.2946 8.15588H28.4933C28.9595 8.15588 29.3374 8.53382 29.3374 9.00002C29.3374 9.46623 28.9595 9.84417 28.4933 9.84417H15.2946C14.8284 9.84417 14.4504 9.46623 14.4504 9.00002Z" fill="currentColor"></path><path fill-rule="evenodd" clip-rule="evenodd" d="M14.4504 14.6978C14.4504 14.2316 14.8284 13.8537 15.2946 13.8537H28.4933C28.9595 13.8537 29.3374 14.2316 29.3374 14.6978C29.3374 15.164 28.9595 15.542 28.4933 15.542H15.2946C14.8284 15.542 14.4504 15.164 14.4504 14.6978Z" fill="currentColor"></path><path fill-rule="evenodd" clip-rule="evenodd" d="M1.94549 6.87377C2.27514 6.54411 2.80962 6.54411 3.13928 6.87377L6.23458 9.96907L9.32988 6.87377C9.65954 6.54411 10.194 6.54411 10.5237 6.87377C10.8533 7.20343 10.8533 7.73791 10.5237 8.06756L6.23458 12.3567L1.94549 8.06756C1.61583 7.73791 1.61583 7.20343 1.94549 6.87377Z" fill="currentColor"></path></svg>
92
+
93
+ </button>
94
+
95
+
96
+ </div></li>
97
+ <li><hr class="h-5 w-0.5 border-none bg-gray-100 dark:bg-gray-800"></li>
98
+ <li><a class="block cursor-pointer whitespace-nowrap px-2 py-0.5 hover:text-gray-500 dark:text-gray-300 dark:hover:text-gray-100" href="/login">Log In
99
+ </a></li>
100
+ <li><a class="whitespace-nowrap rounded-full border border-transparent bg-gray-900 px-3 py-1 leading-none text-white hover:border-black hover:bg-white hover:text-black" href="/join">Sign Up
101
+ </a></li></ul></nav></div></header></div>
102
+
103
+
104
+
105
+ <div class="SVELTE_HYDRATER contents" data-target="SSOBanner" data-props="{}"></div>
106
+
107
+
108
+
109
+ <main class="flex flex-1 flex-col"><div class="SVELTE_HYDRATER contents" data-target="DatasetHeader" data-props="{&quot;activeTab&quot;:&quot;files&quot;,&quot;author&quot;:{&quot;avatarUrl&quot;:&quot;https://cdn-avatars.huggingface.co/v1/production/uploads/5dd96eb166059660ed1ee413/vcvpdt_HHRaFiR1K-n09V.png&quot;,&quot;fullname&quot;:&quot;choosealicense.com mirror&quot;,&quot;name&quot;:&quot;choosealicense&quot;,&quot;type&quot;:&quot;org&quot;,&quot;isHf&quot;:false,&quot;isMod&quot;:false,&quot;isEnterprise&quot;:false,&quot;followerCount&quot;:12},&quot;canReadRepoSettings&quot;:false,&quot;dataset&quot;:{&quot;author&quot;:&quot;choosealicense&quot;,&quot;cardData&quot;:{&quot;license&quot;:&quot;mit&quot;,&quot;viewer&quot;:false},&quot;cardExists&quot;:true,&quot;createdAt&quot;:&quot;2024-04-17T10:10:07.000Z&quot;,&quot;description&quot;:&quot;\n\t\n\t\t\n\t\n\t\n\t\tCommon license info\n\t\n\n\n\t\n\t\t\n\t\n\t\n\t\tExtracted from https://github.com/github/choosealicense.com\n\t\n\n\n\t\n\t\t\nlicense id\n\n\n\t\t\n0bsd\n\n\nafl-3.0\n\n\nagpl-3.0\n\n\napache-2.0\n\n\nartistic-2.0\n\n\nblueoak-1.0.0\n\n\nbsd-2-clause-patent\n\n\nbsd-2-clause\n\n\nbsd-3-clause-clear\n\n\nbsd-3-clause\n\n\nbsd-4-clause\n\n\nbsl-1.0\n\n\ncc-by-4.0\n\n\ncc-by-sa-4.0\n\n\ncc0-1.0\n\n\ncecill-2.1\n\n\ncern-ohl-p-2.0\n\n\ncern-ohl-s-2.0\n\n\ncern-ohl-w-2.0\n\n\necl-2.0\n\n\nepl-1.0\n\n\nepl-2.0\n\n\neupl-1.1\n\n\neupl-1.2\n\n\ngfdl-1.3\n\n\ngpl-2.0\n\n\ngpl-3.0\n\n\nisc… See the full description on the dataset page: https://huggingface.co/datasets/choosealicense/licenses.&quot;,&quot;downloads&quot;:787,&quot;downloadsAllTime&quot;:2017,&quot;id&quot;:&quot;choosealicense/licenses&quot;,&quot;isLikedByUser&quot;:false,&quot;lastModified&quot;:&quot;2024-04-17T10:17:35.000Z&quot;,&quot;likes&quot;:10,&quot;discussionsDisabled&quot;:false,&quot;repoType&quot;:&quot;dataset&quot;,&quot;private&quot;:false,&quot;gated&quot;:false,&quot;tags&quot;:[&quot;license:mit&quot;,&quot;region:us&quot;],&quot;tag_objs&quot;:[{&quot;id&quot;:&quot;license:mit&quot;,&quot;label&quot;:&quot;mit&quot;,&quot;type&quot;:&quot;license&quot;},{&quot;type&quot;:&quot;region&quot;,&quot;label&quot;:&quot;🇺🇸 Region: US&quot;,&quot;id&quot;:&quot;region:us&quot;}],&quot;hasBlockedOids&quot;:false},&quot;discussionsStats&quot;:{&quot;closed&quot;:13,&quot;open&quot;:17,&quot;total&quot;:30}}"><header class="from-gray-50-to-white border-b border-gray-100 bg-gradient-to-t via-white dark:via-gray-950 pt-6 sm:pt-9"><div class="container relative "><h1 class="flex flex-wrap items-center leading-tight mb-3 text-lg max-sm:gap-y-1.5 md:text-xl"><a href="/datasets" class="group flex items-center"><svg class="mr-1.5 text-gray-400" style="" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 25 25"><ellipse cx="12.5" cy="5" fill="currentColor" fill-opacity="0.25" rx="7.5" ry="2"></ellipse><path d="M12.5 15C16.6421 15 20 14.1046 20 13V20C20 21.1046 16.6421 22 12.5 22C8.35786 22 5 21.1046 5 20V13C5 14.1046 8.35786 15 12.5 15Z" fill="currentColor" opacity="0.5"></path><path d="M12.5 7C16.6421 7 20 6.10457 20 5V11.5C20 12.6046 16.6421 13.5 12.5 13.5C8.35786 13.5 5 12.6046 5 11.5V5C5 6.10457 8.35786 7 12.5 7Z" fill="currentColor" opacity="0.5"></path><path d="M5.23628 12C5.08204 12.1598 5 12.8273 5 13C5 14.1046 8.35786 15 12.5 15C16.6421 15 20 14.1046 20 13C20 12.8273 19.918 12.1598 19.7637 12C18.9311 12.8626 15.9947 13.5 12.5 13.5C9.0053 13.5 6.06886 12.8626 5.23628 12Z" fill="currentColor"></path></svg>
110
+ <span class="mr-2.5 font-semibold text-gray-400 group-hover:text-gray-500">Datasets:</span></a>
111
+ <div class="group flex flex-none items-center"><div class="relative mr-1 flex items-center">
112
+
113
+
114
+
115
+ <span class="inline-block "><span class="contents"><a href="/choosealicense" class="text-gray-400 hover:text-blue-600"><img alt="" class="w-3.5 h-3.5 rounded flex-none" src="https://cdn-avatars.huggingface.co/v1/production/uploads/5dd96eb166059660ed1ee413/vcvpdt_HHRaFiR1K-n09V.png" crossorigin="anonymous"></a></span>
116
+ </span></div>
117
+
118
+
119
+ <span class="inline-block "><span class="contents"><a href="/choosealicense" class="text-gray-400 hover:text-blue-600">choosealicense</a></span>
120
+ </span>
121
+ <div class="mx-0.5 text-gray-300">/</div></div>
122
+
123
+ <div class="max-w-full "><a class="break-words font-mono font-semibold hover:text-blue-600 " href="/datasets/choosealicense/licenses">licenses</a>
124
+ <button class="relative text-sm mr-4 inline-flex cursor-pointer items-center text-sm focus:outline-none mx-0.5 text-gray-600 " title="Copy dataset name to clipboard" type="button"><svg class="" xmlns="http://www.w3.org/2000/svg" aria-hidden="true" fill="currentColor" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32"><path d="M28,10V28H10V10H28m0-2H10a2,2,0,0,0-2,2V28a2,2,0,0,0,2,2H28a2,2,0,0,0,2-2V10a2,2,0,0,0-2-2Z" transform="translate(0)"></path><path d="M4,18H2V4A2,2,0,0,1,4,2H18V4H4Z" transform="translate(0)"></path><rect fill="none" width="32" height="32"></rect></svg>
125
+
126
+ </button></div>
127
+ <div class="inline-flex items-center overflow-hidden whitespace-nowrap rounded-md border bg-white text-sm leading-none text-gray-500 mr-2"><button class="relative flex items-center overflow-hidden from-red-50 to-transparent dark:from-red-900 px-1.5 py-1 hover:bg-gradient-to-t focus:outline-none" title="Like"><svg class="left-1.5 absolute" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32" fill="currentColor"><path d="M22.45,6a5.47,5.47,0,0,1,3.91,1.64,5.7,5.7,0,0,1,0,8L16,26.13,5.64,15.64a5.7,5.7,0,0,1,0-8,5.48,5.48,0,0,1,7.82,0L16,10.24l2.53-2.58A5.44,5.44,0,0,1,22.45,6m0-2a7.47,7.47,0,0,0-5.34,2.24L16,7.36,14.89,6.24a7.49,7.49,0,0,0-10.68,0,7.72,7.72,0,0,0,0,10.82L16,29,27.79,17.06a7.72,7.72,0,0,0,0-10.82A7.49,7.49,0,0,0,22.45,4Z"></path></svg>
128
+
129
+
130
+ <span class="ml-4 pl-0.5 ">like</span></button>
131
+ <button class="flex items-center border-l px-1.5 py-1 text-gray-400 hover:bg-gray-50 focus:bg-gray-100 focus:outline-none dark:hover:bg-gray-900 dark:focus:bg-gray-800" title="See users who liked this repository">10</button></div>
132
+
133
+
134
+
135
+
136
+ <div class="relative flex items-center gap-1.5 "><div class="mr-2 inline-flex h-6 items-center overflow-hidden whitespace-nowrap rounded-md border text-sm text-gray-500"><button class="relative flex h-full max-w-56 items-center gap-1.5 overflow-hidden px-1.5 hover:bg-gray-50 focus:bg-gray-100 focus:outline-none dark:hover:bg-gray-900 dark:focus:bg-gray-800" type="button" ><div class="flex h-full flex-1 items-center justify-center ">Follow</div>
137
+ <img alt="" class="size-3 flex-none rounded-sm" src="https://cdn-avatars.huggingface.co/v1/production/uploads/5dd96eb166059660ed1ee413/vcvpdt_HHRaFiR1K-n09V.png">
138
+ <span class="truncate">choosealicense.com mirror</span></button>
139
+ <button class="flex h-full items-center border-l pl-1.5 pr-1.5 text-gray-400 hover:bg-gray-50 focus:bg-gray-100 focus:outline-none dark:hover:bg-gray-900 dark:focus:bg-gray-800" title="Show choosealicense.com mirror's followers" type="button">12</button></div>
140
+
141
+ </div>
142
+
143
+ </h1>
144
+ <div class="mb-3 flex flex-wrap md:mb-4"><div class="mr-1 flex flex-wrap items-center"><span class="mb-1 mr-1 p-1 text-sm leading-tight text-gray-400 md:mb-1.5">License:
145
+ </span>
146
+ <div class="relative inline-block ">
147
+ <button class="group mr-1 mb-1 md:mr-1.5 md:mb-1.5 rounded-full rounded-br-none " type="button">
148
+ <div class="tag tag-white rounded-full relative rounded-br-none pr-2.5">
149
+ <svg class="text-xs text-gray-900" width="1em" height="1em" viewBox="0 0 10 10" fill="none" xmlns="http://www.w3.org/2000/svg"><path d="M1.46009 5.0945V6.88125C1.46009 7.25201 1.75937 7.55129 2.13012 7.55129C2.50087 7.55129 2.80016 7.25201 2.80016 6.88125V5.0945C2.80016 4.72375 2.50087 4.42446 2.13012 4.42446C1.75937 4.42446 1.46009 4.72375 1.46009 5.0945ZM4.14022 5.0945V6.88125C4.14022 7.25201 4.4395 7.55129 4.81026 7.55129C5.18101 7.55129 5.48029 7.25201 5.48029 6.88125V5.0945C5.48029 4.72375 5.18101 4.42446 4.81026 4.42446C4.4395 4.42446 4.14022 4.72375 4.14022 5.0945ZM1.23674 9.78473H8.38377C8.75452 9.78473 9.0538 9.48545 9.0538 9.1147C9.0538 8.74395 8.75452 8.44466 8.38377 8.44466H1.23674C0.865993 8.44466 0.566711 8.74395 0.566711 9.1147C0.566711 9.48545 0.865993 9.78473 1.23674 9.78473ZM6.82036 5.0945V6.88125C6.82036 7.25201 7.11964 7.55129 7.49039 7.55129C7.86114 7.55129 8.16042 7.25201 8.16042 6.88125V5.0945C8.16042 4.72375 7.86114 4.42446 7.49039 4.42446C7.11964 4.42446 6.82036 4.72375 6.82036 5.0945ZM4.39484 0.623142L0.865993 2.48137C0.682851 2.57517 0.566711 2.76725 0.566711 2.97273C0.566711 3.28094 0.816857 3.53109 1.12507 3.53109H8.49991C8.80365 3.53109 9.0538 3.28094 9.0538 2.97273C9.0538 2.76725 8.93766 2.57517 8.75452 2.48137L5.22568 0.623142C4.9666 0.484669 4.65391 0.484669 4.39484 0.623142V0.623142Z" fill="currentColor"></path></svg>
150
+
151
+
152
+
153
+ <span>mit</span>
154
+
155
+
156
+ <div class="border-br-gray-200 absolute bottom-0.5 right-0.5 h-1 w-1 border-[3px] border-l-transparent border-t-transparent border-b-gray-200 border-r-gray-200 dark:border-b-gray-700 dark:border-r-gray-700"></div></div>
157
+
158
+ </button>
159
+
160
+
161
+ </div>
162
+
163
+ </div></div>
164
+
165
+ <div class="flex flex-col-reverse lg:flex-row lg:items-center lg:justify-between"><div class="-mb-px flex h-12 items-center overflow-x-auto overflow-y-hidden "><a class="tab-alternate " href="/datasets/choosealicense/licenses"><svg class="mr-1.5 text-gray-400 flex-none" style="" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 24 24"><path class="uim-quaternary" d="M20.23 7.24L12 12L3.77 7.24a1.98 1.98 0 0 1 .7-.71L11 2.76c.62-.35 1.38-.35 2 0l6.53 3.77c.29.173.531.418.7.71z" opacity=".25" fill="currentColor"></path><path class="uim-tertiary" d="M12 12v9.5a2.09 2.09 0 0 1-.91-.21L4.5 17.48a2.003 2.003 0 0 1-1-1.73v-7.5a2.06 2.06 0 0 1 .27-1.01L12 12z" opacity=".5" fill="currentColor"></path><path class="uim-primary" d="M20.5 8.25v7.5a2.003 2.003 0 0 1-1 1.73l-6.62 3.82c-.275.13-.576.198-.88.2V12l8.23-4.76c.175.308.268.656.27 1.01z" fill="currentColor"></path></svg>
166
+ Dataset card
167
+
168
+
169
+ </a><a class="tab-alternate active" href="/datasets/choosealicense/licenses/tree/main"><svg class="mr-1.5 text-gray-400 flex-none" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 24 24"><path class="uim-tertiary" d="M21 19h-8a1 1 0 0 1 0-2h8a1 1 0 0 1 0 2zm0-4h-8a1 1 0 0 1 0-2h8a1 1 0 0 1 0 2zm0-8h-8a1 1 0 0 1 0-2h8a1 1 0 0 1 0 2zm0 4h-8a1 1 0 0 1 0-2h8a1 1 0 0 1 0 2z" opacity=".5" fill="currentColor"></path><path class="uim-primary" d="M9 19a1 1 0 0 1-1-1V6a1 1 0 0 1 2 0v12a1 1 0 0 1-1 1zm-6-4.333a1 1 0 0 1-.64-1.769L3.438 12l-1.078-.898a1 1 0 0 1 1.28-1.538l2 1.667a1 1 0 0 1 0 1.538l-2 1.667a.999.999 0 0 1-.64.231z" fill="currentColor"></path></svg>
170
+ <span class="xl:hidden">Files</span>
171
+ <span class="hidden xl:inline">Files and versions</span>
172
+
173
+
174
+ </a><a class="tab-alternate " href="/datasets/choosealicense/licenses/discussions"><svg class="mr-1.5 text-gray-400 flex-none" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32"><path d="M20.6081 3C21.7684 3 22.8053 3.49196 23.5284 4.38415C23.9756 4.93678 24.4428 5.82749 24.4808 7.16133C24.9674 7.01707 25.4353 6.93643 25.8725 6.93643C26.9833 6.93643 27.9865 7.37587 28.696 8.17411C29.6075 9.19872 30.0124 10.4579 29.8361 11.7177C29.7523 12.3177 29.5581 12.8555 29.2678 13.3534C29.8798 13.8646 30.3306 14.5763 30.5485 15.4322C30.719 16.1032 30.8939 17.5006 29.9808 18.9403C30.0389 19.0342 30.0934 19.1319 30.1442 19.2318C30.6932 20.3074 30.7283 21.5229 30.2439 22.6548C29.5093 24.3704 27.6841 25.7219 24.1397 27.1727C21.9347 28.0753 19.9174 28.6523 19.8994 28.6575C16.9842 29.4379 14.3477 29.8345 12.0653 29.8345C7.87017 29.8345 4.8668 28.508 3.13831 25.8921C0.356375 21.6797 0.754104 17.8269 4.35369 14.1131C6.34591 12.058 7.67023 9.02782 7.94613 8.36275C8.50224 6.39343 9.97271 4.20438 12.4172 4.20438H12.4179C12.6236 4.20438 12.8314 4.2214 13.0364 4.25468C14.107 4.42854 15.0428 5.06476 15.7115 6.02205C16.4331 5.09583 17.134 4.359 17.7682 3.94323C18.7242 3.31737 19.6794 3 20.6081 3ZM20.6081 5.95917C20.2427 5.95917 19.7963 6.1197 19.3039 6.44225C17.7754 7.44319 14.8258 12.6772 13.7458 14.7131C13.3839 15.3952 12.7655 15.6837 12.2086 15.6837C11.1036 15.6837 10.2408 14.5497 12.1076 13.1085C14.9146 10.9402 13.9299 7.39584 12.5898 7.1776C12.5311 7.16799 12.4731 7.16355 12.4172 7.16355C11.1989 7.16355 10.6615 9.33114 10.6615 9.33114C10.6615 9.33114 9.0863 13.4148 6.38031 16.206C3.67434 18.998 3.5346 21.2388 5.50675 24.2246C6.85185 26.2606 9.42666 26.8753 12.0653 26.8753C14.8021 26.8753 17.6077 26.2139 19.1799 25.793C19.2574 25.7723 28.8193 22.984 27.6081 20.6107C27.4046 20.212 27.0693 20.0522 26.6471 20.0522C24.9416 20.0522 21.8393 22.6726 20.5057 22.6726C20.2076 22.6726 19.9976 22.5416 19.9116 22.222C19.3433 20.1173 28.552 19.2325 27.7758 16.1839C27.639 15.6445 27.2677 15.4256 26.746 15.4263C24.4923 15.4263 19.4358 19.5181 18.3759 19.5181C18.2949 19.5181 18.2368 19.4937 18.2053 19.4419C17.6743 18.557 17.9653 17.9394 21.7082 15.6009C25.4511 13.2617 28.0783 11.8545 26.5841 10.1752C26.4121 9.98141 26.1684 9.8956 25.8725 9.8956C23.6001 9.89634 18.2311 14.9403 18.2311 14.9403C18.2311 14.9403 16.7821 16.496 15.9057 16.496C15.7043 16.496 15.533 16.4139 15.4169 16.2112C14.7956 15.1296 21.1879 10.1286 21.5484 8.06535C21.7928 6.66715 21.3771 5.95917 20.6081 5.95917Z" fill="#FF9D00"></path><path d="M5.50686 24.2246C3.53472 21.2387 3.67446 18.9979 6.38043 16.206C9.08641 13.4147 10.6615 9.33111 10.6615 9.33111C10.6615 9.33111 11.2499 6.95933 12.59 7.17757C13.93 7.39581 14.9139 10.9401 12.1069 13.1084C9.29997 15.276 12.6659 16.7489 13.7459 14.713C14.8258 12.6772 17.7747 7.44316 19.304 6.44221C20.8326 5.44128 21.9089 6.00204 21.5484 8.06532C21.188 10.1286 14.795 15.1295 15.4171 16.2118C16.0391 17.2934 18.2312 14.9402 18.2312 14.9402C18.2312 14.9402 25.0907 8.49588 26.5842 10.1752C28.0776 11.8545 25.4512 13.2616 21.7082 15.6008C17.9646 17.9393 17.6744 18.557 18.2054 19.4418C18.7372 20.3266 26.9998 13.1351 27.7759 16.1838C28.5513 19.2324 19.3434 20.1173 19.9117 22.2219C20.48 24.3274 26.3979 18.2382 27.6082 20.6107C28.8193 22.9839 19.2574 25.7722 19.18 25.7929C16.0914 26.62 8.24723 28.3726 5.50686 24.2246Z" fill="#FFD21E"></path></svg>
175
+ Community
176
+ <div class="ml-1.5 flex h-4 min-w-[1rem] items-center justify-center rounded px-1 text-xs leading-none shadow-sm bg-black text-white dark:bg-gray-800 dark:text-gray-200">30
177
+ </div>
178
+
179
+ </a>
180
+ </div>
181
+ </div></div></header>
182
+
183
+
184
+ </div>
185
+
186
+ <div class="container relative flex flex-col md:grid md:space-y-0 w-full md:grid-cols-12 space-y-4 md:gap-6 mb-16"><section class="pt-8 border-gray-100 col-span-full"><header class="flex flex-wrap items-center justify-start pb-2 md:justify-end lg:flex-nowrap"><div class="relative mr-4 flex min-w-0 basis-auto flex-wrap items-center md:flex-grow md:basis-full lg:basis-auto lg:flex-nowrap"><div class="SVELTE_HYDRATER contents" data-target="BranchSelector" data-props="{&quot;path&quot;:&quot;markdown/apache-2.0.md&quot;,&quot;repoName&quot;:&quot;choosealicense/licenses&quot;,&quot;repoType&quot;:&quot;dataset&quot;,&quot;rev&quot;:&quot;main&quot;,&quot;refs&quot;:{&quot;branches&quot;:[{&quot;name&quot;:&quot;main&quot;,&quot;ref&quot;:&quot;refs/heads/main&quot;,&quot;targetCommit&quot;:&quot;20edaed2b9e7dccd366d0654d4536fb377850680&quot;}],&quot;tags&quot;:[],&quot;converts&quot;:[]},&quot;view&quot;:&quot;blob&quot;}"><div class="relative mr-4 mb-2">
187
+ <button class="text-sm md:text-base btn w-full cursor-pointer text-sm" type="button">
188
+ <svg class="mr-1.5 text-gray-700 dark:text-gray-400" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 24 24" style="transform: rotate(360deg);"><path d="M13 14c-3.36 0-4.46 1.35-4.82 2.24C9.25 16.7 10 17.76 10 19a3 3 0 0 1-3 3a3 3 0 0 1-3-3c0-1.31.83-2.42 2-2.83V7.83A2.99 2.99 0 0 1 4 5a3 3 0 0 1 3-3a3 3 0 0 1 3 3c0 1.31-.83 2.42-2 2.83v5.29c.88-.65 2.16-1.12 4-1.12c2.67 0 3.56-1.34 3.85-2.23A3.006 3.006 0 0 1 14 7a3 3 0 0 1 3-3a3 3 0 0 1 3 3c0 1.34-.88 2.5-2.09 2.86C17.65 11.29 16.68 14 13 14m-6 4a1 1 0 0 0-1 1a1 1 0 0 0 1 1a1 1 0 0 0 1-1a1 1 0 0 0-1-1M7 4a1 1 0 0 0-1 1a1 1 0 0 0 1 1a1 1 0 0 0 1-1a1 1 0 0 0-1-1m10 2a1 1 0 0 0-1 1a1 1 0 0 0 1 1a1 1 0 0 0 1-1a1 1 0 0 0-1-1z" fill="currentColor"></path></svg>
189
+ main
190
+ <svg class="-mr-1 text-gray-500" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 24 24"><path d="M16.293 9.293L12 13.586L7.707 9.293l-1.414 1.414L12 16.414l5.707-5.707z" fill="currentColor"></path></svg></button>
191
+
192
+
193
+ </div></div>
194
+ <div class="relative mb-2 flex flex-wrap items-center"><a class="truncate text-gray-800 hover:underline" href="/datasets/choosealicense/licenses/tree/main">licenses</a>
195
+ <span class="mx-1 text-gray-300">/</span>
196
+ <a class="truncate hover:underline dark:text-gray-300" href="/datasets/choosealicense/licenses/tree/main/markdown">markdown
197
+ </a>
198
+ <span class="mx-1 text-gray-300">/</span><span class="dark:text-gray-300">apache-2.0.md</span>
199
+ <div class="SVELTE_HYDRATER contents" data-target="CopyButton" data-props="{&quot;value&quot;:&quot;markdown/apache-2.0.md&quot;,&quot;classNames&quot;:&quot;text-xs ml-2&quot;,&quot;title&quot;:&quot;Copy path&quot;}"><button class="relative text-xs ml-2 inline-flex cursor-pointer items-center text-sm focus:outline-none mx-0.5 text-gray-600 " title="Copy path" type="button"><svg class="" xmlns="http://www.w3.org/2000/svg" aria-hidden="true" fill="currentColor" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32"><path d="M28,10V28H10V10H28m0-2H10a2,2,0,0,0-2,2V28a2,2,0,0,0,2,2H28a2,2,0,0,0,2-2V10a2,2,0,0,0-2-2Z" transform="translate(0)"></path><path d="M4,18H2V4A2,2,0,0,1,4,2H18V4H4Z" transform="translate(0)"></path><rect fill="none" width="32" height="32"></rect></svg>
200
+
201
+ </button></div></div></div>
202
+
203
+
204
+ </header>
205
+ <div class="SVELTE_HYDRATER contents" data-target="LastCommit" data-props="{&quot;commitLast&quot;:{&quot;date&quot;:&quot;2024-04-17T10:17:35.000Z&quot;,&quot;verified&quot;:&quot;verified&quot;,&quot;subject&quot;:&quot;import&quot;,&quot;authors&quot;:[{&quot;_id&quot;:&quot;5dd96eb166059660ed1ee413&quot;,&quot;avatar&quot;:&quot;https://cdn-avatars.huggingface.co/v1/production/uploads/5dd96eb166059660ed1ee413/NQtzmrDdbG0H8qkZvRyGk.jpeg&quot;,&quot;isHf&quot;:true,&quot;user&quot;:&quot;julien-c&quot;}],&quot;commit&quot;:{&quot;id&quot;:&quot;20edaed2b9e7dccd366d0654d4536fb377850680&quot;,&quot;parentIds&quot;:[&quot;360db1f90dd48765f6dcaadb64b94203e71aade4&quot;]},&quot;title&quot;:&quot;import&quot;},&quot;repo&quot;:{&quot;name&quot;:&quot;choosealicense/licenses&quot;,&quot;type&quot;:&quot;dataset&quot;}}"><div class="from-gray-100-to-white flex items-baseline rounded-t-lg border border-b-0 bg-gradient-to-t px-3 py-2 dark:border-gray-800"><img class="mr-2.5 mt-0.5 h-4 w-4 self-center rounded-full" alt="julien-c's picture" src="https://cdn-avatars.huggingface.co/v1/production/uploads/5dd96eb166059660ed1ee413/NQtzmrDdbG0H8qkZvRyGk.jpeg">
206
+ <div class="mr-4 flex flex-none items-center truncate"><a class="hover:underline" href="/julien-c">julien-c
207
+ </a>
208
+ <span class="max-w-[175px] truncate border font-semibold leading-snug sm:max-w-xs rounded px-1 border-yellow-100 bg-yellow-50 text-yellow-500 dark:bg-yellow-800 dark:text-yellow-400 text-xs mt-0.5 ml-1.5 uppercase" title="member of the Hugging Face team">HF staff
209
+ </span>
210
+ </div>
211
+ <div class="mr-4 truncate font-mono text-sm text-gray-500 hover:prose-a:underline"><!-- HTML_TAG_START -->import<!-- HTML_TAG_END --></div>
212
+ <a class="rounded border bg-gray-50 px-1.5 text-sm hover:underline dark:border-gray-800 dark:bg-gray-900" href="/datasets/choosealicense/licenses/commit/20edaed2b9e7dccd366d0654d4536fb377850680">20edaed</a>
213
+ <span class="mx-2 text-green-500 dark:text-green-600 px-1.5 border-green-100 dark:border-green-800 rounded-full border text-xs uppercase" title="This commit is signed and the signature is verified">verified</span>
214
+ <time class="ml-auto hidden flex-none truncate pl-2 text-gray-500 dark:text-gray-400 lg:block" datetime="2024-04-17T10:17:35" title="Wed, 17 Apr 2024 10:17:35 GMT">8 months ago</time></div></div>
215
+ <div class="relative flex flex-wrap items-center border px-3 py-1.5 text-sm text-gray-800 dark:border-gray-800 dark:bg-gray-900 "><div class="flex items-center gap-3 text-sm font-medium"><a class="rounded-md px-1.5 capitalize bg-gray-200 dark:bg-gray-800" href="/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md">preview</a>
216
+ <a class="rounded-md px-1.5 capitalize " href="/datasets/choosealicense/licenses/blob/main/markdown/apache-2.0.md?code=true">code</a></div>
217
+ <div class="mx-4 text-gray-200">|</div>
218
+ <a class="my-1 mr-4 flex items-center hover:underline " href="/datasets/choosealicense/licenses/raw/main/markdown/apache-2.0.md"><svg class="mr-1.5" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32" style="transform: rotate(360deg);"><path d="M31 16l-7 7l-1.41-1.41L28.17 16l-5.58-5.59L24 9l7 7z" fill="currentColor"></path><path d="M1 16l7-7l1.41 1.41L3.83 16l5.58 5.59L8 23l-7-7z" fill="currentColor"></path><path d="M12.419 25.484L17.639 6l1.932.518L14.35 26z" fill="currentColor"></path></svg>
219
+ raw
220
+ </a><div class="SVELTE_HYDRATER contents" data-target="CopyButton" data-props="{&quot;value&quot;:&quot;https://huggingface.co/datasets/choosealicense/licenses/resolve/main/markdown/apache-2.0.md&quot;,&quot;style&quot;:&quot;blank&quot;,&quot;label&quot;:&quot;Copy download link&quot;,&quot;classNames&quot;:&quot;my-1 mr-4 flex items-center no-underline hover:underline&quot;}"><button class="relative my-1 mr-4 flex items-center no-underline hover:underline " title="Copy download link" type="button"><svg class="" xmlns="http://www.w3.org/2000/svg" aria-hidden="true" fill="currentColor" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32"><path d="M28,10V28H10V10H28m0-2H10a2,2,0,0,0-2,2V28a2,2,0,0,0,2,2H28a2,2,0,0,0,2-2V10a2,2,0,0,0-2-2Z" transform="translate(0)"></path><path d="M4,18H2V4A2,2,0,0,1,4,2H18V4H4Z" transform="translate(0)"></path><rect fill="none" width="32" height="32"></rect></svg>
221
+ <span class="ml-1.5 ">Copy download link</span>
222
+ </button></div><a class="my-1 mr-4 flex items-center hover:underline " href="/datasets/choosealicense/licenses/commits/main/markdown/apache-2.0.md"><svg class="mr-1.5" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32" style="transform: rotate(360deg);"><path d="M16 4C9.383 4 4 9.383 4 16s5.383 12 12 12s12-5.383 12-12S22.617 4 16 4zm0 2c5.535 0 10 4.465 10 10s-4.465 10-10 10S6 21.535 6 16S10.465 6 16 6zm-1 2v9h7v-2h-5V8z" fill="currentColor"></path></svg>
223
+ history
224
+ </a><a class="my-1 mr-4 flex items-center hover:underline " href="/datasets/choosealicense/licenses/blame/main/markdown/apache-2.0.md"><svg class="mr-1.5" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32" style="transform: rotate(360deg);"><path d="M16 2a14 14 0 1 0 14 14A14 14 0 0 0 16 2zm0 26a12 12 0 1 1 12-12a12 12 0 0 1-12 12z" fill="currentColor"></path><path d="M11.5 11a2.5 2.5 0 1 0 2.5 2.5a2.48 2.48 0 0 0-2.5-2.5z" fill="currentColor"></path><path d="M20.5 11a2.5 2.5 0 1 0 2.5 2.5a2.48 2.48 0 0 0-2.5-2.5z" fill="currentColor"></path></svg>
225
+ blame
226
+ </a><a class="my-1 mr-4 flex items-center hover:underline text-green-600 dark:text-gray-300" href="/datasets/choosealicense/licenses/edit/main/markdown/apache-2.0.md"><svg class="mr-1.5" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32"><path d="M2 26h28v2H2z" fill="currentColor"></path><path d="M25.4 9c.8-.8.8-2 0-2.8l-3.6-3.6c-.8-.8-2-.8-2.8 0l-15 15V24h6.4l15-15zm-5-5L24 7.6l-3 3L17.4 7l3-3zM6 22v-3.6l10-10l3.6 3.6l-10 10H6z" fill="currentColor"></path></svg>
227
+ contribute
228
+ </a><a class="my-1 mr-4 flex items-center hover:underline " href="/datasets/choosealicense/licenses/delete/main/markdown/apache-2.0.md"><svg class="mr-1.5" xmlns="http://www.w3.org/2000/svg" xmlns:xlink="http://www.w3.org/1999/xlink" aria-hidden="true" focusable="false" role="img" width="1em" height="1em" preserveAspectRatio="xMidYMid meet" viewBox="0 0 32 32"><path d="M12 12h2v12h-2z" fill="currentColor"></path><path d="M18 12h2v12h-2z" fill="currentColor"></path><path d="M4 6v2h2v20a2 2 0 0 0 2 2h16a2 2 0 0 0 2-2V8h2V6zm4 22V8h16v20z" fill="currentColor"></path><path d="M12 2h8v2h-8z" fill="currentColor"></path></svg>
229
+ delete
230
+ </a>
231
+
232
+ <div class="mr-4 flex items-center"><div class="SVELTE_HYDRATER contents" data-target="ScanStatusBadge" data-props="{&quot;classNames&quot;:&quot;mr-2&quot;,&quot;scanStatus&quot;:{&quot;status&quot;:&quot;safe&quot;,&quot;protectAiScan&quot;:{&quot;status&quot;:&quot;unscanned&quot;},&quot;avScan&quot;:{&quot;status&quot;:&quot;safe&quot;},&quot;pickleImportScan&quot;:{&quot;status&quot;:&quot;unscanned&quot;,&quot;pickleImports&quot;:[]}},&quot;repo&quot;:{&quot;_id&quot;:&quot;661f9fff32baa05e065be707&quot;,&quot;gitalyUid&quot;:&quot;3a16b4ced9f9d02b5debb80bb526cf577594dbaa403bbaed9e1bf1556dcd4b63&quot;,&quot;type&quot;:&quot;dataset&quot;,&quot;name&quot;:&quot;choosealicense/licenses&quot;,&quot;config&quot;:{&quot;private&quot;:false,&quot;gated&quot;:false,&quot;discussionsDisabled&quot;:false,&quot;duplicationDisabled&quot;:false,&quot;region&quot;:&quot;us&quot;,&quot;gitaly&quot;:{&quot;storage&quot;:&quot;default&quot;,&quot;repoUid&quot;:&quot;3a16b4ced9f9d02b5debb80bb526cf577594dbaa403bbaed9e1bf1556dcd4b63&quot;,&quot;region&quot;:&quot;us&quot;},&quot;lfs&quot;:{&quot;bucket&quot;:&quot;hf-hub-lfs-us-east-1&quot;,&quot;prefix&quot;:&quot;repos/3a/16/3a16b4ced9f9d02b5debb80bb526cf577594dbaa403bbaed9e1bf1556dcd4b63&quot;,&quot;usedStorage&quot;:73635203,&quot;usedStorageDatasetsServer&quot;:0},&quot;lastDiscussion&quot;:30},&quot;updatedAt&quot;:&quot;2024-04-17T10:10:07.391Z&quot;,&quot;authorId&quot;:&quot;661f9fb248cf6faa5fe2f700&quot;,&quot;creatorId&quot;:&quot;5dd96eb166059660ed1ee413&quot;},&quot;revision&quot;:&quot;main&quot;,&quot;filePath&quot;:&quot;markdown/apache-2.0.md&quot;,&quot;openByDefault&quot;:false}"><div class="sm:relative mr-2"><button class="flex h-[1.125rem] select-none items-center gap-0.5 rounded border pl-0.5 pr-0.5 text-xs leading-tight text-gray-400 hover:cursor-pointer text-gray-400 hover:border-gray-200 hover:bg-gray-50 hover:text-gray-500 dark:border-gray-800 dark:hover:bg-gray-800 dark:hover:text-gray-200 "><svg class="flex-none" width="1em" height="1em" viewBox="0 0 22 28" fill="none" xmlns="http://www.w3.org/2000/svg"><path fill-rule="evenodd" clip-rule="evenodd" d="M15.3634 10.3639C15.8486 10.8491 15.8486 11.6357 15.3634 12.1209L10.9292 16.5551C10.6058 16.8785 10.0814 16.8785 9.7579 16.5551L7.03051 13.8277C6.54532 13.3425 6.54532 12.5558 7.03051 12.0707C7.51569 11.5855 8.30234 11.5855 8.78752 12.0707L9.7579 13.041C10.0814 13.3645 10.6058 13.3645 10.9292 13.041L13.6064 10.3639C14.0916 9.8787 14.8782 9.8787 15.3634 10.3639Z" fill="currentColor"></path><path fill-rule="evenodd" clip-rule="evenodd" d="M10.6666 27.12C4.93329 25.28 0 19.2267 0 12.7867V6.52001C0 5.40001 0.693334 4.41334 1.73333 4.01334L9.73333 1.01334C10.3333 0.786673 11 0.786673 11.6 1.02667L19.6 4.02667C20.1083 4.21658 20.5465 4.55701 20.8562 5.00252C21.1659 5.44803 21.3324 5.97742 21.3333 6.52001V12.7867C21.3333 19.24 16.4 25.28 10.6666 27.12Z" fill="currentColor" fill-opacity="0.22"></path><path d="M10.0845 1.94967L10.0867 1.94881C10.4587 1.8083 10.8666 1.81036 11.2286 1.95515L11.2387 1.95919L11.2489 1.963L19.2489 4.963L19.25 4.96342C19.5677 5.08211 19.8416 5.29488 20.0351 5.57333C20.2285 5.85151 20.3326 6.18203 20.3333 6.52082C20.3333 6.52113 20.3333 6.52144 20.3333 6.52176L20.3333 12.7867C20.3333 18.6535 15.8922 24.2319 10.6666 26.0652C5.44153 24.2316 1 18.6409 1 12.7867V6.52001C1 5.82357 1.42893 5.20343 2.08883 4.94803L10.0845 1.94967Z" stroke="currentColor" stroke-opacity="0.30" stroke-width="2"></path></svg>
233
+
234
+ <span class="mr-0.5 max-sm:hidden">Safe</span></button>
235
+
236
+ </div></div>
237
+ </div>
238
+
239
+ <div class="flex items-center gap-x-3 dark:text-gray-300 sm:ml-auto">
240
+ 12.6 kB</div></div>
241
+
242
+ <div class="relative min-h-[100px] rounded-b-lg border border-t-0 leading-tight dark:border-gray-800 dark:bg-gray-925">
243
+
244
+ <div class="py-4 px-4 sm:px-6 prose hf-sanitized hf-sanitized-uvHZzz4Js8_cYPxEWqtUb">
245
+ <!-- HTML_TAG_START --><pre><code> Apache License
246
+ Version 2.0, January 2004
247
+ http://www.apache.org/licenses/
248
+ </code></pre>
249
+ <p> TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION</p>
250
+ <ol>
251
+ <li><p>Definitions.</p>
252
+ <p>"License" shall mean the terms and conditions for use, reproduction,
253
+ and distribution as defined by Sections 1 through 9 of this document.</p>
254
+ <p>"Licensor" shall mean the copyright owner or entity authorized by
255
+ the copyright owner that is granting the License.</p>
256
+ <p>"Legal Entity" shall mean the union of the acting entity and all
257
+ other entities that control, are controlled by, or are under common
258
+ control with that entity. For the purposes of this definition,
259
+ "control" means (i) the power, direct or indirect, to cause the
260
+ direction or management of such entity, whether by contract or
261
+ otherwise, or (ii) ownership of fifty percent (50%) or more of the
262
+ outstanding shares, or (iii) beneficial ownership of such entity.</p>
263
+ <p>"You" (or "Your") shall mean an individual or Legal Entity
264
+ exercising permissions granted by this License.</p>
265
+ <p>"Source" form shall mean the preferred form for making modifications,
266
+ including but not limited to software source code, documentation
267
+ source, and configuration files.</p>
268
+ <p>"Object" form shall mean any form resulting from mechanical
269
+ transformation or translation of a Source form, including but
270
+ not limited to compiled object code, generated documentation,
271
+ and conversions to other media types.</p>
272
+ <p>"Work" shall mean the work of authorship, whether in Source or
273
+ Object form, made available under the License, as indicated by a
274
+ copyright notice that is included in or attached to the work
275
+ (an example is provided in the Appendix below).</p>
276
+ <p>"Derivative Works" shall mean any work, whether in Source or Object
277
+ form, that is based on (or derived from) the Work and for which the
278
+ editorial revisions, annotations, elaborations, or other modifications
279
+ represent, as a whole, an original work of authorship. For the purposes
280
+ of this License, Derivative Works shall not include works that remain
281
+ separable from, or merely link (or bind by name) to the interfaces of,
282
+ the Work and Derivative Works thereof.</p>
283
+ <p>"Contribution" shall mean any work of authorship, including
284
+ the original version of the Work and any modifications or additions
285
+ to that Work or Derivative Works thereof, that is intentionally
286
+ submitted to Licensor for inclusion in the Work by the copyright owner
287
+ or by an individual or Legal Entity authorized to submit on behalf of
288
+ the copyright owner. For the purposes of this definition, "submitted"
289
+ means any form of electronic, verbal, or written communication sent
290
+ to the Licensor or its representatives, including but not limited to
291
+ communication on electronic mailing lists, source code control systems,
292
+ and issue tracking systems that are managed by, or on behalf of, the
293
+ Licensor for the purpose of discussing and improving the Work, but
294
+ excluding communication that is conspicuously marked or otherwise
295
+ designated in writing by the copyright owner as "Not a Contribution."</p>
296
+ <p>"Contributor" shall mean Licensor and any individual or Legal Entity
297
+ on behalf of whom a Contribution has been received by Licensor and
298
+ subsequently incorporated within the Work.</p>
299
+ </li>
300
+ <li><p>Grant of Copyright License. Subject to the terms and conditions of
301
+ this License, each Contributor hereby grants to You a perpetual,
302
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
303
+ copyright license to reproduce, prepare Derivative Works of,
304
+ publicly display, publicly perform, sublicense, and distribute the
305
+ Work and such Derivative Works in Source or Object form.</p>
306
+ </li>
307
+ <li><p>Grant of Patent License. Subject to the terms and conditions of
308
+ this License, each Contributor hereby grants to You a perpetual,
309
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
310
+ (except as stated in this section) patent license to make, have made,
311
+ use, offer to sell, sell, import, and otherwise transfer the Work,
312
+ where such license applies only to those patent claims licensable
313
+ by such Contributor that are necessarily infringed by their
314
+ Contribution(s) alone or by combination of their Contribution(s)
315
+ with the Work to which such Contribution(s) was submitted. If You
316
+ institute patent litigation against any entity (including a
317
+ cross-claim or counterclaim in a lawsuit) alleging that the Work
318
+ or a Contribution incorporated within the Work constitutes direct
319
+ or contributory patent infringement, then any patent licenses
320
+ granted to You under this License for that Work shall terminate
321
+ as of the date such litigation is filed.</p>
322
+ </li>
323
+ <li><p>Redistribution. You may reproduce and distribute copies of the
324
+ Work or Derivative Works thereof in any medium, with or without
325
+ modifications, and in Source or Object form, provided that You
326
+ meet the following conditions:</p>
327
+ <p>(a) You must give any other recipients of the Work or
328
+ Derivative Works a copy of this License; and
329
+ (b) You must cause any modified files to carry prominent notices
330
+ stating that You changed the files; and
331
+ (c) You must retain, in the Source form of any Derivative Works
332
+ that You distribute, all copyright, patent, trademark, and
333
+ attribution notices from the Source form of the Work,
334
+ excluding those notices that do not pertain to any part of
335
+ the Derivative Works; and
336
+ (d) If the Work includes a "NOTICE" text file as part of its
337
+ distribution, then any Derivative Works that You distribute must
338
+ include a readable copy of the attribution notices contained
339
+ within such NOTICE file, excluding those notices that do not
340
+ pertain to any part of the Derivative Works, in at least one
341
+ of the following places: within a NOTICE text file distributed
342
+ as part of the Derivative Works; within the Source form or
343
+ documentation, if provided along with the Derivative Works; or,
344
+ within a display generated by the Derivative Works, if and
345
+ wherever such third-party notices normally appear. The contents
346
+ of the NOTICE file are for informational purposes only and
347
+ do not modify the License. You may add Your own attribution
348
+ notices within Derivative Works that You distribute, alongside
349
+ or as an addendum to the NOTICE text from the Work, provided
350
+ that such additional attribution notices cannot be construed
351
+ as modifying the License.
352
+ You may add Your own copyright statement to Your modifications and
353
+ may provide additional or different license terms and conditions
354
+ for use, reproduction, or distribution of Your modifications, or
355
+ for any such Derivative Works as a whole, provided Your use,
356
+ reproduction, and distribution of the Work otherwise complies with
357
+ the conditions stated in this License.</p>
358
+ </li>
359
+ <li><p>Submission of Contributions. Unless You explicitly state otherwise,
360
+ any Contribution intentionally submitted for inclusion in the Work
361
+ by You to the Licensor shall be under the terms and conditions of
362
+ this License, without any additional terms or conditions.
363
+ Notwithstanding the above, nothing herein shall supersede or modify
364
+ the terms of any separate license agreement you may have executed
365
+ with Licensor regarding such Contributions.</p>
366
+ </li>
367
+ <li><p>Trademarks. This License does not grant permission to use the trade
368
+ names, trademarks, service marks, or product names of the Licensor,
369
+ except as required for reasonable and customary use in describing the
370
+ origin of the Work and reproducing the content of the NOTICE file.</p>
371
+ </li>
372
+ <li><p>Disclaimer of Warranty. Unless required by applicable law or
373
+ agreed to in writing, Licensor provides the Work (and each
374
+ Contributor provides its Contributions) on an "AS IS" BASIS,
375
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
376
+ implied, including, without limitation, any warranties or conditions
377
+ of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
378
+ PARTICULAR PURPOSE. You are solely responsible for determining the
379
+ appropriateness of using or redistributing the Work and assume any
380
+ risks associated with Your exercise of permissions under this License.</p>
381
+ </li>
382
+ <li><p>Limitation of Liability. In no event and under no legal theory,
383
+ whether in tort (including negligence), contract, or otherwise,
384
+ unless required by applicable law (such as deliberate and grossly
385
+ negligent acts) or agreed to in writing, shall any Contributor be
386
+ liable to You for damages, including any direct, indirect, special,
387
+ incidental, or consequential damages of any character arising as a
388
+ result of this License or out of the use or inability to use the
389
+ Work (including but not limited to damages for loss of goodwill,
390
+ work stoppage, computer failure or malfunction, or any and all
391
+ other commercial damages or losses), even if such Contributor
392
+ has been advised of the possibility of such damages.</p>
393
+ </li>
394
+ <li><p>Accepting Warranty or Additional Liability. While redistributing
395
+ the Work or Derivative Works thereof, You may choose to offer,
396
+ and charge a fee for, acceptance of support, warranty, indemnity,
397
+ or other liability obligations and/or rights consistent with this
398
+ License. However, in accepting such obligations, You may act only
399
+ on Your own behalf and on Your sole responsibility, not on behalf
400
+ of any other Contributor, and only if You agree to indemnify,
401
+ defend, and hold each Contributor harmless for any liability
402
+ incurred by, or claims asserted against, such Contributor by reason
403
+ of your accepting any such warranty or additional liability.</p>
404
+ </li>
405
+ </ol>
406
+ <p> END OF TERMS AND CONDITIONS</p>
407
+ <p> APPENDIX: How to apply the Apache License to your work.</p>
408
+ <pre><code> To apply the Apache License to your work, attach the following
409
+ boilerplate notice, with the fields enclosed by brackets "[]"
410
+ replaced with your own identifying information. (Don't include
411
+ the brackets!) The text should be enclosed in the appropriate
412
+ comment syntax for the file format. We also recommend that a
413
+ file or class name and description of purpose be included on the
414
+ same "printed page" as the copyright notice for easier
415
+ identification within third-party archives.
416
+ </code></pre>
417
+ <p> Copyright [yyyy] [name of copyright owner]</p>
418
+ <p> Licensed under the Apache License, Version 2.0 (the "License");
419
+ you may not use this file except in compliance with the License.
420
+ You may obtain a copy of the License at</p>
421
+ <pre><code> http://www.apache.org/licenses/LICENSE-2.0
422
+ </code></pre>
423
+ <p> Unless required by applicable law or agreed to in writing, software
424
+ distributed under the License is distributed on an "AS IS" BASIS,
425
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
426
+ See the License for the specific language governing permissions and
427
+ limitations under the License.</p>
428
+ <!-- HTML_TAG_END --></div>
429
+ </div></section></div></main>
430
+
431
+ </div>
432
+
433
+ <script>
434
+ import("\/front\/build\/kube-f68c51b\/index.js");
435
+ window.moonSha = "kube-f68c51b\/";
436
+ window.__hf_deferred = {};
437
+ </script>
438
+
439
+ <!-- Stripe -->
440
+ <script>
441
+ if (["hf.co", "huggingface.co"].includes(window.location.hostname)) {
442
+ const script = document.createElement("script");
443
+ script.src = "https://js.stripe.com/v3/";
444
+ script.async = true;
445
+ document.head.appendChild(script);
446
+ }
447
+ </script>
448
+ </body>
449
+ </html>
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MixtralForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 4096,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 14336,
12
+ "max_position_embeddings": 32768,
13
+ "model_type": "mixtral",
14
+ "num_attention_heads": 32,
15
+ "num_experts_per_tok": 2,
16
+ "num_hidden_layers": 32,
17
+ "num_key_value_heads": 8,
18
+ "num_local_experts": 8,
19
+ "output_router_logits": false,
20
+ "rms_norm_eps": 1e-05,
21
+ "rope_theta": 1000000.0,
22
+ "router_aux_loss_coef": 0.02,
23
+ "sliding_window": null,
24
+ "tie_word_embeddings": false,
25
+ "torch_dtype": "bfloat16",
26
+ "transformers_version": "4.36.0.dev0",
27
+ "use_cache": true,
28
+ "vocab_size": 32000
29
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.36.0.dev0"
6
+ }
model-00001-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a3193f0eab51d9cd1bcb05eab88a2c79dfda50757182ea019ac833329c7e5eb
3
+ size 4920052720
model-00002-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e05b0842bc3ee92b898e6a21effa513566993e3274028a0a53ae0391296b6ad9
3
+ size 4865559920
model-00003-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2b84a4c79a9b65f7003ecd0a2c18820314609c654137ba8b18f252854f64327
3
+ size 4865559920
model-00004-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a19935f9ffc5547642420de501d07ba8e02094577a83acc6a77b6976824714e
3
+ size 4865559920
model-00005-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bde64f94e5f93464130fc50006426bab95781af4efd28e5498ab5084b0686abb
3
+ size 4865559920
model-00006-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b665498c4929138af419a0671d484d8a256774e8e73dd140ae1c8866359feec4
3
+ size 4932504264
model-00007-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1032ca26df1deb0d7e95a6b357c871319ad83cd94e820379446bada20e761c7e
3
+ size 4865559912
model-00008-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:420c33428b05c5fde204e93ab2dac8de73b6f0cc5eaedee12d7da256ab3c6b6b
3
+ size 4865559920
model-00009-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:765a1cc5edc41c977aee05b0227c6a71ed3e16b4b7ff955b2e9832a0dbfee4e6
3
+ size 4865559920
model-00010-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88a4e7903ffcc69b12b7a6607f600c9f054484725f3eaabadbc7d0901f4f4798
3
+ size 4865559920
model-00011-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a83fb10bea4aa348a9c353b5b502d44a0f47b0fd6a4bbfb1798c78b547713c5
3
+ size 4865559920
model-00012-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec8723486bb3e804463a0f5a1190d6690257182a0229dcc89c0c0999fb85c387
3
+ size 4999646240
model-00013-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46af28a6483d7a71c87b67f8b41341a1432c4dbcaa7fa18880dc97c5c381e40a
3
+ size 4798417968
model-00014-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:065a252ea4c470842e802c9bc10aed6cc2095882e7dc5471e5da765901e9ce1d
3
+ size 4865559944
model-00015-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:058801599ee1474327d5a36eb19b702614f3c297b6e8258756a7b5a3e3837155
3
+ size 4865559944
model-00016-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8a4e97f2e531d7e83b8d6c568e0a8a377d708b22f3363264c9e719f029a2156
3
+ size 4865559944
model-00017-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84a9f6193c86af06b9269062cb3b8cc6dfd9a8f980351525d34d6253e139c71d
3
+ size 4865559944
model-00018-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a612a532620f17f64a5c834c1954edf1497c2440ac584abab851e6d1c9c1190
3
+ size 4865559944
model-00019-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69785b654ba8420dfcc5563fb0581fac8fa9a8af8553e66e13d4f22aab012c88
3
+ size 4932504280
model-00020-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e75deed93e97eb9833b9a5cb1d249bb3881fcdbace05a8a47e6a4be7d661d0ee
3
+ size 4865559944
model-00021-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89da2f5603989e1b4980313f176f4dd07d2916428526ff7485479ee1bae6b2ad
3
+ size 4865559944
model-00022-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94024537639748f57e2eca8b04a4e9b1df624f65abb137b2674a780865cd38fa
3
+ size 4865559944
model-00023-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e900da246ae4e89f852caa57b62513f2c35be47e7a35c642ff8664f8522e8257
3
+ size 4865559944
model-00024-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fa9a18f5f3cba3772d5d48deff29defbdeb1f4580de6ffadc53ac12454924ae
3
+ size 4865559944
model-00025-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d413c4efe147a695dfc92fcf52ad5b25879cd2decd62401ef33c35b104834ed
3
+ size 4932504280
model-00026-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:245e6ece6e026e40f7ca8b35a6ea3af5ca098e9ed06383c88629f9e0fb89b3c3
3
+ size 4865559944
model-00027-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d0b035d11f1ea48c508d53bb710c998670abd2fd980ec6ab6afb588b00f2813
3
+ size 4865559944
model-00028-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9daec6dcbe843294f3c94d7adbeb86e2add36b65ab71df1dea8d377fb83dda6
3
+ size 4865559944
model-00029-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99d62713b696159db600f77720a7962294455c64ba14e73da5d31de40bce76dd
3
+ size 4865559944
model-00030-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd3dfd95ea0ec512cdb94768e51ab556dcdb2f93f742971e1547492beb646d26
3
+ size 4865559944
model-00031-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48df382fb3f55cccea78aca62dcba44b300d3470a11cd32f5d794c07ad5e701a
3
+ size 4932504280
model-00032-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80a36fc4f1f7e646d1470ba5ca6dcab69851bfeb0256ab66a663d9c78fcc739d
3
+ size 4865559944
model-00033-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8db2cfd63a739e57f5433dc43321004847d92133040197d956a8d684ad471f18
3
+ size 4865559944
model-00034-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d817387fd7642accfcd553716060cc56b7c9a74daf0d86a3649587950c1f83cf
3
+ size 4865559944
model-00035-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96e613a1c148840a9aacae66335517bb3484ab37cd9ef4984cf1c784c82cc0d9
3
+ size 4865559944
model-00036-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7997e191eba4c2be252f58e41fdb52985d0a37603dbf8fe895e1b4026a59940
3
+ size 4865559944
model-00037-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1b2a643a38e1242d49078ee94796e4d4f324964ea08bdc166f110b392d093d7
3
+ size 4999646264
model-00038-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5de718c1e9919cf9cb710d7030716468b946c6b29d48a443883d3b270e782bc6
3
+ size 4798417968
model-00039-of-00039.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e7db3e7e09db7d3dbe09715853d753958362cb31b45252a20c870d83a5fc7a7
3
+ size 1463862216
model.safetensors.index.json ADDED
@@ -0,0 +1,1002 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 186811170816
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00039-of-00039.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00039.safetensors",
8
+ "model.layers.0.block_sparse_moe.experts.0.w1.weight": "model-00001-of-00039.safetensors",
9
+ "model.layers.0.block_sparse_moe.experts.0.w2.weight": "model-00001-of-00039.safetensors",
10
+ "model.layers.0.block_sparse_moe.experts.0.w3.weight": "model-00001-of-00039.safetensors",
11
+ "model.layers.0.block_sparse_moe.experts.1.w1.weight": "model-00001-of-00039.safetensors",
12
+ "model.layers.0.block_sparse_moe.experts.1.w2.weight": "model-00001-of-00039.safetensors",
13
+ "model.layers.0.block_sparse_moe.experts.1.w3.weight": "model-00001-of-00039.safetensors",
14
+ "model.layers.0.block_sparse_moe.experts.2.w1.weight": "model-00001-of-00039.safetensors",
15
+ "model.layers.0.block_sparse_moe.experts.2.w2.weight": "model-00001-of-00039.safetensors",
16
+ "model.layers.0.block_sparse_moe.experts.2.w3.weight": "model-00001-of-00039.safetensors",
17
+ "model.layers.0.block_sparse_moe.experts.3.w1.weight": "model-00001-of-00039.safetensors",
18
+ "model.layers.0.block_sparse_moe.experts.3.w2.weight": "model-00001-of-00039.safetensors",
19
+ "model.layers.0.block_sparse_moe.experts.3.w3.weight": "model-00001-of-00039.safetensors",
20
+ "model.layers.0.block_sparse_moe.experts.4.w1.weight": "model-00001-of-00039.safetensors",
21
+ "model.layers.0.block_sparse_moe.experts.4.w2.weight": "model-00001-of-00039.safetensors",
22
+ "model.layers.0.block_sparse_moe.experts.4.w3.weight": "model-00001-of-00039.safetensors",
23
+ "model.layers.0.block_sparse_moe.experts.5.w1.weight": "model-00001-of-00039.safetensors",
24
+ "model.layers.0.block_sparse_moe.experts.5.w2.weight": "model-00001-of-00039.safetensors",
25
+ "model.layers.0.block_sparse_moe.experts.5.w3.weight": "model-00001-of-00039.safetensors",
26
+ "model.layers.0.block_sparse_moe.experts.6.w1.weight": "model-00002-of-00039.safetensors",
27
+ "model.layers.0.block_sparse_moe.experts.6.w2.weight": "model-00002-of-00039.safetensors",
28
+ "model.layers.0.block_sparse_moe.experts.6.w3.weight": "model-00002-of-00039.safetensors",
29
+ "model.layers.0.block_sparse_moe.experts.7.w1.weight": "model-00002-of-00039.safetensors",
30
+ "model.layers.0.block_sparse_moe.experts.7.w2.weight": "model-00002-of-00039.safetensors",
31
+ "model.layers.0.block_sparse_moe.experts.7.w3.weight": "model-00002-of-00039.safetensors",
32
+ "model.layers.0.block_sparse_moe.gate.weight": "model-00001-of-00039.safetensors",
33
+ "model.layers.0.input_layernorm.weight": "model-00002-of-00039.safetensors",
34
+ "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00039.safetensors",
35
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00039.safetensors",
36
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00039.safetensors",
37
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00039.safetensors",
38
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00039.safetensors",
39
+ "model.layers.1.block_sparse_moe.experts.0.w1.weight": "model-00002-of-00039.safetensors",
40
+ "model.layers.1.block_sparse_moe.experts.0.w2.weight": "model-00002-of-00039.safetensors",
41
+ "model.layers.1.block_sparse_moe.experts.0.w3.weight": "model-00002-of-00039.safetensors",
42
+ "model.layers.1.block_sparse_moe.experts.1.w1.weight": "model-00002-of-00039.safetensors",
43
+ "model.layers.1.block_sparse_moe.experts.1.w2.weight": "model-00002-of-00039.safetensors",
44
+ "model.layers.1.block_sparse_moe.experts.1.w3.weight": "model-00002-of-00039.safetensors",
45
+ "model.layers.1.block_sparse_moe.experts.2.w1.weight": "model-00002-of-00039.safetensors",
46
+ "model.layers.1.block_sparse_moe.experts.2.w2.weight": "model-00002-of-00039.safetensors",
47
+ "model.layers.1.block_sparse_moe.experts.2.w3.weight": "model-00002-of-00039.safetensors",
48
+ "model.layers.1.block_sparse_moe.experts.3.w1.weight": "model-00002-of-00039.safetensors",
49
+ "model.layers.1.block_sparse_moe.experts.3.w2.weight": "model-00002-of-00039.safetensors",
50
+ "model.layers.1.block_sparse_moe.experts.3.w3.weight": "model-00002-of-00039.safetensors",
51
+ "model.layers.1.block_sparse_moe.experts.4.w1.weight": "model-00002-of-00039.safetensors",
52
+ "model.layers.1.block_sparse_moe.experts.4.w2.weight": "model-00002-of-00039.safetensors",
53
+ "model.layers.1.block_sparse_moe.experts.4.w3.weight": "model-00003-of-00039.safetensors",
54
+ "model.layers.1.block_sparse_moe.experts.5.w1.weight": "model-00003-of-00039.safetensors",
55
+ "model.layers.1.block_sparse_moe.experts.5.w2.weight": "model-00003-of-00039.safetensors",
56
+ "model.layers.1.block_sparse_moe.experts.5.w3.weight": "model-00003-of-00039.safetensors",
57
+ "model.layers.1.block_sparse_moe.experts.6.w1.weight": "model-00003-of-00039.safetensors",
58
+ "model.layers.1.block_sparse_moe.experts.6.w2.weight": "model-00003-of-00039.safetensors",
59
+ "model.layers.1.block_sparse_moe.experts.6.w3.weight": "model-00003-of-00039.safetensors",
60
+ "model.layers.1.block_sparse_moe.experts.7.w1.weight": "model-00003-of-00039.safetensors",
61
+ "model.layers.1.block_sparse_moe.experts.7.w2.weight": "model-00003-of-00039.safetensors",
62
+ "model.layers.1.block_sparse_moe.experts.7.w3.weight": "model-00003-of-00039.safetensors",
63
+ "model.layers.1.block_sparse_moe.gate.weight": "model-00002-of-00039.safetensors",
64
+ "model.layers.1.input_layernorm.weight": "model-00003-of-00039.safetensors",
65
+ "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00039.safetensors",
66
+ "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00039.safetensors",
67
+ "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00039.safetensors",
68
+ "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00039.safetensors",
69
+ "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00039.safetensors",
70
+ "model.layers.10.block_sparse_moe.experts.0.w1.weight": "model-00013-of-00039.safetensors",
71
+ "model.layers.10.block_sparse_moe.experts.0.w2.weight": "model-00013-of-00039.safetensors",
72
+ "model.layers.10.block_sparse_moe.experts.0.w3.weight": "model-00013-of-00039.safetensors",
73
+ "model.layers.10.block_sparse_moe.experts.1.w1.weight": "model-00013-of-00039.safetensors",
74
+ "model.layers.10.block_sparse_moe.experts.1.w2.weight": "model-00013-of-00039.safetensors",
75
+ "model.layers.10.block_sparse_moe.experts.1.w3.weight": "model-00013-of-00039.safetensors",
76
+ "model.layers.10.block_sparse_moe.experts.2.w1.weight": "model-00013-of-00039.safetensors",
77
+ "model.layers.10.block_sparse_moe.experts.2.w2.weight": "model-00013-of-00039.safetensors",
78
+ "model.layers.10.block_sparse_moe.experts.2.w3.weight": "model-00013-of-00039.safetensors",
79
+ "model.layers.10.block_sparse_moe.experts.3.w1.weight": "model-00013-of-00039.safetensors",
80
+ "model.layers.10.block_sparse_moe.experts.3.w2.weight": "model-00013-of-00039.safetensors",
81
+ "model.layers.10.block_sparse_moe.experts.3.w3.weight": "model-00013-of-00039.safetensors",
82
+ "model.layers.10.block_sparse_moe.experts.4.w1.weight": "model-00013-of-00039.safetensors",
83
+ "model.layers.10.block_sparse_moe.experts.4.w2.weight": "model-00013-of-00039.safetensors",
84
+ "model.layers.10.block_sparse_moe.experts.4.w3.weight": "model-00013-of-00039.safetensors",
85
+ "model.layers.10.block_sparse_moe.experts.5.w1.weight": "model-00013-of-00039.safetensors",
86
+ "model.layers.10.block_sparse_moe.experts.5.w2.weight": "model-00013-of-00039.safetensors",
87
+ "model.layers.10.block_sparse_moe.experts.5.w3.weight": "model-00013-of-00039.safetensors",
88
+ "model.layers.10.block_sparse_moe.experts.6.w1.weight": "model-00013-of-00039.safetensors",
89
+ "model.layers.10.block_sparse_moe.experts.6.w2.weight": "model-00013-of-00039.safetensors",
90
+ "model.layers.10.block_sparse_moe.experts.6.w3.weight": "model-00014-of-00039.safetensors",
91
+ "model.layers.10.block_sparse_moe.experts.7.w1.weight": "model-00014-of-00039.safetensors",
92
+ "model.layers.10.block_sparse_moe.experts.7.w2.weight": "model-00014-of-00039.safetensors",
93
+ "model.layers.10.block_sparse_moe.experts.7.w3.weight": "model-00014-of-00039.safetensors",
94
+ "model.layers.10.block_sparse_moe.gate.weight": "model-00013-of-00039.safetensors",
95
+ "model.layers.10.input_layernorm.weight": "model-00014-of-00039.safetensors",
96
+ "model.layers.10.post_attention_layernorm.weight": "model-00014-of-00039.safetensors",
97
+ "model.layers.10.self_attn.k_proj.weight": "model-00013-of-00039.safetensors",
98
+ "model.layers.10.self_attn.o_proj.weight": "model-00013-of-00039.safetensors",
99
+ "model.layers.10.self_attn.q_proj.weight": "model-00012-of-00039.safetensors",
100
+ "model.layers.10.self_attn.v_proj.weight": "model-00013-of-00039.safetensors",
101
+ "model.layers.11.block_sparse_moe.experts.0.w1.weight": "model-00014-of-00039.safetensors",
102
+ "model.layers.11.block_sparse_moe.experts.0.w2.weight": "model-00014-of-00039.safetensors",
103
+ "model.layers.11.block_sparse_moe.experts.0.w3.weight": "model-00014-of-00039.safetensors",
104
+ "model.layers.11.block_sparse_moe.experts.1.w1.weight": "model-00014-of-00039.safetensors",
105
+ "model.layers.11.block_sparse_moe.experts.1.w2.weight": "model-00014-of-00039.safetensors",
106
+ "model.layers.11.block_sparse_moe.experts.1.w3.weight": "model-00014-of-00039.safetensors",
107
+ "model.layers.11.block_sparse_moe.experts.2.w1.weight": "model-00014-of-00039.safetensors",
108
+ "model.layers.11.block_sparse_moe.experts.2.w2.weight": "model-00014-of-00039.safetensors",
109
+ "model.layers.11.block_sparse_moe.experts.2.w3.weight": "model-00014-of-00039.safetensors",
110
+ "model.layers.11.block_sparse_moe.experts.3.w1.weight": "model-00014-of-00039.safetensors",
111
+ "model.layers.11.block_sparse_moe.experts.3.w2.weight": "model-00014-of-00039.safetensors",
112
+ "model.layers.11.block_sparse_moe.experts.3.w3.weight": "model-00014-of-00039.safetensors",
113
+ "model.layers.11.block_sparse_moe.experts.4.w1.weight": "model-00014-of-00039.safetensors",
114
+ "model.layers.11.block_sparse_moe.experts.4.w2.weight": "model-00014-of-00039.safetensors",
115
+ "model.layers.11.block_sparse_moe.experts.4.w3.weight": "model-00014-of-00039.safetensors",
116
+ "model.layers.11.block_sparse_moe.experts.5.w1.weight": "model-00014-of-00039.safetensors",
117
+ "model.layers.11.block_sparse_moe.experts.5.w2.weight": "model-00015-of-00039.safetensors",
118
+ "model.layers.11.block_sparse_moe.experts.5.w3.weight": "model-00015-of-00039.safetensors",
119
+ "model.layers.11.block_sparse_moe.experts.6.w1.weight": "model-00015-of-00039.safetensors",
120
+ "model.layers.11.block_sparse_moe.experts.6.w2.weight": "model-00015-of-00039.safetensors",
121
+ "model.layers.11.block_sparse_moe.experts.6.w3.weight": "model-00015-of-00039.safetensors",
122
+ "model.layers.11.block_sparse_moe.experts.7.w1.weight": "model-00015-of-00039.safetensors",
123
+ "model.layers.11.block_sparse_moe.experts.7.w2.weight": "model-00015-of-00039.safetensors",
124
+ "model.layers.11.block_sparse_moe.experts.7.w3.weight": "model-00015-of-00039.safetensors",
125
+ "model.layers.11.block_sparse_moe.gate.weight": "model-00014-of-00039.safetensors",
126
+ "model.layers.11.input_layernorm.weight": "model-00015-of-00039.safetensors",
127
+ "model.layers.11.post_attention_layernorm.weight": "model-00015-of-00039.safetensors",
128
+ "model.layers.11.self_attn.k_proj.weight": "model-00014-of-00039.safetensors",
129
+ "model.layers.11.self_attn.o_proj.weight": "model-00014-of-00039.safetensors",
130
+ "model.layers.11.self_attn.q_proj.weight": "model-00014-of-00039.safetensors",
131
+ "model.layers.11.self_attn.v_proj.weight": "model-00014-of-00039.safetensors",
132
+ "model.layers.12.block_sparse_moe.experts.0.w1.weight": "model-00015-of-00039.safetensors",
133
+ "model.layers.12.block_sparse_moe.experts.0.w2.weight": "model-00015-of-00039.safetensors",
134
+ "model.layers.12.block_sparse_moe.experts.0.w3.weight": "model-00015-of-00039.safetensors",
135
+ "model.layers.12.block_sparse_moe.experts.1.w1.weight": "model-00015-of-00039.safetensors",
136
+ "model.layers.12.block_sparse_moe.experts.1.w2.weight": "model-00015-of-00039.safetensors",
137
+ "model.layers.12.block_sparse_moe.experts.1.w3.weight": "model-00015-of-00039.safetensors",
138
+ "model.layers.12.block_sparse_moe.experts.2.w1.weight": "model-00015-of-00039.safetensors",
139
+ "model.layers.12.block_sparse_moe.experts.2.w2.weight": "model-00015-of-00039.safetensors",
140
+ "model.layers.12.block_sparse_moe.experts.2.w3.weight": "model-00015-of-00039.safetensors",
141
+ "model.layers.12.block_sparse_moe.experts.3.w1.weight": "model-00015-of-00039.safetensors",
142
+ "model.layers.12.block_sparse_moe.experts.3.w2.weight": "model-00015-of-00039.safetensors",
143
+ "model.layers.12.block_sparse_moe.experts.3.w3.weight": "model-00015-of-00039.safetensors",
144
+ "model.layers.12.block_sparse_moe.experts.4.w1.weight": "model-00016-of-00039.safetensors",
145
+ "model.layers.12.block_sparse_moe.experts.4.w2.weight": "model-00016-of-00039.safetensors",
146
+ "model.layers.12.block_sparse_moe.experts.4.w3.weight": "model-00016-of-00039.safetensors",
147
+ "model.layers.12.block_sparse_moe.experts.5.w1.weight": "model-00016-of-00039.safetensors",
148
+ "model.layers.12.block_sparse_moe.experts.5.w2.weight": "model-00016-of-00039.safetensors",
149
+ "model.layers.12.block_sparse_moe.experts.5.w3.weight": "model-00016-of-00039.safetensors",
150
+ "model.layers.12.block_sparse_moe.experts.6.w1.weight": "model-00016-of-00039.safetensors",
151
+ "model.layers.12.block_sparse_moe.experts.6.w2.weight": "model-00016-of-00039.safetensors",
152
+ "model.layers.12.block_sparse_moe.experts.6.w3.weight": "model-00016-of-00039.safetensors",
153
+ "model.layers.12.block_sparse_moe.experts.7.w1.weight": "model-00016-of-00039.safetensors",
154
+ "model.layers.12.block_sparse_moe.experts.7.w2.weight": "model-00016-of-00039.safetensors",
155
+ "model.layers.12.block_sparse_moe.experts.7.w3.weight": "model-00016-of-00039.safetensors",
156
+ "model.layers.12.block_sparse_moe.gate.weight": "model-00015-of-00039.safetensors",
157
+ "model.layers.12.input_layernorm.weight": "model-00016-of-00039.safetensors",
158
+ "model.layers.12.post_attention_layernorm.weight": "model-00016-of-00039.safetensors",
159
+ "model.layers.12.self_attn.k_proj.weight": "model-00015-of-00039.safetensors",
160
+ "model.layers.12.self_attn.o_proj.weight": "model-00015-of-00039.safetensors",
161
+ "model.layers.12.self_attn.q_proj.weight": "model-00015-of-00039.safetensors",
162
+ "model.layers.12.self_attn.v_proj.weight": "model-00015-of-00039.safetensors",
163
+ "model.layers.13.block_sparse_moe.experts.0.w1.weight": "model-00016-of-00039.safetensors",
164
+ "model.layers.13.block_sparse_moe.experts.0.w2.weight": "model-00016-of-00039.safetensors",
165
+ "model.layers.13.block_sparse_moe.experts.0.w3.weight": "model-00016-of-00039.safetensors",
166
+ "model.layers.13.block_sparse_moe.experts.1.w1.weight": "model-00016-of-00039.safetensors",
167
+ "model.layers.13.block_sparse_moe.experts.1.w2.weight": "model-00016-of-00039.safetensors",
168
+ "model.layers.13.block_sparse_moe.experts.1.w3.weight": "model-00016-of-00039.safetensors",
169
+ "model.layers.13.block_sparse_moe.experts.2.w1.weight": "model-00016-of-00039.safetensors",
170
+ "model.layers.13.block_sparse_moe.experts.2.w2.weight": "model-00016-of-00039.safetensors",
171
+ "model.layers.13.block_sparse_moe.experts.2.w3.weight": "model-00017-of-00039.safetensors",
172
+ "model.layers.13.block_sparse_moe.experts.3.w1.weight": "model-00017-of-00039.safetensors",
173
+ "model.layers.13.block_sparse_moe.experts.3.w2.weight": "model-00017-of-00039.safetensors",
174
+ "model.layers.13.block_sparse_moe.experts.3.w3.weight": "model-00017-of-00039.safetensors",
175
+ "model.layers.13.block_sparse_moe.experts.4.w1.weight": "model-00017-of-00039.safetensors",
176
+ "model.layers.13.block_sparse_moe.experts.4.w2.weight": "model-00017-of-00039.safetensors",
177
+ "model.layers.13.block_sparse_moe.experts.4.w3.weight": "model-00017-of-00039.safetensors",
178
+ "model.layers.13.block_sparse_moe.experts.5.w1.weight": "model-00017-of-00039.safetensors",
179
+ "model.layers.13.block_sparse_moe.experts.5.w2.weight": "model-00017-of-00039.safetensors",
180
+ "model.layers.13.block_sparse_moe.experts.5.w3.weight": "model-00017-of-00039.safetensors",
181
+ "model.layers.13.block_sparse_moe.experts.6.w1.weight": "model-00017-of-00039.safetensors",
182
+ "model.layers.13.block_sparse_moe.experts.6.w2.weight": "model-00017-of-00039.safetensors",
183
+ "model.layers.13.block_sparse_moe.experts.6.w3.weight": "model-00017-of-00039.safetensors",
184
+ "model.layers.13.block_sparse_moe.experts.7.w1.weight": "model-00017-of-00039.safetensors",
185
+ "model.layers.13.block_sparse_moe.experts.7.w2.weight": "model-00017-of-00039.safetensors",
186
+ "model.layers.13.block_sparse_moe.experts.7.w3.weight": "model-00017-of-00039.safetensors",
187
+ "model.layers.13.block_sparse_moe.gate.weight": "model-00016-of-00039.safetensors",
188
+ "model.layers.13.input_layernorm.weight": "model-00017-of-00039.safetensors",
189
+ "model.layers.13.post_attention_layernorm.weight": "model-00017-of-00039.safetensors",
190
+ "model.layers.13.self_attn.k_proj.weight": "model-00016-of-00039.safetensors",
191
+ "model.layers.13.self_attn.o_proj.weight": "model-00016-of-00039.safetensors",
192
+ "model.layers.13.self_attn.q_proj.weight": "model-00016-of-00039.safetensors",
193
+ "model.layers.13.self_attn.v_proj.weight": "model-00016-of-00039.safetensors",
194
+ "model.layers.14.block_sparse_moe.experts.0.w1.weight": "model-00017-of-00039.safetensors",
195
+ "model.layers.14.block_sparse_moe.experts.0.w2.weight": "model-00017-of-00039.safetensors",
196
+ "model.layers.14.block_sparse_moe.experts.0.w3.weight": "model-00017-of-00039.safetensors",
197
+ "model.layers.14.block_sparse_moe.experts.1.w1.weight": "model-00017-of-00039.safetensors",
198
+ "model.layers.14.block_sparse_moe.experts.1.w2.weight": "model-00018-of-00039.safetensors",
199
+ "model.layers.14.block_sparse_moe.experts.1.w3.weight": "model-00018-of-00039.safetensors",
200
+ "model.layers.14.block_sparse_moe.experts.2.w1.weight": "model-00018-of-00039.safetensors",
201
+ "model.layers.14.block_sparse_moe.experts.2.w2.weight": "model-00018-of-00039.safetensors",
202
+ "model.layers.14.block_sparse_moe.experts.2.w3.weight": "model-00018-of-00039.safetensors",
203
+ "model.layers.14.block_sparse_moe.experts.3.w1.weight": "model-00018-of-00039.safetensors",
204
+ "model.layers.14.block_sparse_moe.experts.3.w2.weight": "model-00018-of-00039.safetensors",
205
+ "model.layers.14.block_sparse_moe.experts.3.w3.weight": "model-00018-of-00039.safetensors",
206
+ "model.layers.14.block_sparse_moe.experts.4.w1.weight": "model-00018-of-00039.safetensors",
207
+ "model.layers.14.block_sparse_moe.experts.4.w2.weight": "model-00018-of-00039.safetensors",
208
+ "model.layers.14.block_sparse_moe.experts.4.w3.weight": "model-00018-of-00039.safetensors",
209
+ "model.layers.14.block_sparse_moe.experts.5.w1.weight": "model-00018-of-00039.safetensors",
210
+ "model.layers.14.block_sparse_moe.experts.5.w2.weight": "model-00018-of-00039.safetensors",
211
+ "model.layers.14.block_sparse_moe.experts.5.w3.weight": "model-00018-of-00039.safetensors",
212
+ "model.layers.14.block_sparse_moe.experts.6.w1.weight": "model-00018-of-00039.safetensors",
213
+ "model.layers.14.block_sparse_moe.experts.6.w2.weight": "model-00018-of-00039.safetensors",
214
+ "model.layers.14.block_sparse_moe.experts.6.w3.weight": "model-00018-of-00039.safetensors",
215
+ "model.layers.14.block_sparse_moe.experts.7.w1.weight": "model-00018-of-00039.safetensors",
216
+ "model.layers.14.block_sparse_moe.experts.7.w2.weight": "model-00018-of-00039.safetensors",
217
+ "model.layers.14.block_sparse_moe.experts.7.w3.weight": "model-00018-of-00039.safetensors",
218
+ "model.layers.14.block_sparse_moe.gate.weight": "model-00017-of-00039.safetensors",
219
+ "model.layers.14.input_layernorm.weight": "model-00018-of-00039.safetensors",
220
+ "model.layers.14.post_attention_layernorm.weight": "model-00018-of-00039.safetensors",
221
+ "model.layers.14.self_attn.k_proj.weight": "model-00017-of-00039.safetensors",
222
+ "model.layers.14.self_attn.o_proj.weight": "model-00017-of-00039.safetensors",
223
+ "model.layers.14.self_attn.q_proj.weight": "model-00017-of-00039.safetensors",
224
+ "model.layers.14.self_attn.v_proj.weight": "model-00017-of-00039.safetensors",
225
+ "model.layers.15.block_sparse_moe.experts.0.w1.weight": "model-00019-of-00039.safetensors",
226
+ "model.layers.15.block_sparse_moe.experts.0.w2.weight": "model-00019-of-00039.safetensors",
227
+ "model.layers.15.block_sparse_moe.experts.0.w3.weight": "model-00019-of-00039.safetensors",
228
+ "model.layers.15.block_sparse_moe.experts.1.w1.weight": "model-00019-of-00039.safetensors",
229
+ "model.layers.15.block_sparse_moe.experts.1.w2.weight": "model-00019-of-00039.safetensors",
230
+ "model.layers.15.block_sparse_moe.experts.1.w3.weight": "model-00019-of-00039.safetensors",
231
+ "model.layers.15.block_sparse_moe.experts.2.w1.weight": "model-00019-of-00039.safetensors",
232
+ "model.layers.15.block_sparse_moe.experts.2.w2.weight": "model-00019-of-00039.safetensors",
233
+ "model.layers.15.block_sparse_moe.experts.2.w3.weight": "model-00019-of-00039.safetensors",
234
+ "model.layers.15.block_sparse_moe.experts.3.w1.weight": "model-00019-of-00039.safetensors",
235
+ "model.layers.15.block_sparse_moe.experts.3.w2.weight": "model-00019-of-00039.safetensors",
236
+ "model.layers.15.block_sparse_moe.experts.3.w3.weight": "model-00019-of-00039.safetensors",
237
+ "model.layers.15.block_sparse_moe.experts.4.w1.weight": "model-00019-of-00039.safetensors",
238
+ "model.layers.15.block_sparse_moe.experts.4.w2.weight": "model-00019-of-00039.safetensors",
239
+ "model.layers.15.block_sparse_moe.experts.4.w3.weight": "model-00019-of-00039.safetensors",
240
+ "model.layers.15.block_sparse_moe.experts.5.w1.weight": "model-00019-of-00039.safetensors",
241
+ "model.layers.15.block_sparse_moe.experts.5.w2.weight": "model-00019-of-00039.safetensors",
242
+ "model.layers.15.block_sparse_moe.experts.5.w3.weight": "model-00019-of-00039.safetensors",
243
+ "model.layers.15.block_sparse_moe.experts.6.w1.weight": "model-00019-of-00039.safetensors",
244
+ "model.layers.15.block_sparse_moe.experts.6.w2.weight": "model-00019-of-00039.safetensors",
245
+ "model.layers.15.block_sparse_moe.experts.6.w3.weight": "model-00019-of-00039.safetensors",
246
+ "model.layers.15.block_sparse_moe.experts.7.w1.weight": "model-00020-of-00039.safetensors",
247
+ "model.layers.15.block_sparse_moe.experts.7.w2.weight": "model-00020-of-00039.safetensors",
248
+ "model.layers.15.block_sparse_moe.experts.7.w3.weight": "model-00020-of-00039.safetensors",
249
+ "model.layers.15.block_sparse_moe.gate.weight": "model-00018-of-00039.safetensors",
250
+ "model.layers.15.input_layernorm.weight": "model-00020-of-00039.safetensors",
251
+ "model.layers.15.post_attention_layernorm.weight": "model-00020-of-00039.safetensors",
252
+ "model.layers.15.self_attn.k_proj.weight": "model-00018-of-00039.safetensors",
253
+ "model.layers.15.self_attn.o_proj.weight": "model-00018-of-00039.safetensors",
254
+ "model.layers.15.self_attn.q_proj.weight": "model-00018-of-00039.safetensors",
255
+ "model.layers.15.self_attn.v_proj.weight": "model-00018-of-00039.safetensors",
256
+ "model.layers.16.block_sparse_moe.experts.0.w1.weight": "model-00020-of-00039.safetensors",
257
+ "model.layers.16.block_sparse_moe.experts.0.w2.weight": "model-00020-of-00039.safetensors",
258
+ "model.layers.16.block_sparse_moe.experts.0.w3.weight": "model-00020-of-00039.safetensors",
259
+ "model.layers.16.block_sparse_moe.experts.1.w1.weight": "model-00020-of-00039.safetensors",
260
+ "model.layers.16.block_sparse_moe.experts.1.w2.weight": "model-00020-of-00039.safetensors",
261
+ "model.layers.16.block_sparse_moe.experts.1.w3.weight": "model-00020-of-00039.safetensors",
262
+ "model.layers.16.block_sparse_moe.experts.2.w1.weight": "model-00020-of-00039.safetensors",
263
+ "model.layers.16.block_sparse_moe.experts.2.w2.weight": "model-00020-of-00039.safetensors",
264
+ "model.layers.16.block_sparse_moe.experts.2.w3.weight": "model-00020-of-00039.safetensors",
265
+ "model.layers.16.block_sparse_moe.experts.3.w1.weight": "model-00020-of-00039.safetensors",
266
+ "model.layers.16.block_sparse_moe.experts.3.w2.weight": "model-00020-of-00039.safetensors",
267
+ "model.layers.16.block_sparse_moe.experts.3.w3.weight": "model-00020-of-00039.safetensors",
268
+ "model.layers.16.block_sparse_moe.experts.4.w1.weight": "model-00020-of-00039.safetensors",
269
+ "model.layers.16.block_sparse_moe.experts.4.w2.weight": "model-00020-of-00039.safetensors",
270
+ "model.layers.16.block_sparse_moe.experts.4.w3.weight": "model-00020-of-00039.safetensors",
271
+ "model.layers.16.block_sparse_moe.experts.5.w1.weight": "model-00020-of-00039.safetensors",
272
+ "model.layers.16.block_sparse_moe.experts.5.w2.weight": "model-00020-of-00039.safetensors",
273
+ "model.layers.16.block_sparse_moe.experts.5.w3.weight": "model-00021-of-00039.safetensors",
274
+ "model.layers.16.block_sparse_moe.experts.6.w1.weight": "model-00021-of-00039.safetensors",
275
+ "model.layers.16.block_sparse_moe.experts.6.w2.weight": "model-00021-of-00039.safetensors",
276
+ "model.layers.16.block_sparse_moe.experts.6.w3.weight": "model-00021-of-00039.safetensors",
277
+ "model.layers.16.block_sparse_moe.experts.7.w1.weight": "model-00021-of-00039.safetensors",
278
+ "model.layers.16.block_sparse_moe.experts.7.w2.weight": "model-00021-of-00039.safetensors",
279
+ "model.layers.16.block_sparse_moe.experts.7.w3.weight": "model-00021-of-00039.safetensors",
280
+ "model.layers.16.block_sparse_moe.gate.weight": "model-00020-of-00039.safetensors",
281
+ "model.layers.16.input_layernorm.weight": "model-00021-of-00039.safetensors",
282
+ "model.layers.16.post_attention_layernorm.weight": "model-00021-of-00039.safetensors",
283
+ "model.layers.16.self_attn.k_proj.weight": "model-00020-of-00039.safetensors",
284
+ "model.layers.16.self_attn.o_proj.weight": "model-00020-of-00039.safetensors",
285
+ "model.layers.16.self_attn.q_proj.weight": "model-00020-of-00039.safetensors",
286
+ "model.layers.16.self_attn.v_proj.weight": "model-00020-of-00039.safetensors",
287
+ "model.layers.17.block_sparse_moe.experts.0.w1.weight": "model-00021-of-00039.safetensors",
288
+ "model.layers.17.block_sparse_moe.experts.0.w2.weight": "model-00021-of-00039.safetensors",
289
+ "model.layers.17.block_sparse_moe.experts.0.w3.weight": "model-00021-of-00039.safetensors",
290
+ "model.layers.17.block_sparse_moe.experts.1.w1.weight": "model-00021-of-00039.safetensors",
291
+ "model.layers.17.block_sparse_moe.experts.1.w2.weight": "model-00021-of-00039.safetensors",
292
+ "model.layers.17.block_sparse_moe.experts.1.w3.weight": "model-00021-of-00039.safetensors",
293
+ "model.layers.17.block_sparse_moe.experts.2.w1.weight": "model-00021-of-00039.safetensors",
294
+ "model.layers.17.block_sparse_moe.experts.2.w2.weight": "model-00021-of-00039.safetensors",
295
+ "model.layers.17.block_sparse_moe.experts.2.w3.weight": "model-00021-of-00039.safetensors",
296
+ "model.layers.17.block_sparse_moe.experts.3.w1.weight": "model-00021-of-00039.safetensors",
297
+ "model.layers.17.block_sparse_moe.experts.3.w2.weight": "model-00021-of-00039.safetensors",
298
+ "model.layers.17.block_sparse_moe.experts.3.w3.weight": "model-00021-of-00039.safetensors",
299
+ "model.layers.17.block_sparse_moe.experts.4.w1.weight": "model-00021-of-00039.safetensors",
300
+ "model.layers.17.block_sparse_moe.experts.4.w2.weight": "model-00022-of-00039.safetensors",
301
+ "model.layers.17.block_sparse_moe.experts.4.w3.weight": "model-00022-of-00039.safetensors",
302
+ "model.layers.17.block_sparse_moe.experts.5.w1.weight": "model-00022-of-00039.safetensors",
303
+ "model.layers.17.block_sparse_moe.experts.5.w2.weight": "model-00022-of-00039.safetensors",
304
+ "model.layers.17.block_sparse_moe.experts.5.w3.weight": "model-00022-of-00039.safetensors",
305
+ "model.layers.17.block_sparse_moe.experts.6.w1.weight": "model-00022-of-00039.safetensors",
306
+ "model.layers.17.block_sparse_moe.experts.6.w2.weight": "model-00022-of-00039.safetensors",
307
+ "model.layers.17.block_sparse_moe.experts.6.w3.weight": "model-00022-of-00039.safetensors",
308
+ "model.layers.17.block_sparse_moe.experts.7.w1.weight": "model-00022-of-00039.safetensors",
309
+ "model.layers.17.block_sparse_moe.experts.7.w2.weight": "model-00022-of-00039.safetensors",
310
+ "model.layers.17.block_sparse_moe.experts.7.w3.weight": "model-00022-of-00039.safetensors",
311
+ "model.layers.17.block_sparse_moe.gate.weight": "model-00021-of-00039.safetensors",
312
+ "model.layers.17.input_layernorm.weight": "model-00022-of-00039.safetensors",
313
+ "model.layers.17.post_attention_layernorm.weight": "model-00022-of-00039.safetensors",
314
+ "model.layers.17.self_attn.k_proj.weight": "model-00021-of-00039.safetensors",
315
+ "model.layers.17.self_attn.o_proj.weight": "model-00021-of-00039.safetensors",
316
+ "model.layers.17.self_attn.q_proj.weight": "model-00021-of-00039.safetensors",
317
+ "model.layers.17.self_attn.v_proj.weight": "model-00021-of-00039.safetensors",
318
+ "model.layers.18.block_sparse_moe.experts.0.w1.weight": "model-00022-of-00039.safetensors",
319
+ "model.layers.18.block_sparse_moe.experts.0.w2.weight": "model-00022-of-00039.safetensors",
320
+ "model.layers.18.block_sparse_moe.experts.0.w3.weight": "model-00022-of-00039.safetensors",
321
+ "model.layers.18.block_sparse_moe.experts.1.w1.weight": "model-00022-of-00039.safetensors",
322
+ "model.layers.18.block_sparse_moe.experts.1.w2.weight": "model-00022-of-00039.safetensors",
323
+ "model.layers.18.block_sparse_moe.experts.1.w3.weight": "model-00022-of-00039.safetensors",
324
+ "model.layers.18.block_sparse_moe.experts.2.w1.weight": "model-00022-of-00039.safetensors",
325
+ "model.layers.18.block_sparse_moe.experts.2.w2.weight": "model-00022-of-00039.safetensors",
326
+ "model.layers.18.block_sparse_moe.experts.2.w3.weight": "model-00022-of-00039.safetensors",
327
+ "model.layers.18.block_sparse_moe.experts.3.w1.weight": "model-00023-of-00039.safetensors",
328
+ "model.layers.18.block_sparse_moe.experts.3.w2.weight": "model-00023-of-00039.safetensors",
329
+ "model.layers.18.block_sparse_moe.experts.3.w3.weight": "model-00023-of-00039.safetensors",
330
+ "model.layers.18.block_sparse_moe.experts.4.w1.weight": "model-00023-of-00039.safetensors",
331
+ "model.layers.18.block_sparse_moe.experts.4.w2.weight": "model-00023-of-00039.safetensors",
332
+ "model.layers.18.block_sparse_moe.experts.4.w3.weight": "model-00023-of-00039.safetensors",
333
+ "model.layers.18.block_sparse_moe.experts.5.w1.weight": "model-00023-of-00039.safetensors",
334
+ "model.layers.18.block_sparse_moe.experts.5.w2.weight": "model-00023-of-00039.safetensors",
335
+ "model.layers.18.block_sparse_moe.experts.5.w3.weight": "model-00023-of-00039.safetensors",
336
+ "model.layers.18.block_sparse_moe.experts.6.w1.weight": "model-00023-of-00039.safetensors",
337
+ "model.layers.18.block_sparse_moe.experts.6.w2.weight": "model-00023-of-00039.safetensors",
338
+ "model.layers.18.block_sparse_moe.experts.6.w3.weight": "model-00023-of-00039.safetensors",
339
+ "model.layers.18.block_sparse_moe.experts.7.w1.weight": "model-00023-of-00039.safetensors",
340
+ "model.layers.18.block_sparse_moe.experts.7.w2.weight": "model-00023-of-00039.safetensors",
341
+ "model.layers.18.block_sparse_moe.experts.7.w3.weight": "model-00023-of-00039.safetensors",
342
+ "model.layers.18.block_sparse_moe.gate.weight": "model-00022-of-00039.safetensors",
343
+ "model.layers.18.input_layernorm.weight": "model-00023-of-00039.safetensors",
344
+ "model.layers.18.post_attention_layernorm.weight": "model-00023-of-00039.safetensors",
345
+ "model.layers.18.self_attn.k_proj.weight": "model-00022-of-00039.safetensors",
346
+ "model.layers.18.self_attn.o_proj.weight": "model-00022-of-00039.safetensors",
347
+ "model.layers.18.self_attn.q_proj.weight": "model-00022-of-00039.safetensors",
348
+ "model.layers.18.self_attn.v_proj.weight": "model-00022-of-00039.safetensors",
349
+ "model.layers.19.block_sparse_moe.experts.0.w1.weight": "model-00023-of-00039.safetensors",
350
+ "model.layers.19.block_sparse_moe.experts.0.w2.weight": "model-00023-of-00039.safetensors",
351
+ "model.layers.19.block_sparse_moe.experts.0.w3.weight": "model-00023-of-00039.safetensors",
352
+ "model.layers.19.block_sparse_moe.experts.1.w1.weight": "model-00023-of-00039.safetensors",
353
+ "model.layers.19.block_sparse_moe.experts.1.w2.weight": "model-00023-of-00039.safetensors",
354
+ "model.layers.19.block_sparse_moe.experts.1.w3.weight": "model-00024-of-00039.safetensors",
355
+ "model.layers.19.block_sparse_moe.experts.2.w1.weight": "model-00024-of-00039.safetensors",
356
+ "model.layers.19.block_sparse_moe.experts.2.w2.weight": "model-00024-of-00039.safetensors",
357
+ "model.layers.19.block_sparse_moe.experts.2.w3.weight": "model-00024-of-00039.safetensors",
358
+ "model.layers.19.block_sparse_moe.experts.3.w1.weight": "model-00024-of-00039.safetensors",
359
+ "model.layers.19.block_sparse_moe.experts.3.w2.weight": "model-00024-of-00039.safetensors",
360
+ "model.layers.19.block_sparse_moe.experts.3.w3.weight": "model-00024-of-00039.safetensors",
361
+ "model.layers.19.block_sparse_moe.experts.4.w1.weight": "model-00024-of-00039.safetensors",
362
+ "model.layers.19.block_sparse_moe.experts.4.w2.weight": "model-00024-of-00039.safetensors",
363
+ "model.layers.19.block_sparse_moe.experts.4.w3.weight": "model-00024-of-00039.safetensors",
364
+ "model.layers.19.block_sparse_moe.experts.5.w1.weight": "model-00024-of-00039.safetensors",
365
+ "model.layers.19.block_sparse_moe.experts.5.w2.weight": "model-00024-of-00039.safetensors",
366
+ "model.layers.19.block_sparse_moe.experts.5.w3.weight": "model-00024-of-00039.safetensors",
367
+ "model.layers.19.block_sparse_moe.experts.6.w1.weight": "model-00024-of-00039.safetensors",
368
+ "model.layers.19.block_sparse_moe.experts.6.w2.weight": "model-00024-of-00039.safetensors",
369
+ "model.layers.19.block_sparse_moe.experts.6.w3.weight": "model-00024-of-00039.safetensors",
370
+ "model.layers.19.block_sparse_moe.experts.7.w1.weight": "model-00024-of-00039.safetensors",
371
+ "model.layers.19.block_sparse_moe.experts.7.w2.weight": "model-00024-of-00039.safetensors",
372
+ "model.layers.19.block_sparse_moe.experts.7.w3.weight": "model-00024-of-00039.safetensors",
373
+ "model.layers.19.block_sparse_moe.gate.weight": "model-00023-of-00039.safetensors",
374
+ "model.layers.19.input_layernorm.weight": "model-00024-of-00039.safetensors",
375
+ "model.layers.19.post_attention_layernorm.weight": "model-00024-of-00039.safetensors",
376
+ "model.layers.19.self_attn.k_proj.weight": "model-00023-of-00039.safetensors",
377
+ "model.layers.19.self_attn.o_proj.weight": "model-00023-of-00039.safetensors",
378
+ "model.layers.19.self_attn.q_proj.weight": "model-00023-of-00039.safetensors",
379
+ "model.layers.19.self_attn.v_proj.weight": "model-00023-of-00039.safetensors",
380
+ "model.layers.2.block_sparse_moe.experts.0.w1.weight": "model-00003-of-00039.safetensors",
381
+ "model.layers.2.block_sparse_moe.experts.0.w2.weight": "model-00003-of-00039.safetensors",
382
+ "model.layers.2.block_sparse_moe.experts.0.w3.weight": "model-00003-of-00039.safetensors",
383
+ "model.layers.2.block_sparse_moe.experts.1.w1.weight": "model-00003-of-00039.safetensors",
384
+ "model.layers.2.block_sparse_moe.experts.1.w2.weight": "model-00003-of-00039.safetensors",
385
+ "model.layers.2.block_sparse_moe.experts.1.w3.weight": "model-00003-of-00039.safetensors",
386
+ "model.layers.2.block_sparse_moe.experts.2.w1.weight": "model-00003-of-00039.safetensors",
387
+ "model.layers.2.block_sparse_moe.experts.2.w2.weight": "model-00003-of-00039.safetensors",
388
+ "model.layers.2.block_sparse_moe.experts.2.w3.weight": "model-00003-of-00039.safetensors",
389
+ "model.layers.2.block_sparse_moe.experts.3.w1.weight": "model-00003-of-00039.safetensors",
390
+ "model.layers.2.block_sparse_moe.experts.3.w2.weight": "model-00004-of-00039.safetensors",
391
+ "model.layers.2.block_sparse_moe.experts.3.w3.weight": "model-00004-of-00039.safetensors",
392
+ "model.layers.2.block_sparse_moe.experts.4.w1.weight": "model-00004-of-00039.safetensors",
393
+ "model.layers.2.block_sparse_moe.experts.4.w2.weight": "model-00004-of-00039.safetensors",
394
+ "model.layers.2.block_sparse_moe.experts.4.w3.weight": "model-00004-of-00039.safetensors",
395
+ "model.layers.2.block_sparse_moe.experts.5.w1.weight": "model-00004-of-00039.safetensors",
396
+ "model.layers.2.block_sparse_moe.experts.5.w2.weight": "model-00004-of-00039.safetensors",
397
+ "model.layers.2.block_sparse_moe.experts.5.w3.weight": "model-00004-of-00039.safetensors",
398
+ "model.layers.2.block_sparse_moe.experts.6.w1.weight": "model-00004-of-00039.safetensors",
399
+ "model.layers.2.block_sparse_moe.experts.6.w2.weight": "model-00004-of-00039.safetensors",
400
+ "model.layers.2.block_sparse_moe.experts.6.w3.weight": "model-00004-of-00039.safetensors",
401
+ "model.layers.2.block_sparse_moe.experts.7.w1.weight": "model-00004-of-00039.safetensors",
402
+ "model.layers.2.block_sparse_moe.experts.7.w2.weight": "model-00004-of-00039.safetensors",
403
+ "model.layers.2.block_sparse_moe.experts.7.w3.weight": "model-00004-of-00039.safetensors",
404
+ "model.layers.2.block_sparse_moe.gate.weight": "model-00003-of-00039.safetensors",
405
+ "model.layers.2.input_layernorm.weight": "model-00004-of-00039.safetensors",
406
+ "model.layers.2.post_attention_layernorm.weight": "model-00004-of-00039.safetensors",
407
+ "model.layers.2.self_attn.k_proj.weight": "model-00003-of-00039.safetensors",
408
+ "model.layers.2.self_attn.o_proj.weight": "model-00003-of-00039.safetensors",
409
+ "model.layers.2.self_attn.q_proj.weight": "model-00003-of-00039.safetensors",
410
+ "model.layers.2.self_attn.v_proj.weight": "model-00003-of-00039.safetensors",
411
+ "model.layers.20.block_sparse_moe.experts.0.w1.weight": "model-00024-of-00039.safetensors",
412
+ "model.layers.20.block_sparse_moe.experts.0.w2.weight": "model-00025-of-00039.safetensors",
413
+ "model.layers.20.block_sparse_moe.experts.0.w3.weight": "model-00025-of-00039.safetensors",
414
+ "model.layers.20.block_sparse_moe.experts.1.w1.weight": "model-00025-of-00039.safetensors",
415
+ "model.layers.20.block_sparse_moe.experts.1.w2.weight": "model-00025-of-00039.safetensors",
416
+ "model.layers.20.block_sparse_moe.experts.1.w3.weight": "model-00025-of-00039.safetensors",
417
+ "model.layers.20.block_sparse_moe.experts.2.w1.weight": "model-00025-of-00039.safetensors",
418
+ "model.layers.20.block_sparse_moe.experts.2.w2.weight": "model-00025-of-00039.safetensors",
419
+ "model.layers.20.block_sparse_moe.experts.2.w3.weight": "model-00025-of-00039.safetensors",
420
+ "model.layers.20.block_sparse_moe.experts.3.w1.weight": "model-00025-of-00039.safetensors",
421
+ "model.layers.20.block_sparse_moe.experts.3.w2.weight": "model-00025-of-00039.safetensors",
422
+ "model.layers.20.block_sparse_moe.experts.3.w3.weight": "model-00025-of-00039.safetensors",
423
+ "model.layers.20.block_sparse_moe.experts.4.w1.weight": "model-00025-of-00039.safetensors",
424
+ "model.layers.20.block_sparse_moe.experts.4.w2.weight": "model-00025-of-00039.safetensors",
425
+ "model.layers.20.block_sparse_moe.experts.4.w3.weight": "model-00025-of-00039.safetensors",
426
+ "model.layers.20.block_sparse_moe.experts.5.w1.weight": "model-00025-of-00039.safetensors",
427
+ "model.layers.20.block_sparse_moe.experts.5.w2.weight": "model-00025-of-00039.safetensors",
428
+ "model.layers.20.block_sparse_moe.experts.5.w3.weight": "model-00025-of-00039.safetensors",
429
+ "model.layers.20.block_sparse_moe.experts.6.w1.weight": "model-00025-of-00039.safetensors",
430
+ "model.layers.20.block_sparse_moe.experts.6.w2.weight": "model-00025-of-00039.safetensors",
431
+ "model.layers.20.block_sparse_moe.experts.6.w3.weight": "model-00025-of-00039.safetensors",
432
+ "model.layers.20.block_sparse_moe.experts.7.w1.weight": "model-00025-of-00039.safetensors",
433
+ "model.layers.20.block_sparse_moe.experts.7.w2.weight": "model-00026-of-00039.safetensors",
434
+ "model.layers.20.block_sparse_moe.experts.7.w3.weight": "model-00026-of-00039.safetensors",
435
+ "model.layers.20.block_sparse_moe.gate.weight": "model-00024-of-00039.safetensors",
436
+ "model.layers.20.input_layernorm.weight": "model-00026-of-00039.safetensors",
437
+ "model.layers.20.post_attention_layernorm.weight": "model-00026-of-00039.safetensors",
438
+ "model.layers.20.self_attn.k_proj.weight": "model-00024-of-00039.safetensors",
439
+ "model.layers.20.self_attn.o_proj.weight": "model-00024-of-00039.safetensors",
440
+ "model.layers.20.self_attn.q_proj.weight": "model-00024-of-00039.safetensors",
441
+ "model.layers.20.self_attn.v_proj.weight": "model-00024-of-00039.safetensors",
442
+ "model.layers.21.block_sparse_moe.experts.0.w1.weight": "model-00026-of-00039.safetensors",
443
+ "model.layers.21.block_sparse_moe.experts.0.w2.weight": "model-00026-of-00039.safetensors",
444
+ "model.layers.21.block_sparse_moe.experts.0.w3.weight": "model-00026-of-00039.safetensors",
445
+ "model.layers.21.block_sparse_moe.experts.1.w1.weight": "model-00026-of-00039.safetensors",
446
+ "model.layers.21.block_sparse_moe.experts.1.w2.weight": "model-00026-of-00039.safetensors",
447
+ "model.layers.21.block_sparse_moe.experts.1.w3.weight": "model-00026-of-00039.safetensors",
448
+ "model.layers.21.block_sparse_moe.experts.2.w1.weight": "model-00026-of-00039.safetensors",
449
+ "model.layers.21.block_sparse_moe.experts.2.w2.weight": "model-00026-of-00039.safetensors",
450
+ "model.layers.21.block_sparse_moe.experts.2.w3.weight": "model-00026-of-00039.safetensors",
451
+ "model.layers.21.block_sparse_moe.experts.3.w1.weight": "model-00026-of-00039.safetensors",
452
+ "model.layers.21.block_sparse_moe.experts.3.w2.weight": "model-00026-of-00039.safetensors",
453
+ "model.layers.21.block_sparse_moe.experts.3.w3.weight": "model-00026-of-00039.safetensors",
454
+ "model.layers.21.block_sparse_moe.experts.4.w1.weight": "model-00026-of-00039.safetensors",
455
+ "model.layers.21.block_sparse_moe.experts.4.w2.weight": "model-00026-of-00039.safetensors",
456
+ "model.layers.21.block_sparse_moe.experts.4.w3.weight": "model-00026-of-00039.safetensors",
457
+ "model.layers.21.block_sparse_moe.experts.5.w1.weight": "model-00026-of-00039.safetensors",
458
+ "model.layers.21.block_sparse_moe.experts.5.w2.weight": "model-00026-of-00039.safetensors",
459
+ "model.layers.21.block_sparse_moe.experts.5.w3.weight": "model-00026-of-00039.safetensors",
460
+ "model.layers.21.block_sparse_moe.experts.6.w1.weight": "model-00027-of-00039.safetensors",
461
+ "model.layers.21.block_sparse_moe.experts.6.w2.weight": "model-00027-of-00039.safetensors",
462
+ "model.layers.21.block_sparse_moe.experts.6.w3.weight": "model-00027-of-00039.safetensors",
463
+ "model.layers.21.block_sparse_moe.experts.7.w1.weight": "model-00027-of-00039.safetensors",
464
+ "model.layers.21.block_sparse_moe.experts.7.w2.weight": "model-00027-of-00039.safetensors",
465
+ "model.layers.21.block_sparse_moe.experts.7.w3.weight": "model-00027-of-00039.safetensors",
466
+ "model.layers.21.block_sparse_moe.gate.weight": "model-00026-of-00039.safetensors",
467
+ "model.layers.21.input_layernorm.weight": "model-00027-of-00039.safetensors",
468
+ "model.layers.21.post_attention_layernorm.weight": "model-00027-of-00039.safetensors",
469
+ "model.layers.21.self_attn.k_proj.weight": "model-00026-of-00039.safetensors",
470
+ "model.layers.21.self_attn.o_proj.weight": "model-00026-of-00039.safetensors",
471
+ "model.layers.21.self_attn.q_proj.weight": "model-00026-of-00039.safetensors",
472
+ "model.layers.21.self_attn.v_proj.weight": "model-00026-of-00039.safetensors",
473
+ "model.layers.22.block_sparse_moe.experts.0.w1.weight": "model-00027-of-00039.safetensors",
474
+ "model.layers.22.block_sparse_moe.experts.0.w2.weight": "model-00027-of-00039.safetensors",
475
+ "model.layers.22.block_sparse_moe.experts.0.w3.weight": "model-00027-of-00039.safetensors",
476
+ "model.layers.22.block_sparse_moe.experts.1.w1.weight": "model-00027-of-00039.safetensors",
477
+ "model.layers.22.block_sparse_moe.experts.1.w2.weight": "model-00027-of-00039.safetensors",
478
+ "model.layers.22.block_sparse_moe.experts.1.w3.weight": "model-00027-of-00039.safetensors",
479
+ "model.layers.22.block_sparse_moe.experts.2.w1.weight": "model-00027-of-00039.safetensors",
480
+ "model.layers.22.block_sparse_moe.experts.2.w2.weight": "model-00027-of-00039.safetensors",
481
+ "model.layers.22.block_sparse_moe.experts.2.w3.weight": "model-00027-of-00039.safetensors",
482
+ "model.layers.22.block_sparse_moe.experts.3.w1.weight": "model-00027-of-00039.safetensors",
483
+ "model.layers.22.block_sparse_moe.experts.3.w2.weight": "model-00027-of-00039.safetensors",
484
+ "model.layers.22.block_sparse_moe.experts.3.w3.weight": "model-00027-of-00039.safetensors",
485
+ "model.layers.22.block_sparse_moe.experts.4.w1.weight": "model-00027-of-00039.safetensors",
486
+ "model.layers.22.block_sparse_moe.experts.4.w2.weight": "model-00027-of-00039.safetensors",
487
+ "model.layers.22.block_sparse_moe.experts.4.w3.weight": "model-00028-of-00039.safetensors",
488
+ "model.layers.22.block_sparse_moe.experts.5.w1.weight": "model-00028-of-00039.safetensors",
489
+ "model.layers.22.block_sparse_moe.experts.5.w2.weight": "model-00028-of-00039.safetensors",
490
+ "model.layers.22.block_sparse_moe.experts.5.w3.weight": "model-00028-of-00039.safetensors",
491
+ "model.layers.22.block_sparse_moe.experts.6.w1.weight": "model-00028-of-00039.safetensors",
492
+ "model.layers.22.block_sparse_moe.experts.6.w2.weight": "model-00028-of-00039.safetensors",
493
+ "model.layers.22.block_sparse_moe.experts.6.w3.weight": "model-00028-of-00039.safetensors",
494
+ "model.layers.22.block_sparse_moe.experts.7.w1.weight": "model-00028-of-00039.safetensors",
495
+ "model.layers.22.block_sparse_moe.experts.7.w2.weight": "model-00028-of-00039.safetensors",
496
+ "model.layers.22.block_sparse_moe.experts.7.w3.weight": "model-00028-of-00039.safetensors",
497
+ "model.layers.22.block_sparse_moe.gate.weight": "model-00027-of-00039.safetensors",
498
+ "model.layers.22.input_layernorm.weight": "model-00028-of-00039.safetensors",
499
+ "model.layers.22.post_attention_layernorm.weight": "model-00028-of-00039.safetensors",
500
+ "model.layers.22.self_attn.k_proj.weight": "model-00027-of-00039.safetensors",
501
+ "model.layers.22.self_attn.o_proj.weight": "model-00027-of-00039.safetensors",
502
+ "model.layers.22.self_attn.q_proj.weight": "model-00027-of-00039.safetensors",
503
+ "model.layers.22.self_attn.v_proj.weight": "model-00027-of-00039.safetensors",
504
+ "model.layers.23.block_sparse_moe.experts.0.w1.weight": "model-00028-of-00039.safetensors",
505
+ "model.layers.23.block_sparse_moe.experts.0.w2.weight": "model-00028-of-00039.safetensors",
506
+ "model.layers.23.block_sparse_moe.experts.0.w3.weight": "model-00028-of-00039.safetensors",
507
+ "model.layers.23.block_sparse_moe.experts.1.w1.weight": "model-00028-of-00039.safetensors",
508
+ "model.layers.23.block_sparse_moe.experts.1.w2.weight": "model-00028-of-00039.safetensors",
509
+ "model.layers.23.block_sparse_moe.experts.1.w3.weight": "model-00028-of-00039.safetensors",
510
+ "model.layers.23.block_sparse_moe.experts.2.w1.weight": "model-00028-of-00039.safetensors",
511
+ "model.layers.23.block_sparse_moe.experts.2.w2.weight": "model-00028-of-00039.safetensors",
512
+ "model.layers.23.block_sparse_moe.experts.2.w3.weight": "model-00028-of-00039.safetensors",
513
+ "model.layers.23.block_sparse_moe.experts.3.w1.weight": "model-00028-of-00039.safetensors",
514
+ "model.layers.23.block_sparse_moe.experts.3.w2.weight": "model-00029-of-00039.safetensors",
515
+ "model.layers.23.block_sparse_moe.experts.3.w3.weight": "model-00029-of-00039.safetensors",
516
+ "model.layers.23.block_sparse_moe.experts.4.w1.weight": "model-00029-of-00039.safetensors",
517
+ "model.layers.23.block_sparse_moe.experts.4.w2.weight": "model-00029-of-00039.safetensors",
518
+ "model.layers.23.block_sparse_moe.experts.4.w3.weight": "model-00029-of-00039.safetensors",
519
+ "model.layers.23.block_sparse_moe.experts.5.w1.weight": "model-00029-of-00039.safetensors",
520
+ "model.layers.23.block_sparse_moe.experts.5.w2.weight": "model-00029-of-00039.safetensors",
521
+ "model.layers.23.block_sparse_moe.experts.5.w3.weight": "model-00029-of-00039.safetensors",
522
+ "model.layers.23.block_sparse_moe.experts.6.w1.weight": "model-00029-of-00039.safetensors",
523
+ "model.layers.23.block_sparse_moe.experts.6.w2.weight": "model-00029-of-00039.safetensors",
524
+ "model.layers.23.block_sparse_moe.experts.6.w3.weight": "model-00029-of-00039.safetensors",
525
+ "model.layers.23.block_sparse_moe.experts.7.w1.weight": "model-00029-of-00039.safetensors",
526
+ "model.layers.23.block_sparse_moe.experts.7.w2.weight": "model-00029-of-00039.safetensors",
527
+ "model.layers.23.block_sparse_moe.experts.7.w3.weight": "model-00029-of-00039.safetensors",
528
+ "model.layers.23.block_sparse_moe.gate.weight": "model-00028-of-00039.safetensors",
529
+ "model.layers.23.input_layernorm.weight": "model-00029-of-00039.safetensors",
530
+ "model.layers.23.post_attention_layernorm.weight": "model-00029-of-00039.safetensors",
531
+ "model.layers.23.self_attn.k_proj.weight": "model-00028-of-00039.safetensors",
532
+ "model.layers.23.self_attn.o_proj.weight": "model-00028-of-00039.safetensors",
533
+ "model.layers.23.self_attn.q_proj.weight": "model-00028-of-00039.safetensors",
534
+ "model.layers.23.self_attn.v_proj.weight": "model-00028-of-00039.safetensors",
535
+ "model.layers.24.block_sparse_moe.experts.0.w1.weight": "model-00029-of-00039.safetensors",
536
+ "model.layers.24.block_sparse_moe.experts.0.w2.weight": "model-00029-of-00039.safetensors",
537
+ "model.layers.24.block_sparse_moe.experts.0.w3.weight": "model-00029-of-00039.safetensors",
538
+ "model.layers.24.block_sparse_moe.experts.1.w1.weight": "model-00029-of-00039.safetensors",
539
+ "model.layers.24.block_sparse_moe.experts.1.w2.weight": "model-00029-of-00039.safetensors",
540
+ "model.layers.24.block_sparse_moe.experts.1.w3.weight": "model-00029-of-00039.safetensors",
541
+ "model.layers.24.block_sparse_moe.experts.2.w1.weight": "model-00030-of-00039.safetensors",
542
+ "model.layers.24.block_sparse_moe.experts.2.w2.weight": "model-00030-of-00039.safetensors",
543
+ "model.layers.24.block_sparse_moe.experts.2.w3.weight": "model-00030-of-00039.safetensors",
544
+ "model.layers.24.block_sparse_moe.experts.3.w1.weight": "model-00030-of-00039.safetensors",
545
+ "model.layers.24.block_sparse_moe.experts.3.w2.weight": "model-00030-of-00039.safetensors",
546
+ "model.layers.24.block_sparse_moe.experts.3.w3.weight": "model-00030-of-00039.safetensors",
547
+ "model.layers.24.block_sparse_moe.experts.4.w1.weight": "model-00030-of-00039.safetensors",
548
+ "model.layers.24.block_sparse_moe.experts.4.w2.weight": "model-00030-of-00039.safetensors",
549
+ "model.layers.24.block_sparse_moe.experts.4.w3.weight": "model-00030-of-00039.safetensors",
550
+ "model.layers.24.block_sparse_moe.experts.5.w1.weight": "model-00030-of-00039.safetensors",
551
+ "model.layers.24.block_sparse_moe.experts.5.w2.weight": "model-00030-of-00039.safetensors",
552
+ "model.layers.24.block_sparse_moe.experts.5.w3.weight": "model-00030-of-00039.safetensors",
553
+ "model.layers.24.block_sparse_moe.experts.6.w1.weight": "model-00030-of-00039.safetensors",
554
+ "model.layers.24.block_sparse_moe.experts.6.w2.weight": "model-00030-of-00039.safetensors",
555
+ "model.layers.24.block_sparse_moe.experts.6.w3.weight": "model-00030-of-00039.safetensors",
556
+ "model.layers.24.block_sparse_moe.experts.7.w1.weight": "model-00030-of-00039.safetensors",
557
+ "model.layers.24.block_sparse_moe.experts.7.w2.weight": "model-00030-of-00039.safetensors",
558
+ "model.layers.24.block_sparse_moe.experts.7.w3.weight": "model-00030-of-00039.safetensors",
559
+ "model.layers.24.block_sparse_moe.gate.weight": "model-00029-of-00039.safetensors",
560
+ "model.layers.24.input_layernorm.weight": "model-00030-of-00039.safetensors",
561
+ "model.layers.24.post_attention_layernorm.weight": "model-00030-of-00039.safetensors",
562
+ "model.layers.24.self_attn.k_proj.weight": "model-00029-of-00039.safetensors",
563
+ "model.layers.24.self_attn.o_proj.weight": "model-00029-of-00039.safetensors",
564
+ "model.layers.24.self_attn.q_proj.weight": "model-00029-of-00039.safetensors",
565
+ "model.layers.24.self_attn.v_proj.weight": "model-00029-of-00039.safetensors",
566
+ "model.layers.25.block_sparse_moe.experts.0.w1.weight": "model-00030-of-00039.safetensors",
567
+ "model.layers.25.block_sparse_moe.experts.0.w2.weight": "model-00030-of-00039.safetensors",
568
+ "model.layers.25.block_sparse_moe.experts.0.w3.weight": "model-00031-of-00039.safetensors",
569
+ "model.layers.25.block_sparse_moe.experts.1.w1.weight": "model-00031-of-00039.safetensors",
570
+ "model.layers.25.block_sparse_moe.experts.1.w2.weight": "model-00031-of-00039.safetensors",
571
+ "model.layers.25.block_sparse_moe.experts.1.w3.weight": "model-00031-of-00039.safetensors",
572
+ "model.layers.25.block_sparse_moe.experts.2.w1.weight": "model-00031-of-00039.safetensors",
573
+ "model.layers.25.block_sparse_moe.experts.2.w2.weight": "model-00031-of-00039.safetensors",
574
+ "model.layers.25.block_sparse_moe.experts.2.w3.weight": "model-00031-of-00039.safetensors",
575
+ "model.layers.25.block_sparse_moe.experts.3.w1.weight": "model-00031-of-00039.safetensors",
576
+ "model.layers.25.block_sparse_moe.experts.3.w2.weight": "model-00031-of-00039.safetensors",
577
+ "model.layers.25.block_sparse_moe.experts.3.w3.weight": "model-00031-of-00039.safetensors",
578
+ "model.layers.25.block_sparse_moe.experts.4.w1.weight": "model-00031-of-00039.safetensors",
579
+ "model.layers.25.block_sparse_moe.experts.4.w2.weight": "model-00031-of-00039.safetensors",
580
+ "model.layers.25.block_sparse_moe.experts.4.w3.weight": "model-00031-of-00039.safetensors",
581
+ "model.layers.25.block_sparse_moe.experts.5.w1.weight": "model-00031-of-00039.safetensors",
582
+ "model.layers.25.block_sparse_moe.experts.5.w2.weight": "model-00031-of-00039.safetensors",
583
+ "model.layers.25.block_sparse_moe.experts.5.w3.weight": "model-00031-of-00039.safetensors",
584
+ "model.layers.25.block_sparse_moe.experts.6.w1.weight": "model-00031-of-00039.safetensors",
585
+ "model.layers.25.block_sparse_moe.experts.6.w2.weight": "model-00031-of-00039.safetensors",
586
+ "model.layers.25.block_sparse_moe.experts.6.w3.weight": "model-00031-of-00039.safetensors",
587
+ "model.layers.25.block_sparse_moe.experts.7.w1.weight": "model-00031-of-00039.safetensors",
588
+ "model.layers.25.block_sparse_moe.experts.7.w2.weight": "model-00031-of-00039.safetensors",
589
+ "model.layers.25.block_sparse_moe.experts.7.w3.weight": "model-00032-of-00039.safetensors",
590
+ "model.layers.25.block_sparse_moe.gate.weight": "model-00030-of-00039.safetensors",
591
+ "model.layers.25.input_layernorm.weight": "model-00032-of-00039.safetensors",
592
+ "model.layers.25.post_attention_layernorm.weight": "model-00032-of-00039.safetensors",
593
+ "model.layers.25.self_attn.k_proj.weight": "model-00030-of-00039.safetensors",
594
+ "model.layers.25.self_attn.o_proj.weight": "model-00030-of-00039.safetensors",
595
+ "model.layers.25.self_attn.q_proj.weight": "model-00030-of-00039.safetensors",
596
+ "model.layers.25.self_attn.v_proj.weight": "model-00030-of-00039.safetensors",
597
+ "model.layers.26.block_sparse_moe.experts.0.w1.weight": "model-00032-of-00039.safetensors",
598
+ "model.layers.26.block_sparse_moe.experts.0.w2.weight": "model-00032-of-00039.safetensors",
599
+ "model.layers.26.block_sparse_moe.experts.0.w3.weight": "model-00032-of-00039.safetensors",
600
+ "model.layers.26.block_sparse_moe.experts.1.w1.weight": "model-00032-of-00039.safetensors",
601
+ "model.layers.26.block_sparse_moe.experts.1.w2.weight": "model-00032-of-00039.safetensors",
602
+ "model.layers.26.block_sparse_moe.experts.1.w3.weight": "model-00032-of-00039.safetensors",
603
+ "model.layers.26.block_sparse_moe.experts.2.w1.weight": "model-00032-of-00039.safetensors",
604
+ "model.layers.26.block_sparse_moe.experts.2.w2.weight": "model-00032-of-00039.safetensors",
605
+ "model.layers.26.block_sparse_moe.experts.2.w3.weight": "model-00032-of-00039.safetensors",
606
+ "model.layers.26.block_sparse_moe.experts.3.w1.weight": "model-00032-of-00039.safetensors",
607
+ "model.layers.26.block_sparse_moe.experts.3.w2.weight": "model-00032-of-00039.safetensors",
608
+ "model.layers.26.block_sparse_moe.experts.3.w3.weight": "model-00032-of-00039.safetensors",
609
+ "model.layers.26.block_sparse_moe.experts.4.w1.weight": "model-00032-of-00039.safetensors",
610
+ "model.layers.26.block_sparse_moe.experts.4.w2.weight": "model-00032-of-00039.safetensors",
611
+ "model.layers.26.block_sparse_moe.experts.4.w3.weight": "model-00032-of-00039.safetensors",
612
+ "model.layers.26.block_sparse_moe.experts.5.w1.weight": "model-00032-of-00039.safetensors",
613
+ "model.layers.26.block_sparse_moe.experts.5.w2.weight": "model-00032-of-00039.safetensors",
614
+ "model.layers.26.block_sparse_moe.experts.5.w3.weight": "model-00032-of-00039.safetensors",
615
+ "model.layers.26.block_sparse_moe.experts.6.w1.weight": "model-00032-of-00039.safetensors",
616
+ "model.layers.26.block_sparse_moe.experts.6.w2.weight": "model-00033-of-00039.safetensors",
617
+ "model.layers.26.block_sparse_moe.experts.6.w3.weight": "model-00033-of-00039.safetensors",
618
+ "model.layers.26.block_sparse_moe.experts.7.w1.weight": "model-00033-of-00039.safetensors",
619
+ "model.layers.26.block_sparse_moe.experts.7.w2.weight": "model-00033-of-00039.safetensors",
620
+ "model.layers.26.block_sparse_moe.experts.7.w3.weight": "model-00033-of-00039.safetensors",
621
+ "model.layers.26.block_sparse_moe.gate.weight": "model-00032-of-00039.safetensors",
622
+ "model.layers.26.input_layernorm.weight": "model-00033-of-00039.safetensors",
623
+ "model.layers.26.post_attention_layernorm.weight": "model-00033-of-00039.safetensors",
624
+ "model.layers.26.self_attn.k_proj.weight": "model-00032-of-00039.safetensors",
625
+ "model.layers.26.self_attn.o_proj.weight": "model-00032-of-00039.safetensors",
626
+ "model.layers.26.self_attn.q_proj.weight": "model-00032-of-00039.safetensors",
627
+ "model.layers.26.self_attn.v_proj.weight": "model-00032-of-00039.safetensors",
628
+ "model.layers.27.block_sparse_moe.experts.0.w1.weight": "model-00033-of-00039.safetensors",
629
+ "model.layers.27.block_sparse_moe.experts.0.w2.weight": "model-00033-of-00039.safetensors",
630
+ "model.layers.27.block_sparse_moe.experts.0.w3.weight": "model-00033-of-00039.safetensors",
631
+ "model.layers.27.block_sparse_moe.experts.1.w1.weight": "model-00033-of-00039.safetensors",
632
+ "model.layers.27.block_sparse_moe.experts.1.w2.weight": "model-00033-of-00039.safetensors",
633
+ "model.layers.27.block_sparse_moe.experts.1.w3.weight": "model-00033-of-00039.safetensors",
634
+ "model.layers.27.block_sparse_moe.experts.2.w1.weight": "model-00033-of-00039.safetensors",
635
+ "model.layers.27.block_sparse_moe.experts.2.w2.weight": "model-00033-of-00039.safetensors",
636
+ "model.layers.27.block_sparse_moe.experts.2.w3.weight": "model-00033-of-00039.safetensors",
637
+ "model.layers.27.block_sparse_moe.experts.3.w1.weight": "model-00033-of-00039.safetensors",
638
+ "model.layers.27.block_sparse_moe.experts.3.w2.weight": "model-00033-of-00039.safetensors",
639
+ "model.layers.27.block_sparse_moe.experts.3.w3.weight": "model-00033-of-00039.safetensors",
640
+ "model.layers.27.block_sparse_moe.experts.4.w1.weight": "model-00033-of-00039.safetensors",
641
+ "model.layers.27.block_sparse_moe.experts.4.w2.weight": "model-00033-of-00039.safetensors",
642
+ "model.layers.27.block_sparse_moe.experts.4.w3.weight": "model-00033-of-00039.safetensors",
643
+ "model.layers.27.block_sparse_moe.experts.5.w1.weight": "model-00034-of-00039.safetensors",
644
+ "model.layers.27.block_sparse_moe.experts.5.w2.weight": "model-00034-of-00039.safetensors",
645
+ "model.layers.27.block_sparse_moe.experts.5.w3.weight": "model-00034-of-00039.safetensors",
646
+ "model.layers.27.block_sparse_moe.experts.6.w1.weight": "model-00034-of-00039.safetensors",
647
+ "model.layers.27.block_sparse_moe.experts.6.w2.weight": "model-00034-of-00039.safetensors",
648
+ "model.layers.27.block_sparse_moe.experts.6.w3.weight": "model-00034-of-00039.safetensors",
649
+ "model.layers.27.block_sparse_moe.experts.7.w1.weight": "model-00034-of-00039.safetensors",
650
+ "model.layers.27.block_sparse_moe.experts.7.w2.weight": "model-00034-of-00039.safetensors",
651
+ "model.layers.27.block_sparse_moe.experts.7.w3.weight": "model-00034-of-00039.safetensors",
652
+ "model.layers.27.block_sparse_moe.gate.weight": "model-00033-of-00039.safetensors",
653
+ "model.layers.27.input_layernorm.weight": "model-00034-of-00039.safetensors",
654
+ "model.layers.27.post_attention_layernorm.weight": "model-00034-of-00039.safetensors",
655
+ "model.layers.27.self_attn.k_proj.weight": "model-00033-of-00039.safetensors",
656
+ "model.layers.27.self_attn.o_proj.weight": "model-00033-of-00039.safetensors",
657
+ "model.layers.27.self_attn.q_proj.weight": "model-00033-of-00039.safetensors",
658
+ "model.layers.27.self_attn.v_proj.weight": "model-00033-of-00039.safetensors",
659
+ "model.layers.28.block_sparse_moe.experts.0.w1.weight": "model-00034-of-00039.safetensors",
660
+ "model.layers.28.block_sparse_moe.experts.0.w2.weight": "model-00034-of-00039.safetensors",
661
+ "model.layers.28.block_sparse_moe.experts.0.w3.weight": "model-00034-of-00039.safetensors",
662
+ "model.layers.28.block_sparse_moe.experts.1.w1.weight": "model-00034-of-00039.safetensors",
663
+ "model.layers.28.block_sparse_moe.experts.1.w2.weight": "model-00034-of-00039.safetensors",
664
+ "model.layers.28.block_sparse_moe.experts.1.w3.weight": "model-00034-of-00039.safetensors",
665
+ "model.layers.28.block_sparse_moe.experts.2.w1.weight": "model-00034-of-00039.safetensors",
666
+ "model.layers.28.block_sparse_moe.experts.2.w2.weight": "model-00034-of-00039.safetensors",
667
+ "model.layers.28.block_sparse_moe.experts.2.w3.weight": "model-00034-of-00039.safetensors",
668
+ "model.layers.28.block_sparse_moe.experts.3.w1.weight": "model-00034-of-00039.safetensors",
669
+ "model.layers.28.block_sparse_moe.experts.3.w2.weight": "model-00034-of-00039.safetensors",
670
+ "model.layers.28.block_sparse_moe.experts.3.w3.weight": "model-00035-of-00039.safetensors",
671
+ "model.layers.28.block_sparse_moe.experts.4.w1.weight": "model-00035-of-00039.safetensors",
672
+ "model.layers.28.block_sparse_moe.experts.4.w2.weight": "model-00035-of-00039.safetensors",
673
+ "model.layers.28.block_sparse_moe.experts.4.w3.weight": "model-00035-of-00039.safetensors",
674
+ "model.layers.28.block_sparse_moe.experts.5.w1.weight": "model-00035-of-00039.safetensors",
675
+ "model.layers.28.block_sparse_moe.experts.5.w2.weight": "model-00035-of-00039.safetensors",
676
+ "model.layers.28.block_sparse_moe.experts.5.w3.weight": "model-00035-of-00039.safetensors",
677
+ "model.layers.28.block_sparse_moe.experts.6.w1.weight": "model-00035-of-00039.safetensors",
678
+ "model.layers.28.block_sparse_moe.experts.6.w2.weight": "model-00035-of-00039.safetensors",
679
+ "model.layers.28.block_sparse_moe.experts.6.w3.weight": "model-00035-of-00039.safetensors",
680
+ "model.layers.28.block_sparse_moe.experts.7.w1.weight": "model-00035-of-00039.safetensors",
681
+ "model.layers.28.block_sparse_moe.experts.7.w2.weight": "model-00035-of-00039.safetensors",
682
+ "model.layers.28.block_sparse_moe.experts.7.w3.weight": "model-00035-of-00039.safetensors",
683
+ "model.layers.28.block_sparse_moe.gate.weight": "model-00034-of-00039.safetensors",
684
+ "model.layers.28.input_layernorm.weight": "model-00035-of-00039.safetensors",
685
+ "model.layers.28.post_attention_layernorm.weight": "model-00035-of-00039.safetensors",
686
+ "model.layers.28.self_attn.k_proj.weight": "model-00034-of-00039.safetensors",
687
+ "model.layers.28.self_attn.o_proj.weight": "model-00034-of-00039.safetensors",
688
+ "model.layers.28.self_attn.q_proj.weight": "model-00034-of-00039.safetensors",
689
+ "model.layers.28.self_attn.v_proj.weight": "model-00034-of-00039.safetensors",
690
+ "model.layers.29.block_sparse_moe.experts.0.w1.weight": "model-00035-of-00039.safetensors",
691
+ "model.layers.29.block_sparse_moe.experts.0.w2.weight": "model-00035-of-00039.safetensors",
692
+ "model.layers.29.block_sparse_moe.experts.0.w3.weight": "model-00035-of-00039.safetensors",
693
+ "model.layers.29.block_sparse_moe.experts.1.w1.weight": "model-00035-of-00039.safetensors",
694
+ "model.layers.29.block_sparse_moe.experts.1.w2.weight": "model-00035-of-00039.safetensors",
695
+ "model.layers.29.block_sparse_moe.experts.1.w3.weight": "model-00035-of-00039.safetensors",
696
+ "model.layers.29.block_sparse_moe.experts.2.w1.weight": "model-00035-of-00039.safetensors",
697
+ "model.layers.29.block_sparse_moe.experts.2.w2.weight": "model-00036-of-00039.safetensors",
698
+ "model.layers.29.block_sparse_moe.experts.2.w3.weight": "model-00036-of-00039.safetensors",
699
+ "model.layers.29.block_sparse_moe.experts.3.w1.weight": "model-00036-of-00039.safetensors",
700
+ "model.layers.29.block_sparse_moe.experts.3.w2.weight": "model-00036-of-00039.safetensors",
701
+ "model.layers.29.block_sparse_moe.experts.3.w3.weight": "model-00036-of-00039.safetensors",
702
+ "model.layers.29.block_sparse_moe.experts.4.w1.weight": "model-00036-of-00039.safetensors",
703
+ "model.layers.29.block_sparse_moe.experts.4.w2.weight": "model-00036-of-00039.safetensors",
704
+ "model.layers.29.block_sparse_moe.experts.4.w3.weight": "model-00036-of-00039.safetensors",
705
+ "model.layers.29.block_sparse_moe.experts.5.w1.weight": "model-00036-of-00039.safetensors",
706
+ "model.layers.29.block_sparse_moe.experts.5.w2.weight": "model-00036-of-00039.safetensors",
707
+ "model.layers.29.block_sparse_moe.experts.5.w3.weight": "model-00036-of-00039.safetensors",
708
+ "model.layers.29.block_sparse_moe.experts.6.w1.weight": "model-00036-of-00039.safetensors",
709
+ "model.layers.29.block_sparse_moe.experts.6.w2.weight": "model-00036-of-00039.safetensors",
710
+ "model.layers.29.block_sparse_moe.experts.6.w3.weight": "model-00036-of-00039.safetensors",
711
+ "model.layers.29.block_sparse_moe.experts.7.w1.weight": "model-00036-of-00039.safetensors",
712
+ "model.layers.29.block_sparse_moe.experts.7.w2.weight": "model-00036-of-00039.safetensors",
713
+ "model.layers.29.block_sparse_moe.experts.7.w3.weight": "model-00036-of-00039.safetensors",
714
+ "model.layers.29.block_sparse_moe.gate.weight": "model-00035-of-00039.safetensors",
715
+ "model.layers.29.input_layernorm.weight": "model-00036-of-00039.safetensors",
716
+ "model.layers.29.post_attention_layernorm.weight": "model-00036-of-00039.safetensors",
717
+ "model.layers.29.self_attn.k_proj.weight": "model-00035-of-00039.safetensors",
718
+ "model.layers.29.self_attn.o_proj.weight": "model-00035-of-00039.safetensors",
719
+ "model.layers.29.self_attn.q_proj.weight": "model-00035-of-00039.safetensors",
720
+ "model.layers.29.self_attn.v_proj.weight": "model-00035-of-00039.safetensors",
721
+ "model.layers.3.block_sparse_moe.experts.0.w1.weight": "model-00004-of-00039.safetensors",
722
+ "model.layers.3.block_sparse_moe.experts.0.w2.weight": "model-00004-of-00039.safetensors",
723
+ "model.layers.3.block_sparse_moe.experts.0.w3.weight": "model-00004-of-00039.safetensors",
724
+ "model.layers.3.block_sparse_moe.experts.1.w1.weight": "model-00004-of-00039.safetensors",
725
+ "model.layers.3.block_sparse_moe.experts.1.w2.weight": "model-00004-of-00039.safetensors",
726
+ "model.layers.3.block_sparse_moe.experts.1.w3.weight": "model-00004-of-00039.safetensors",
727
+ "model.layers.3.block_sparse_moe.experts.2.w1.weight": "model-00005-of-00039.safetensors",
728
+ "model.layers.3.block_sparse_moe.experts.2.w2.weight": "model-00005-of-00039.safetensors",
729
+ "model.layers.3.block_sparse_moe.experts.2.w3.weight": "model-00005-of-00039.safetensors",
730
+ "model.layers.3.block_sparse_moe.experts.3.w1.weight": "model-00005-of-00039.safetensors",
731
+ "model.layers.3.block_sparse_moe.experts.3.w2.weight": "model-00005-of-00039.safetensors",
732
+ "model.layers.3.block_sparse_moe.experts.3.w3.weight": "model-00005-of-00039.safetensors",
733
+ "model.layers.3.block_sparse_moe.experts.4.w1.weight": "model-00005-of-00039.safetensors",
734
+ "model.layers.3.block_sparse_moe.experts.4.w2.weight": "model-00005-of-00039.safetensors",
735
+ "model.layers.3.block_sparse_moe.experts.4.w3.weight": "model-00005-of-00039.safetensors",
736
+ "model.layers.3.block_sparse_moe.experts.5.w1.weight": "model-00005-of-00039.safetensors",
737
+ "model.layers.3.block_sparse_moe.experts.5.w2.weight": "model-00005-of-00039.safetensors",
738
+ "model.layers.3.block_sparse_moe.experts.5.w3.weight": "model-00005-of-00039.safetensors",
739
+ "model.layers.3.block_sparse_moe.experts.6.w1.weight": "model-00005-of-00039.safetensors",
740
+ "model.layers.3.block_sparse_moe.experts.6.w2.weight": "model-00005-of-00039.safetensors",
741
+ "model.layers.3.block_sparse_moe.experts.6.w3.weight": "model-00005-of-00039.safetensors",
742
+ "model.layers.3.block_sparse_moe.experts.7.w1.weight": "model-00005-of-00039.safetensors",
743
+ "model.layers.3.block_sparse_moe.experts.7.w2.weight": "model-00005-of-00039.safetensors",
744
+ "model.layers.3.block_sparse_moe.experts.7.w3.weight": "model-00005-of-00039.safetensors",
745
+ "model.layers.3.block_sparse_moe.gate.weight": "model-00004-of-00039.safetensors",
746
+ "model.layers.3.input_layernorm.weight": "model-00005-of-00039.safetensors",
747
+ "model.layers.3.post_attention_layernorm.weight": "model-00005-of-00039.safetensors",
748
+ "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00039.safetensors",
749
+ "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00039.safetensors",
750
+ "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00039.safetensors",
751
+ "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00039.safetensors",
752
+ "model.layers.30.block_sparse_moe.experts.0.w1.weight": "model-00036-of-00039.safetensors",
753
+ "model.layers.30.block_sparse_moe.experts.0.w2.weight": "model-00036-of-00039.safetensors",
754
+ "model.layers.30.block_sparse_moe.experts.0.w3.weight": "model-00036-of-00039.safetensors",
755
+ "model.layers.30.block_sparse_moe.experts.1.w1.weight": "model-00037-of-00039.safetensors",
756
+ "model.layers.30.block_sparse_moe.experts.1.w2.weight": "model-00037-of-00039.safetensors",
757
+ "model.layers.30.block_sparse_moe.experts.1.w3.weight": "model-00037-of-00039.safetensors",
758
+ "model.layers.30.block_sparse_moe.experts.2.w1.weight": "model-00037-of-00039.safetensors",
759
+ "model.layers.30.block_sparse_moe.experts.2.w2.weight": "model-00037-of-00039.safetensors",
760
+ "model.layers.30.block_sparse_moe.experts.2.w3.weight": "model-00037-of-00039.safetensors",
761
+ "model.layers.30.block_sparse_moe.experts.3.w1.weight": "model-00037-of-00039.safetensors",
762
+ "model.layers.30.block_sparse_moe.experts.3.w2.weight": "model-00037-of-00039.safetensors",
763
+ "model.layers.30.block_sparse_moe.experts.3.w3.weight": "model-00037-of-00039.safetensors",
764
+ "model.layers.30.block_sparse_moe.experts.4.w1.weight": "model-00037-of-00039.safetensors",
765
+ "model.layers.30.block_sparse_moe.experts.4.w2.weight": "model-00037-of-00039.safetensors",
766
+ "model.layers.30.block_sparse_moe.experts.4.w3.weight": "model-00037-of-00039.safetensors",
767
+ "model.layers.30.block_sparse_moe.experts.5.w1.weight": "model-00037-of-00039.safetensors",
768
+ "model.layers.30.block_sparse_moe.experts.5.w2.weight": "model-00037-of-00039.safetensors",
769
+ "model.layers.30.block_sparse_moe.experts.5.w3.weight": "model-00037-of-00039.safetensors",
770
+ "model.layers.30.block_sparse_moe.experts.6.w1.weight": "model-00037-of-00039.safetensors",
771
+ "model.layers.30.block_sparse_moe.experts.6.w2.weight": "model-00037-of-00039.safetensors",
772
+ "model.layers.30.block_sparse_moe.experts.6.w3.weight": "model-00037-of-00039.safetensors",
773
+ "model.layers.30.block_sparse_moe.experts.7.w1.weight": "model-00037-of-00039.safetensors",
774
+ "model.layers.30.block_sparse_moe.experts.7.w2.weight": "model-00037-of-00039.safetensors",
775
+ "model.layers.30.block_sparse_moe.experts.7.w3.weight": "model-00037-of-00039.safetensors",
776
+ "model.layers.30.block_sparse_moe.gate.weight": "model-00036-of-00039.safetensors",
777
+ "model.layers.30.input_layernorm.weight": "model-00037-of-00039.safetensors",
778
+ "model.layers.30.post_attention_layernorm.weight": "model-00037-of-00039.safetensors",
779
+ "model.layers.30.self_attn.k_proj.weight": "model-00036-of-00039.safetensors",
780
+ "model.layers.30.self_attn.o_proj.weight": "model-00036-of-00039.safetensors",
781
+ "model.layers.30.self_attn.q_proj.weight": "model-00036-of-00039.safetensors",
782
+ "model.layers.30.self_attn.v_proj.weight": "model-00036-of-00039.safetensors",
783
+ "model.layers.31.block_sparse_moe.experts.0.w1.weight": "model-00038-of-00039.safetensors",
784
+ "model.layers.31.block_sparse_moe.experts.0.w2.weight": "model-00038-of-00039.safetensors",
785
+ "model.layers.31.block_sparse_moe.experts.0.w3.weight": "model-00038-of-00039.safetensors",
786
+ "model.layers.31.block_sparse_moe.experts.1.w1.weight": "model-00038-of-00039.safetensors",
787
+ "model.layers.31.block_sparse_moe.experts.1.w2.weight": "model-00038-of-00039.safetensors",
788
+ "model.layers.31.block_sparse_moe.experts.1.w3.weight": "model-00038-of-00039.safetensors",
789
+ "model.layers.31.block_sparse_moe.experts.2.w1.weight": "model-00038-of-00039.safetensors",
790
+ "model.layers.31.block_sparse_moe.experts.2.w2.weight": "model-00038-of-00039.safetensors",
791
+ "model.layers.31.block_sparse_moe.experts.2.w3.weight": "model-00038-of-00039.safetensors",
792
+ "model.layers.31.block_sparse_moe.experts.3.w1.weight": "model-00038-of-00039.safetensors",
793
+ "model.layers.31.block_sparse_moe.experts.3.w2.weight": "model-00038-of-00039.safetensors",
794
+ "model.layers.31.block_sparse_moe.experts.3.w3.weight": "model-00038-of-00039.safetensors",
795
+ "model.layers.31.block_sparse_moe.experts.4.w1.weight": "model-00038-of-00039.safetensors",
796
+ "model.layers.31.block_sparse_moe.experts.4.w2.weight": "model-00038-of-00039.safetensors",
797
+ "model.layers.31.block_sparse_moe.experts.4.w3.weight": "model-00038-of-00039.safetensors",
798
+ "model.layers.31.block_sparse_moe.experts.5.w1.weight": "model-00038-of-00039.safetensors",
799
+ "model.layers.31.block_sparse_moe.experts.5.w2.weight": "model-00038-of-00039.safetensors",
800
+ "model.layers.31.block_sparse_moe.experts.5.w3.weight": "model-00038-of-00039.safetensors",
801
+ "model.layers.31.block_sparse_moe.experts.6.w1.weight": "model-00038-of-00039.safetensors",
802
+ "model.layers.31.block_sparse_moe.experts.6.w2.weight": "model-00038-of-00039.safetensors",
803
+ "model.layers.31.block_sparse_moe.experts.6.w3.weight": "model-00039-of-00039.safetensors",
804
+ "model.layers.31.block_sparse_moe.experts.7.w1.weight": "model-00039-of-00039.safetensors",
805
+ "model.layers.31.block_sparse_moe.experts.7.w2.weight": "model-00039-of-00039.safetensors",
806
+ "model.layers.31.block_sparse_moe.experts.7.w3.weight": "model-00039-of-00039.safetensors",
807
+ "model.layers.31.block_sparse_moe.gate.weight": "model-00038-of-00039.safetensors",
808
+ "model.layers.31.input_layernorm.weight": "model-00039-of-00039.safetensors",
809
+ "model.layers.31.post_attention_layernorm.weight": "model-00039-of-00039.safetensors",
810
+ "model.layers.31.self_attn.k_proj.weight": "model-00038-of-00039.safetensors",
811
+ "model.layers.31.self_attn.o_proj.weight": "model-00038-of-00039.safetensors",
812
+ "model.layers.31.self_attn.q_proj.weight": "model-00037-of-00039.safetensors",
813
+ "model.layers.31.self_attn.v_proj.weight": "model-00038-of-00039.safetensors",
814
+ "model.layers.4.block_sparse_moe.experts.0.w1.weight": "model-00005-of-00039.safetensors",
815
+ "model.layers.4.block_sparse_moe.experts.0.w2.weight": "model-00005-of-00039.safetensors",
816
+ "model.layers.4.block_sparse_moe.experts.0.w3.weight": "model-00006-of-00039.safetensors",
817
+ "model.layers.4.block_sparse_moe.experts.1.w1.weight": "model-00006-of-00039.safetensors",
818
+ "model.layers.4.block_sparse_moe.experts.1.w2.weight": "model-00006-of-00039.safetensors",
819
+ "model.layers.4.block_sparse_moe.experts.1.w3.weight": "model-00006-of-00039.safetensors",
820
+ "model.layers.4.block_sparse_moe.experts.2.w1.weight": "model-00006-of-00039.safetensors",
821
+ "model.layers.4.block_sparse_moe.experts.2.w2.weight": "model-00006-of-00039.safetensors",
822
+ "model.layers.4.block_sparse_moe.experts.2.w3.weight": "model-00006-of-00039.safetensors",
823
+ "model.layers.4.block_sparse_moe.experts.3.w1.weight": "model-00006-of-00039.safetensors",
824
+ "model.layers.4.block_sparse_moe.experts.3.w2.weight": "model-00006-of-00039.safetensors",
825
+ "model.layers.4.block_sparse_moe.experts.3.w3.weight": "model-00006-of-00039.safetensors",
826
+ "model.layers.4.block_sparse_moe.experts.4.w1.weight": "model-00006-of-00039.safetensors",
827
+ "model.layers.4.block_sparse_moe.experts.4.w2.weight": "model-00006-of-00039.safetensors",
828
+ "model.layers.4.block_sparse_moe.experts.4.w3.weight": "model-00006-of-00039.safetensors",
829
+ "model.layers.4.block_sparse_moe.experts.5.w1.weight": "model-00006-of-00039.safetensors",
830
+ "model.layers.4.block_sparse_moe.experts.5.w2.weight": "model-00006-of-00039.safetensors",
831
+ "model.layers.4.block_sparse_moe.experts.5.w3.weight": "model-00006-of-00039.safetensors",
832
+ "model.layers.4.block_sparse_moe.experts.6.w1.weight": "model-00006-of-00039.safetensors",
833
+ "model.layers.4.block_sparse_moe.experts.6.w2.weight": "model-00006-of-00039.safetensors",
834
+ "model.layers.4.block_sparse_moe.experts.6.w3.weight": "model-00006-of-00039.safetensors",
835
+ "model.layers.4.block_sparse_moe.experts.7.w1.weight": "model-00006-of-00039.safetensors",
836
+ "model.layers.4.block_sparse_moe.experts.7.w2.weight": "model-00006-of-00039.safetensors",
837
+ "model.layers.4.block_sparse_moe.experts.7.w3.weight": "model-00007-of-00039.safetensors",
838
+ "model.layers.4.block_sparse_moe.gate.weight": "model-00005-of-00039.safetensors",
839
+ "model.layers.4.input_layernorm.weight": "model-00007-of-00039.safetensors",
840
+ "model.layers.4.post_attention_layernorm.weight": "model-00007-of-00039.safetensors",
841
+ "model.layers.4.self_attn.k_proj.weight": "model-00005-of-00039.safetensors",
842
+ "model.layers.4.self_attn.o_proj.weight": "model-00005-of-00039.safetensors",
843
+ "model.layers.4.self_attn.q_proj.weight": "model-00005-of-00039.safetensors",
844
+ "model.layers.4.self_attn.v_proj.weight": "model-00005-of-00039.safetensors",
845
+ "model.layers.5.block_sparse_moe.experts.0.w1.weight": "model-00007-of-00039.safetensors",
846
+ "model.layers.5.block_sparse_moe.experts.0.w2.weight": "model-00007-of-00039.safetensors",
847
+ "model.layers.5.block_sparse_moe.experts.0.w3.weight": "model-00007-of-00039.safetensors",
848
+ "model.layers.5.block_sparse_moe.experts.1.w1.weight": "model-00007-of-00039.safetensors",
849
+ "model.layers.5.block_sparse_moe.experts.1.w2.weight": "model-00007-of-00039.safetensors",
850
+ "model.layers.5.block_sparse_moe.experts.1.w3.weight": "model-00007-of-00039.safetensors",
851
+ "model.layers.5.block_sparse_moe.experts.2.w1.weight": "model-00007-of-00039.safetensors",
852
+ "model.layers.5.block_sparse_moe.experts.2.w2.weight": "model-00007-of-00039.safetensors",
853
+ "model.layers.5.block_sparse_moe.experts.2.w3.weight": "model-00007-of-00039.safetensors",
854
+ "model.layers.5.block_sparse_moe.experts.3.w1.weight": "model-00007-of-00039.safetensors",
855
+ "model.layers.5.block_sparse_moe.experts.3.w2.weight": "model-00007-of-00039.safetensors",
856
+ "model.layers.5.block_sparse_moe.experts.3.w3.weight": "model-00007-of-00039.safetensors",
857
+ "model.layers.5.block_sparse_moe.experts.4.w1.weight": "model-00007-of-00039.safetensors",
858
+ "model.layers.5.block_sparse_moe.experts.4.w2.weight": "model-00007-of-00039.safetensors",
859
+ "model.layers.5.block_sparse_moe.experts.4.w3.weight": "model-00007-of-00039.safetensors",
860
+ "model.layers.5.block_sparse_moe.experts.5.w1.weight": "model-00007-of-00039.safetensors",
861
+ "model.layers.5.block_sparse_moe.experts.5.w2.weight": "model-00007-of-00039.safetensors",
862
+ "model.layers.5.block_sparse_moe.experts.5.w3.weight": "model-00007-of-00039.safetensors",
863
+ "model.layers.5.block_sparse_moe.experts.6.w1.weight": "model-00007-of-00039.safetensors",
864
+ "model.layers.5.block_sparse_moe.experts.6.w2.weight": "model-00008-of-00039.safetensors",
865
+ "model.layers.5.block_sparse_moe.experts.6.w3.weight": "model-00008-of-00039.safetensors",
866
+ "model.layers.5.block_sparse_moe.experts.7.w1.weight": "model-00008-of-00039.safetensors",
867
+ "model.layers.5.block_sparse_moe.experts.7.w2.weight": "model-00008-of-00039.safetensors",
868
+ "model.layers.5.block_sparse_moe.experts.7.w3.weight": "model-00008-of-00039.safetensors",
869
+ "model.layers.5.block_sparse_moe.gate.weight": "model-00007-of-00039.safetensors",
870
+ "model.layers.5.input_layernorm.weight": "model-00008-of-00039.safetensors",
871
+ "model.layers.5.post_attention_layernorm.weight": "model-00008-of-00039.safetensors",
872
+ "model.layers.5.self_attn.k_proj.weight": "model-00007-of-00039.safetensors",
873
+ "model.layers.5.self_attn.o_proj.weight": "model-00007-of-00039.safetensors",
874
+ "model.layers.5.self_attn.q_proj.weight": "model-00007-of-00039.safetensors",
875
+ "model.layers.5.self_attn.v_proj.weight": "model-00007-of-00039.safetensors",
876
+ "model.layers.6.block_sparse_moe.experts.0.w1.weight": "model-00008-of-00039.safetensors",
877
+ "model.layers.6.block_sparse_moe.experts.0.w2.weight": "model-00008-of-00039.safetensors",
878
+ "model.layers.6.block_sparse_moe.experts.0.w3.weight": "model-00008-of-00039.safetensors",
879
+ "model.layers.6.block_sparse_moe.experts.1.w1.weight": "model-00008-of-00039.safetensors",
880
+ "model.layers.6.block_sparse_moe.experts.1.w2.weight": "model-00008-of-00039.safetensors",
881
+ "model.layers.6.block_sparse_moe.experts.1.w3.weight": "model-00008-of-00039.safetensors",
882
+ "model.layers.6.block_sparse_moe.experts.2.w1.weight": "model-00008-of-00039.safetensors",
883
+ "model.layers.6.block_sparse_moe.experts.2.w2.weight": "model-00008-of-00039.safetensors",
884
+ "model.layers.6.block_sparse_moe.experts.2.w3.weight": "model-00008-of-00039.safetensors",
885
+ "model.layers.6.block_sparse_moe.experts.3.w1.weight": "model-00008-of-00039.safetensors",
886
+ "model.layers.6.block_sparse_moe.experts.3.w2.weight": "model-00008-of-00039.safetensors",
887
+ "model.layers.6.block_sparse_moe.experts.3.w3.weight": "model-00008-of-00039.safetensors",
888
+ "model.layers.6.block_sparse_moe.experts.4.w1.weight": "model-00008-of-00039.safetensors",
889
+ "model.layers.6.block_sparse_moe.experts.4.w2.weight": "model-00008-of-00039.safetensors",
890
+ "model.layers.6.block_sparse_moe.experts.4.w3.weight": "model-00008-of-00039.safetensors",
891
+ "model.layers.6.block_sparse_moe.experts.5.w1.weight": "model-00009-of-00039.safetensors",
892
+ "model.layers.6.block_sparse_moe.experts.5.w2.weight": "model-00009-of-00039.safetensors",
893
+ "model.layers.6.block_sparse_moe.experts.5.w3.weight": "model-00009-of-00039.safetensors",
894
+ "model.layers.6.block_sparse_moe.experts.6.w1.weight": "model-00009-of-00039.safetensors",
895
+ "model.layers.6.block_sparse_moe.experts.6.w2.weight": "model-00009-of-00039.safetensors",
896
+ "model.layers.6.block_sparse_moe.experts.6.w3.weight": "model-00009-of-00039.safetensors",
897
+ "model.layers.6.block_sparse_moe.experts.7.w1.weight": "model-00009-of-00039.safetensors",
898
+ "model.layers.6.block_sparse_moe.experts.7.w2.weight": "model-00009-of-00039.safetensors",
899
+ "model.layers.6.block_sparse_moe.experts.7.w3.weight": "model-00009-of-00039.safetensors",
900
+ "model.layers.6.block_sparse_moe.gate.weight": "model-00008-of-00039.safetensors",
901
+ "model.layers.6.input_layernorm.weight": "model-00009-of-00039.safetensors",
902
+ "model.layers.6.post_attention_layernorm.weight": "model-00009-of-00039.safetensors",
903
+ "model.layers.6.self_attn.k_proj.weight": "model-00008-of-00039.safetensors",
904
+ "model.layers.6.self_attn.o_proj.weight": "model-00008-of-00039.safetensors",
905
+ "model.layers.6.self_attn.q_proj.weight": "model-00008-of-00039.safetensors",
906
+ "model.layers.6.self_attn.v_proj.weight": "model-00008-of-00039.safetensors",
907
+ "model.layers.7.block_sparse_moe.experts.0.w1.weight": "model-00009-of-00039.safetensors",
908
+ "model.layers.7.block_sparse_moe.experts.0.w2.weight": "model-00009-of-00039.safetensors",
909
+ "model.layers.7.block_sparse_moe.experts.0.w3.weight": "model-00009-of-00039.safetensors",
910
+ "model.layers.7.block_sparse_moe.experts.1.w1.weight": "model-00009-of-00039.safetensors",
911
+ "model.layers.7.block_sparse_moe.experts.1.w2.weight": "model-00009-of-00039.safetensors",
912
+ "model.layers.7.block_sparse_moe.experts.1.w3.weight": "model-00009-of-00039.safetensors",
913
+ "model.layers.7.block_sparse_moe.experts.2.w1.weight": "model-00009-of-00039.safetensors",
914
+ "model.layers.7.block_sparse_moe.experts.2.w2.weight": "model-00009-of-00039.safetensors",
915
+ "model.layers.7.block_sparse_moe.experts.2.w3.weight": "model-00009-of-00039.safetensors",
916
+ "model.layers.7.block_sparse_moe.experts.3.w1.weight": "model-00009-of-00039.safetensors",
917
+ "model.layers.7.block_sparse_moe.experts.3.w2.weight": "model-00009-of-00039.safetensors",
918
+ "model.layers.7.block_sparse_moe.experts.3.w3.weight": "model-00010-of-00039.safetensors",
919
+ "model.layers.7.block_sparse_moe.experts.4.w1.weight": "model-00010-of-00039.safetensors",
920
+ "model.layers.7.block_sparse_moe.experts.4.w2.weight": "model-00010-of-00039.safetensors",
921
+ "model.layers.7.block_sparse_moe.experts.4.w3.weight": "model-00010-of-00039.safetensors",
922
+ "model.layers.7.block_sparse_moe.experts.5.w1.weight": "model-00010-of-00039.safetensors",
923
+ "model.layers.7.block_sparse_moe.experts.5.w2.weight": "model-00010-of-00039.safetensors",
924
+ "model.layers.7.block_sparse_moe.experts.5.w3.weight": "model-00010-of-00039.safetensors",
925
+ "model.layers.7.block_sparse_moe.experts.6.w1.weight": "model-00010-of-00039.safetensors",
926
+ "model.layers.7.block_sparse_moe.experts.6.w2.weight": "model-00010-of-00039.safetensors",
927
+ "model.layers.7.block_sparse_moe.experts.6.w3.weight": "model-00010-of-00039.safetensors",
928
+ "model.layers.7.block_sparse_moe.experts.7.w1.weight": "model-00010-of-00039.safetensors",
929
+ "model.layers.7.block_sparse_moe.experts.7.w2.weight": "model-00010-of-00039.safetensors",
930
+ "model.layers.7.block_sparse_moe.experts.7.w3.weight": "model-00010-of-00039.safetensors",
931
+ "model.layers.7.block_sparse_moe.gate.weight": "model-00009-of-00039.safetensors",
932
+ "model.layers.7.input_layernorm.weight": "model-00010-of-00039.safetensors",
933
+ "model.layers.7.post_attention_layernorm.weight": "model-00010-of-00039.safetensors",
934
+ "model.layers.7.self_attn.k_proj.weight": "model-00009-of-00039.safetensors",
935
+ "model.layers.7.self_attn.o_proj.weight": "model-00009-of-00039.safetensors",
936
+ "model.layers.7.self_attn.q_proj.weight": "model-00009-of-00039.safetensors",
937
+ "model.layers.7.self_attn.v_proj.weight": "model-00009-of-00039.safetensors",
938
+ "model.layers.8.block_sparse_moe.experts.0.w1.weight": "model-00010-of-00039.safetensors",
939
+ "model.layers.8.block_sparse_moe.experts.0.w2.weight": "model-00010-of-00039.safetensors",
940
+ "model.layers.8.block_sparse_moe.experts.0.w3.weight": "model-00010-of-00039.safetensors",
941
+ "model.layers.8.block_sparse_moe.experts.1.w1.weight": "model-00010-of-00039.safetensors",
942
+ "model.layers.8.block_sparse_moe.experts.1.w2.weight": "model-00010-of-00039.safetensors",
943
+ "model.layers.8.block_sparse_moe.experts.1.w3.weight": "model-00010-of-00039.safetensors",
944
+ "model.layers.8.block_sparse_moe.experts.2.w1.weight": "model-00010-of-00039.safetensors",
945
+ "model.layers.8.block_sparse_moe.experts.2.w2.weight": "model-00011-of-00039.safetensors",
946
+ "model.layers.8.block_sparse_moe.experts.2.w3.weight": "model-00011-of-00039.safetensors",
947
+ "model.layers.8.block_sparse_moe.experts.3.w1.weight": "model-00011-of-00039.safetensors",
948
+ "model.layers.8.block_sparse_moe.experts.3.w2.weight": "model-00011-of-00039.safetensors",
949
+ "model.layers.8.block_sparse_moe.experts.3.w3.weight": "model-00011-of-00039.safetensors",
950
+ "model.layers.8.block_sparse_moe.experts.4.w1.weight": "model-00011-of-00039.safetensors",
951
+ "model.layers.8.block_sparse_moe.experts.4.w2.weight": "model-00011-of-00039.safetensors",
952
+ "model.layers.8.block_sparse_moe.experts.4.w3.weight": "model-00011-of-00039.safetensors",
953
+ "model.layers.8.block_sparse_moe.experts.5.w1.weight": "model-00011-of-00039.safetensors",
954
+ "model.layers.8.block_sparse_moe.experts.5.w2.weight": "model-00011-of-00039.safetensors",
955
+ "model.layers.8.block_sparse_moe.experts.5.w3.weight": "model-00011-of-00039.safetensors",
956
+ "model.layers.8.block_sparse_moe.experts.6.w1.weight": "model-00011-of-00039.safetensors",
957
+ "model.layers.8.block_sparse_moe.experts.6.w2.weight": "model-00011-of-00039.safetensors",
958
+ "model.layers.8.block_sparse_moe.experts.6.w3.weight": "model-00011-of-00039.safetensors",
959
+ "model.layers.8.block_sparse_moe.experts.7.w1.weight": "model-00011-of-00039.safetensors",
960
+ "model.layers.8.block_sparse_moe.experts.7.w2.weight": "model-00011-of-00039.safetensors",
961
+ "model.layers.8.block_sparse_moe.experts.7.w3.weight": "model-00011-of-00039.safetensors",
962
+ "model.layers.8.block_sparse_moe.gate.weight": "model-00010-of-00039.safetensors",
963
+ "model.layers.8.input_layernorm.weight": "model-00011-of-00039.safetensors",
964
+ "model.layers.8.post_attention_layernorm.weight": "model-00011-of-00039.safetensors",
965
+ "model.layers.8.self_attn.k_proj.weight": "model-00010-of-00039.safetensors",
966
+ "model.layers.8.self_attn.o_proj.weight": "model-00010-of-00039.safetensors",
967
+ "model.layers.8.self_attn.q_proj.weight": "model-00010-of-00039.safetensors",
968
+ "model.layers.8.self_attn.v_proj.weight": "model-00010-of-00039.safetensors",
969
+ "model.layers.9.block_sparse_moe.experts.0.w1.weight": "model-00011-of-00039.safetensors",
970
+ "model.layers.9.block_sparse_moe.experts.0.w2.weight": "model-00011-of-00039.safetensors",
971
+ "model.layers.9.block_sparse_moe.experts.0.w3.weight": "model-00011-of-00039.safetensors",
972
+ "model.layers.9.block_sparse_moe.experts.1.w1.weight": "model-00012-of-00039.safetensors",
973
+ "model.layers.9.block_sparse_moe.experts.1.w2.weight": "model-00012-of-00039.safetensors",
974
+ "model.layers.9.block_sparse_moe.experts.1.w3.weight": "model-00012-of-00039.safetensors",
975
+ "model.layers.9.block_sparse_moe.experts.2.w1.weight": "model-00012-of-00039.safetensors",
976
+ "model.layers.9.block_sparse_moe.experts.2.w2.weight": "model-00012-of-00039.safetensors",
977
+ "model.layers.9.block_sparse_moe.experts.2.w3.weight": "model-00012-of-00039.safetensors",
978
+ "model.layers.9.block_sparse_moe.experts.3.w1.weight": "model-00012-of-00039.safetensors",
979
+ "model.layers.9.block_sparse_moe.experts.3.w2.weight": "model-00012-of-00039.safetensors",
980
+ "model.layers.9.block_sparse_moe.experts.3.w3.weight": "model-00012-of-00039.safetensors",
981
+ "model.layers.9.block_sparse_moe.experts.4.w1.weight": "model-00012-of-00039.safetensors",
982
+ "model.layers.9.block_sparse_moe.experts.4.w2.weight": "model-00012-of-00039.safetensors",
983
+ "model.layers.9.block_sparse_moe.experts.4.w3.weight": "model-00012-of-00039.safetensors",
984
+ "model.layers.9.block_sparse_moe.experts.5.w1.weight": "model-00012-of-00039.safetensors",
985
+ "model.layers.9.block_sparse_moe.experts.5.w2.weight": "model-00012-of-00039.safetensors",
986
+ "model.layers.9.block_sparse_moe.experts.5.w3.weight": "model-00012-of-00039.safetensors",
987
+ "model.layers.9.block_sparse_moe.experts.6.w1.weight": "model-00012-of-00039.safetensors",
988
+ "model.layers.9.block_sparse_moe.experts.6.w2.weight": "model-00012-of-00039.safetensors",
989
+ "model.layers.9.block_sparse_moe.experts.6.w3.weight": "model-00012-of-00039.safetensors",
990
+ "model.layers.9.block_sparse_moe.experts.7.w1.weight": "model-00012-of-00039.safetensors",
991
+ "model.layers.9.block_sparse_moe.experts.7.w2.weight": "model-00012-of-00039.safetensors",
992
+ "model.layers.9.block_sparse_moe.experts.7.w3.weight": "model-00012-of-00039.safetensors",
993
+ "model.layers.9.block_sparse_moe.gate.weight": "model-00011-of-00039.safetensors",
994
+ "model.layers.9.input_layernorm.weight": "model-00012-of-00039.safetensors",
995
+ "model.layers.9.post_attention_layernorm.weight": "model-00012-of-00039.safetensors",
996
+ "model.layers.9.self_attn.k_proj.weight": "model-00011-of-00039.safetensors",
997
+ "model.layers.9.self_attn.o_proj.weight": "model-00011-of-00039.safetensors",
998
+ "model.layers.9.self_attn.q_proj.weight": "model-00011-of-00039.safetensors",
999
+ "model.layers.9.self_attn.v_proj.weight": "model-00011-of-00039.safetensors",
1000
+ "model.norm.weight": "model-00039-of-00039.safetensors"
1001
+ }
1002
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "unk_token": "<unk>"
5
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc4f0bd70b3709312d9d1d9e5ba674794b6bc5abc17429897a540f93882f25fc
3
+ size 1795303
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "additional_special_tokens": [],
31
+ "bos_token": "<s>",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "</s>",
34
+ "legacy": true,
35
+ "model_max_length": 1000000000000000019884624838656,
36
+ "pad_token": null,
37
+ "sp_model_kwargs": {},
38
+ "spaces_between_special_tokens": false,
39
+ "tokenizer_class": "LlamaTokenizer",
40
+ "unk_token": "<unk>",
41
+ "use_default_system_prompt": false
42
+ }