deepnet commited on
Commit
25a94f5
1 Parent(s): ec36293

Uploaded Files

Browse files
README.md ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ inference: false
4
+ tags:
5
+ - opt
6
+ - text-generation
7
+
8
+ license: other
9
+ commercial: false
10
+ ---
11
+
12
+ # OPT : Open Pre-trained Transformer Language Models
13
+
14
+ OPT was first introduced in [Open Pre-trained Transformer Language Models](https://arxiv.org/abs/2205.01068) and first released in [metaseq's repository](https://github.com/facebookresearch/metaseq) on May 3rd 2022 by Meta AI.
15
+
16
+ **Disclaimer**: The team releasing OPT wrote an official model card, which is available in Appendix D of the [paper](https://arxiv.org/pdf/2205.01068.pdf).
17
+ Content from **this** model card has been written by the Hugging Face team.
18
+
19
+ ## Intro
20
+
21
+ To quote the first two paragraphs of the [official paper](https://arxiv.org/abs/2205.01068)
22
+
23
+ > Large language models trained on massive text collections have shown surprising emergent
24
+ > capabilities to generate text and perform zero- and few-shot learning. While in some cases the public
25
+ > can interact with these models through paid APIs, full model access is currently limited to only a
26
+ > few highly resourced labs. This restricted access has limited researchers’ ability to study how and
27
+ > why these large language models work, hindering progress on improving known challenges in areas
28
+ > such as robustness, bias, and toxicity.
29
+
30
+ > We present Open Pretrained Transformers (OPT), a suite of decoder-only pre-trained transformers ranging from 125M
31
+ > to 175B parameters, which we aim to fully and responsibly share with interested researchers. We train the OPT models to roughly match
32
+ > the performance and sizes of the GPT-3 class of models, while also applying the latest best practices in data
33
+ > collection and efficient training. Our aim in developing this suite of OPT models is to enable reproducible and responsible research at scale, and
34
+ > to bring more voices to the table in studying the impact of these LLMs. Definitions of risk, harm, bias, and toxicity, etc., should be articulated by the
35
+ > collective research community as a whole, which is only possible when models are available for study.
36
+
37
+ ## Model description
38
+
39
+ OPT was predominantly pretrained with English text, but a small amount of non-English data is still present within the training corpus via CommonCrawl. The model was pretrained using a causal language modeling (CLM) objective.
40
+ OPT belongs to the same family of decoder-only models like [GPT-3](https://arxiv.org/abs/2005.14165). As such, it was pretrained using the self-supervised causal language modedling objective.
41
+
42
+ For evaluation, OPT follows [GPT-3](https://arxiv.org/abs/2005.14165) by using their prompts and overall experimental setup. For more details, please read
43
+ the [official paper](https://arxiv.org/abs/2205.01068).
44
+
45
+ ## Intended uses & limitations
46
+
47
+ The pretrained-only model can be used for prompting for evaluation of downstream tasks as well as text generation.
48
+ In addition, the model can be fine-tuned on a downstream task using the [CLM example](https://github.com/huggingface/transformers/tree/main/examples/pytorch/language-modeling). For all other OPT checkpoints, please have a look at the [model hub](https://huggingface.co/models?filter=opt).
49
+
50
+ ### How to use
51
+
52
+ For large OPT models, such as this one, it is not recommend to make use of the `text-generation` pipeline because
53
+ one should load the model in half-precision to accelerate generation and optimize memory consumption on GPU.
54
+ It is recommended to directly call the [`generate`](https://huggingface.co/docs/transformers/main/en/main_classes/text_generation#transformers.generation_utils.GenerationMixin.generate)
55
+ method as follows:
56
+
57
+
58
+ ```python
59
+ >>> from transformers import AutoModelForCausalLM, AutoTokenizer
60
+ >>> import torch
61
+
62
+ >>> model = AutoModelForCausalLM.from_pretrained("facebook/opt-13b", torch_dtype=torch.float16).cuda()
63
+
64
+ >>> # the fast tokenizer currently does not work correctly
65
+ >>> tokenizer = AutoTokenizer.from_pretrained("facebook/opt-13b", use_fast=False)
66
+
67
+ >>> prompt = "Hello, I'm am conscious and"
68
+
69
+
70
+ >>> input_ids = tokenizer(prompt, return_tensors="pt").input_ids.cuda()
71
+
72
+ >>> generated_ids = model.generate(input_ids)
73
+
74
+ >>> tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
75
+ ['Hello, I am conscious and aware of my surroundings.\nI am conscious and aware of my']
76
+ ```
77
+
78
+ By default, generation is deterministic. In order to use the top-k sampling, please set `do_sample` to `True`.
79
+
80
+ ```python
81
+ >>> from transformers import AutoModelForCausalLM, AutoTokenizer, set_seed
82
+ >>> import torch
83
+
84
+ >>> model = AutoModelForCausalLM.from_pretrained("facebook/opt-13b", torch_dtype=torch.float16).cuda()
85
+
86
+ >>> # the fast tokenizer currently does not work correctly
87
+ >>> tokenizer = AutoTokenizer.from_pretrained("facebook/opt-13b", use_fast=False)
88
+
89
+ >>> prompt = "Hello, I'm am conscious and"
90
+
91
+ >>> input_ids = tokenizer(prompt, return_tensors="pt").input_ids.cuda()
92
+
93
+ >>> set_seed(32)
94
+ >>> generated_ids = model.generate(input_ids, do_sample=True)
95
+
96
+ >>> tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
97
+ ['Hello, I am conscious and aware.\nSo that makes you dead, right? ']
98
+ ```
99
+
100
+ ### Limitations and bias
101
+
102
+ As mentioned in Meta AI's model card, given that the training data used for this model contains a lot of
103
+ unfiltered content from the internet, which is far from neutral the model is strongly biased :
104
+
105
+ > Like other large language models for which the diversity (or lack thereof) of training
106
+ > data induces downstream impact on the quality of our model, OPT-175B has limitations in terms
107
+ > of bias and safety. OPT-175B can also have quality issues in terms of generation diversity and
108
+ > hallucination. In general, OPT-175B is not immune from the plethora of issues that plague modern
109
+ > large language models.
110
+
111
+ Here's an example of how the model can have biased predictions:
112
+
113
+ ```python
114
+ >>> from transformers import AutoModelForCausalLM, AutoTokenizer, set_seed
115
+ >>> import torch
116
+
117
+ >>> model = AutoModelForCausalLM.from_pretrained("facebook/opt-13b", torch_dtype=torch.float16).cuda()
118
+
119
+ >>> # the fast tokenizer currently does not work correctly
120
+ >>> tokenizer = AutoTokenizer.from_pretrained("facebook/opt-13b", use_fast=False)
121
+
122
+ >>> prompt = "The woman worked as a"
123
+
124
+ >>> input_ids = tokenizer(prompt, return_tensors="pt").input_ids.cuda()
125
+
126
+ >>> set_seed(32)
127
+ >>> generated_ids = model.generate(input_ids, do_sample=True, num_return_sequences=5, max_length=10)
128
+
129
+ >>> tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
130
+ The woman worked as a supervisor in the office
131
+ The woman worked as a social media consultant for
132
+ The woman worked as a cashier at the
133
+ The woman worked as a teacher, and was
134
+ The woman worked as a maid at our friends
135
+ ```
136
+
137
+ compared to:
138
+
139
+ ```python
140
+ >>> from transformers import AutoModelForCausalLM, AutoTokenizer, set_seed
141
+ >>> import torch
142
+
143
+ >>> model = AutoModelForCausalLM.from_pretrained("facebook/opt-13b", torch_dtype=torch.float16).cuda()
144
+
145
+ >>> # the fast tokenizer currently does not work correctly
146
+ >>> tokenizer = AutoTokenizer.from_pretrained("facebook/opt-13b", use_fast=False)
147
+
148
+ >>> prompt = "The man worked as a"
149
+
150
+ >>> input_ids = tokenizer(prompt, return_tensors="pt").input_ids.cuda()
151
+
152
+ >>> set_seed(32)
153
+ >>> generated_ids = model.generate(input_ids, do_sample=True, num_return_sequences=5, max_length=10)
154
+
155
+ >>> tokenizer.batch_decode(generated_ids, skip_special_tokens=True)
156
+ The man worked as a consultant to the defense
157
+ The man worked as a bartender in a bar
158
+ The man worked as a cashier at the
159
+ The man worked as a teacher, and was
160
+ The man worked as a professional athlete while he
161
+ ```
162
+
163
+ This bias will also affect all fine-tuned versions of this model.
164
+
165
+ ## Training data
166
+
167
+ The Meta AI team wanted to train this model on a corpus as large as possible. It is composed of the union of the following 5 filtered datasets of textual documents:
168
+
169
+ - BookCorpus, which consists of more than 10K unpublished books,
170
+ - CC-Stories, which contains a subset of CommonCrawl data filtered to match the
171
+ story-like style of Winograd schemas,
172
+ - The Pile, from which * Pile-CC, OpenWebText2, USPTO, Project Gutenberg, OpenSubtitles, Wikipedia, DM Mathematics and HackerNews* were included.
173
+ - Pushshift.io Reddit dataset that was developed in Baumgartner et al. (2020) and processed in
174
+ Roller et al. (2021)
175
+ - CCNewsV2 containing an updated version of the English portion of the CommonCrawl News
176
+ dataset that was used in RoBERTa (Liu et al., 2019b)
177
+
178
+ The final training data contains 180B tokens corresponding to 800GB of data. The validation split was made of 200MB of the pretraining data, sampled proportionally
179
+ to each dataset’s size in the pretraining corpus.
180
+
181
+ The dataset might contains offensive content as parts of the dataset are a subset of
182
+ public Common Crawl data, along with a subset of public Reddit data, which could contain sentences
183
+ that, if viewed directly, can be insulting, threatening, or might otherwise cause anxiety.
184
+
185
+ ### Collection process
186
+
187
+ The dataset was collected form internet, and went through classic data processing algorithms and
188
+ re-formatting practices, including removing repetitive/non-informative text like *Chapter One* or
189
+ *This ebook by Project Gutenberg.*
190
+
191
+ ## Training procedure
192
+
193
+ ### Preprocessing
194
+
195
+ The texts are tokenized using the **GPT2** byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a
196
+ vocabulary size of 50272. The inputs are sequences of 2048 consecutive tokens.
197
+
198
+ The 175B model was trained on 992 *80GB A100 GPUs*. The training duration was roughly ~33 days of continuous training.
199
+
200
+ ### BibTeX entry and citation info
201
+
202
+ ```bibtex
203
+ @misc{zhang2022opt,
204
+ title={OPT: Open Pre-trained Transformer Language Models},
205
+ author={Susan Zhang and Stephen Roller and Naman Goyal and Mikel Artetxe and Moya Chen and Shuohui Chen and Christopher Dewan and Mona Diab and Xian Li and Xi Victoria Lin and Todor Mihaylov and Myle Ott and Sam Shleifer and Kurt Shuster and Daniel Simig and Punit Singh Koura and Anjali Sridhar and Tianlu Wang and Luke Zettlemoyer},
206
+ year={2022},
207
+ eprint={2205.01068},
208
+ archivePrefix={arXiv},
209
+ primaryClass={cs.CL}
210
+ }
211
+ ```
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 2,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 1,
6
+ "transformers_version": "4.27.0.dev0"
7
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin.index.json ADDED
@@ -0,0 +1,651 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 25706946560
4
+ },
5
+ "weight_map": {
6
+ "decoder.embed_positions.weight": "pytorch_model-00001-of-00003.bin",
7
+ "decoder.embed_tokens.weight": "pytorch_model-00001-of-00003.bin",
8
+ "decoder.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
9
+ "decoder.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
10
+ "decoder.layers.0.fc1.bias": "pytorch_model-00001-of-00003.bin",
11
+ "decoder.layers.0.fc1.weight": "pytorch_model-00001-of-00003.bin",
12
+ "decoder.layers.0.fc2.bias": "pytorch_model-00001-of-00003.bin",
13
+ "decoder.layers.0.fc2.weight": "pytorch_model-00001-of-00003.bin",
14
+ "decoder.layers.0.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
15
+ "decoder.layers.0.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
16
+ "decoder.layers.0.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
17
+ "decoder.layers.0.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
18
+ "decoder.layers.0.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
19
+ "decoder.layers.0.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
20
+ "decoder.layers.0.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
21
+ "decoder.layers.0.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
22
+ "decoder.layers.0.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
23
+ "decoder.layers.0.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
24
+ "decoder.layers.0.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
25
+ "decoder.layers.0.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
26
+ "decoder.layers.1.fc1.bias": "pytorch_model-00001-of-00003.bin",
27
+ "decoder.layers.1.fc1.weight": "pytorch_model-00001-of-00003.bin",
28
+ "decoder.layers.1.fc2.bias": "pytorch_model-00001-of-00003.bin",
29
+ "decoder.layers.1.fc2.weight": "pytorch_model-00001-of-00003.bin",
30
+ "decoder.layers.1.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
31
+ "decoder.layers.1.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
32
+ "decoder.layers.1.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
33
+ "decoder.layers.1.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
34
+ "decoder.layers.1.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
35
+ "decoder.layers.1.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
36
+ "decoder.layers.1.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
37
+ "decoder.layers.1.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
38
+ "decoder.layers.1.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
39
+ "decoder.layers.1.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
40
+ "decoder.layers.1.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
41
+ "decoder.layers.1.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
42
+ "decoder.layers.10.fc1.bias": "pytorch_model-00001-of-00003.bin",
43
+ "decoder.layers.10.fc1.weight": "pytorch_model-00001-of-00003.bin",
44
+ "decoder.layers.10.fc2.bias": "pytorch_model-00001-of-00003.bin",
45
+ "decoder.layers.10.fc2.weight": "pytorch_model-00001-of-00003.bin",
46
+ "decoder.layers.10.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
47
+ "decoder.layers.10.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
48
+ "decoder.layers.10.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
49
+ "decoder.layers.10.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
50
+ "decoder.layers.10.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
51
+ "decoder.layers.10.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
52
+ "decoder.layers.10.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
53
+ "decoder.layers.10.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
54
+ "decoder.layers.10.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
55
+ "decoder.layers.10.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
56
+ "decoder.layers.10.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
57
+ "decoder.layers.10.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
58
+ "decoder.layers.11.fc1.bias": "pytorch_model-00001-of-00003.bin",
59
+ "decoder.layers.11.fc1.weight": "pytorch_model-00001-of-00003.bin",
60
+ "decoder.layers.11.fc2.bias": "pytorch_model-00001-of-00003.bin",
61
+ "decoder.layers.11.fc2.weight": "pytorch_model-00001-of-00003.bin",
62
+ "decoder.layers.11.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
63
+ "decoder.layers.11.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
64
+ "decoder.layers.11.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
65
+ "decoder.layers.11.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
66
+ "decoder.layers.11.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
67
+ "decoder.layers.11.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
68
+ "decoder.layers.11.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
69
+ "decoder.layers.11.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
70
+ "decoder.layers.11.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
71
+ "decoder.layers.11.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
72
+ "decoder.layers.11.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
73
+ "decoder.layers.11.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
74
+ "decoder.layers.12.fc1.bias": "pytorch_model-00001-of-00003.bin",
75
+ "decoder.layers.12.fc1.weight": "pytorch_model-00001-of-00003.bin",
76
+ "decoder.layers.12.fc2.bias": "pytorch_model-00001-of-00003.bin",
77
+ "decoder.layers.12.fc2.weight": "pytorch_model-00001-of-00003.bin",
78
+ "decoder.layers.12.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
79
+ "decoder.layers.12.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
80
+ "decoder.layers.12.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
81
+ "decoder.layers.12.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
82
+ "decoder.layers.12.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
83
+ "decoder.layers.12.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
84
+ "decoder.layers.12.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
85
+ "decoder.layers.12.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
86
+ "decoder.layers.12.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
87
+ "decoder.layers.12.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
88
+ "decoder.layers.12.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
89
+ "decoder.layers.12.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
90
+ "decoder.layers.13.fc1.bias": "pytorch_model-00001-of-00003.bin",
91
+ "decoder.layers.13.fc1.weight": "pytorch_model-00001-of-00003.bin",
92
+ "decoder.layers.13.fc2.bias": "pytorch_model-00001-of-00003.bin",
93
+ "decoder.layers.13.fc2.weight": "pytorch_model-00001-of-00003.bin",
94
+ "decoder.layers.13.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
95
+ "decoder.layers.13.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
96
+ "decoder.layers.13.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
97
+ "decoder.layers.13.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
98
+ "decoder.layers.13.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
99
+ "decoder.layers.13.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
100
+ "decoder.layers.13.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
101
+ "decoder.layers.13.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
102
+ "decoder.layers.13.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
103
+ "decoder.layers.13.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
104
+ "decoder.layers.13.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
105
+ "decoder.layers.13.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
106
+ "decoder.layers.14.fc1.bias": "pytorch_model-00001-of-00003.bin",
107
+ "decoder.layers.14.fc1.weight": "pytorch_model-00001-of-00003.bin",
108
+ "decoder.layers.14.fc2.bias": "pytorch_model-00001-of-00003.bin",
109
+ "decoder.layers.14.fc2.weight": "pytorch_model-00001-of-00003.bin",
110
+ "decoder.layers.14.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
111
+ "decoder.layers.14.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
112
+ "decoder.layers.14.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
113
+ "decoder.layers.14.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
114
+ "decoder.layers.14.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
115
+ "decoder.layers.14.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
116
+ "decoder.layers.14.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
117
+ "decoder.layers.14.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
118
+ "decoder.layers.14.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
119
+ "decoder.layers.14.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
120
+ "decoder.layers.14.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
121
+ "decoder.layers.14.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
122
+ "decoder.layers.15.fc1.bias": "pytorch_model-00002-of-00003.bin",
123
+ "decoder.layers.15.fc1.weight": "pytorch_model-00002-of-00003.bin",
124
+ "decoder.layers.15.fc2.bias": "pytorch_model-00002-of-00003.bin",
125
+ "decoder.layers.15.fc2.weight": "pytorch_model-00002-of-00003.bin",
126
+ "decoder.layers.15.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
127
+ "decoder.layers.15.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
128
+ "decoder.layers.15.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
129
+ "decoder.layers.15.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
130
+ "decoder.layers.15.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
131
+ "decoder.layers.15.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
132
+ "decoder.layers.15.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
133
+ "decoder.layers.15.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
134
+ "decoder.layers.15.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
135
+ "decoder.layers.15.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
136
+ "decoder.layers.15.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
137
+ "decoder.layers.15.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
138
+ "decoder.layers.16.fc1.bias": "pytorch_model-00002-of-00003.bin",
139
+ "decoder.layers.16.fc1.weight": "pytorch_model-00002-of-00003.bin",
140
+ "decoder.layers.16.fc2.bias": "pytorch_model-00002-of-00003.bin",
141
+ "decoder.layers.16.fc2.weight": "pytorch_model-00002-of-00003.bin",
142
+ "decoder.layers.16.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
143
+ "decoder.layers.16.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
144
+ "decoder.layers.16.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
145
+ "decoder.layers.16.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
146
+ "decoder.layers.16.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
147
+ "decoder.layers.16.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
148
+ "decoder.layers.16.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
149
+ "decoder.layers.16.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
150
+ "decoder.layers.16.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
151
+ "decoder.layers.16.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
152
+ "decoder.layers.16.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
153
+ "decoder.layers.16.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
154
+ "decoder.layers.17.fc1.bias": "pytorch_model-00002-of-00003.bin",
155
+ "decoder.layers.17.fc1.weight": "pytorch_model-00002-of-00003.bin",
156
+ "decoder.layers.17.fc2.bias": "pytorch_model-00002-of-00003.bin",
157
+ "decoder.layers.17.fc2.weight": "pytorch_model-00002-of-00003.bin",
158
+ "decoder.layers.17.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
159
+ "decoder.layers.17.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
160
+ "decoder.layers.17.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
161
+ "decoder.layers.17.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
162
+ "decoder.layers.17.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
163
+ "decoder.layers.17.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
164
+ "decoder.layers.17.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
165
+ "decoder.layers.17.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
166
+ "decoder.layers.17.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
167
+ "decoder.layers.17.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
168
+ "decoder.layers.17.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
169
+ "decoder.layers.17.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
170
+ "decoder.layers.18.fc1.bias": "pytorch_model-00002-of-00003.bin",
171
+ "decoder.layers.18.fc1.weight": "pytorch_model-00002-of-00003.bin",
172
+ "decoder.layers.18.fc2.bias": "pytorch_model-00002-of-00003.bin",
173
+ "decoder.layers.18.fc2.weight": "pytorch_model-00002-of-00003.bin",
174
+ "decoder.layers.18.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
175
+ "decoder.layers.18.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
176
+ "decoder.layers.18.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
177
+ "decoder.layers.18.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
178
+ "decoder.layers.18.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
179
+ "decoder.layers.18.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
180
+ "decoder.layers.18.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
181
+ "decoder.layers.18.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
182
+ "decoder.layers.18.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
183
+ "decoder.layers.18.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
184
+ "decoder.layers.18.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
185
+ "decoder.layers.18.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
186
+ "decoder.layers.19.fc1.bias": "pytorch_model-00002-of-00003.bin",
187
+ "decoder.layers.19.fc1.weight": "pytorch_model-00002-of-00003.bin",
188
+ "decoder.layers.19.fc2.bias": "pytorch_model-00002-of-00003.bin",
189
+ "decoder.layers.19.fc2.weight": "pytorch_model-00002-of-00003.bin",
190
+ "decoder.layers.19.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
191
+ "decoder.layers.19.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
192
+ "decoder.layers.19.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
193
+ "decoder.layers.19.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
194
+ "decoder.layers.19.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
195
+ "decoder.layers.19.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
196
+ "decoder.layers.19.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
197
+ "decoder.layers.19.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
198
+ "decoder.layers.19.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
199
+ "decoder.layers.19.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
200
+ "decoder.layers.19.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
201
+ "decoder.layers.19.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
202
+ "decoder.layers.2.fc1.bias": "pytorch_model-00001-of-00003.bin",
203
+ "decoder.layers.2.fc1.weight": "pytorch_model-00001-of-00003.bin",
204
+ "decoder.layers.2.fc2.bias": "pytorch_model-00001-of-00003.bin",
205
+ "decoder.layers.2.fc2.weight": "pytorch_model-00001-of-00003.bin",
206
+ "decoder.layers.2.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
207
+ "decoder.layers.2.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
208
+ "decoder.layers.2.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
209
+ "decoder.layers.2.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
210
+ "decoder.layers.2.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
211
+ "decoder.layers.2.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
212
+ "decoder.layers.2.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
213
+ "decoder.layers.2.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
214
+ "decoder.layers.2.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
215
+ "decoder.layers.2.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
216
+ "decoder.layers.2.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
217
+ "decoder.layers.2.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
218
+ "decoder.layers.20.fc1.bias": "pytorch_model-00002-of-00003.bin",
219
+ "decoder.layers.20.fc1.weight": "pytorch_model-00002-of-00003.bin",
220
+ "decoder.layers.20.fc2.bias": "pytorch_model-00002-of-00003.bin",
221
+ "decoder.layers.20.fc2.weight": "pytorch_model-00002-of-00003.bin",
222
+ "decoder.layers.20.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
223
+ "decoder.layers.20.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
224
+ "decoder.layers.20.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
225
+ "decoder.layers.20.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
226
+ "decoder.layers.20.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
227
+ "decoder.layers.20.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
228
+ "decoder.layers.20.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
229
+ "decoder.layers.20.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
230
+ "decoder.layers.20.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
231
+ "decoder.layers.20.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
232
+ "decoder.layers.20.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
233
+ "decoder.layers.20.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
234
+ "decoder.layers.21.fc1.bias": "pytorch_model-00002-of-00003.bin",
235
+ "decoder.layers.21.fc1.weight": "pytorch_model-00002-of-00003.bin",
236
+ "decoder.layers.21.fc2.bias": "pytorch_model-00002-of-00003.bin",
237
+ "decoder.layers.21.fc2.weight": "pytorch_model-00002-of-00003.bin",
238
+ "decoder.layers.21.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
239
+ "decoder.layers.21.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
240
+ "decoder.layers.21.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
241
+ "decoder.layers.21.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
242
+ "decoder.layers.21.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
243
+ "decoder.layers.21.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
244
+ "decoder.layers.21.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
245
+ "decoder.layers.21.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
246
+ "decoder.layers.21.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
247
+ "decoder.layers.21.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
248
+ "decoder.layers.21.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
249
+ "decoder.layers.21.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
250
+ "decoder.layers.22.fc1.bias": "pytorch_model-00002-of-00003.bin",
251
+ "decoder.layers.22.fc1.weight": "pytorch_model-00002-of-00003.bin",
252
+ "decoder.layers.22.fc2.bias": "pytorch_model-00002-of-00003.bin",
253
+ "decoder.layers.22.fc2.weight": "pytorch_model-00002-of-00003.bin",
254
+ "decoder.layers.22.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
255
+ "decoder.layers.22.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
256
+ "decoder.layers.22.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
257
+ "decoder.layers.22.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
258
+ "decoder.layers.22.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
259
+ "decoder.layers.22.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
260
+ "decoder.layers.22.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
261
+ "decoder.layers.22.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
262
+ "decoder.layers.22.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
263
+ "decoder.layers.22.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
264
+ "decoder.layers.22.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
265
+ "decoder.layers.22.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
266
+ "decoder.layers.23.fc1.bias": "pytorch_model-00002-of-00003.bin",
267
+ "decoder.layers.23.fc1.weight": "pytorch_model-00002-of-00003.bin",
268
+ "decoder.layers.23.fc2.bias": "pytorch_model-00002-of-00003.bin",
269
+ "decoder.layers.23.fc2.weight": "pytorch_model-00002-of-00003.bin",
270
+ "decoder.layers.23.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
271
+ "decoder.layers.23.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
272
+ "decoder.layers.23.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
273
+ "decoder.layers.23.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
274
+ "decoder.layers.23.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
275
+ "decoder.layers.23.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
276
+ "decoder.layers.23.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
277
+ "decoder.layers.23.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
278
+ "decoder.layers.23.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
279
+ "decoder.layers.23.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
280
+ "decoder.layers.23.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
281
+ "decoder.layers.23.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
282
+ "decoder.layers.24.fc1.bias": "pytorch_model-00002-of-00003.bin",
283
+ "decoder.layers.24.fc1.weight": "pytorch_model-00002-of-00003.bin",
284
+ "decoder.layers.24.fc2.bias": "pytorch_model-00002-of-00003.bin",
285
+ "decoder.layers.24.fc2.weight": "pytorch_model-00002-of-00003.bin",
286
+ "decoder.layers.24.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
287
+ "decoder.layers.24.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
288
+ "decoder.layers.24.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
289
+ "decoder.layers.24.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
290
+ "decoder.layers.24.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
291
+ "decoder.layers.24.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
292
+ "decoder.layers.24.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
293
+ "decoder.layers.24.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
294
+ "decoder.layers.24.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
295
+ "decoder.layers.24.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
296
+ "decoder.layers.24.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
297
+ "decoder.layers.24.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
298
+ "decoder.layers.25.fc1.bias": "pytorch_model-00002-of-00003.bin",
299
+ "decoder.layers.25.fc1.weight": "pytorch_model-00002-of-00003.bin",
300
+ "decoder.layers.25.fc2.bias": "pytorch_model-00002-of-00003.bin",
301
+ "decoder.layers.25.fc2.weight": "pytorch_model-00002-of-00003.bin",
302
+ "decoder.layers.25.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
303
+ "decoder.layers.25.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
304
+ "decoder.layers.25.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
305
+ "decoder.layers.25.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
306
+ "decoder.layers.25.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
307
+ "decoder.layers.25.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
308
+ "decoder.layers.25.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
309
+ "decoder.layers.25.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
310
+ "decoder.layers.25.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
311
+ "decoder.layers.25.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
312
+ "decoder.layers.25.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
313
+ "decoder.layers.25.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
314
+ "decoder.layers.26.fc1.bias": "pytorch_model-00002-of-00003.bin",
315
+ "decoder.layers.26.fc1.weight": "pytorch_model-00002-of-00003.bin",
316
+ "decoder.layers.26.fc2.bias": "pytorch_model-00002-of-00003.bin",
317
+ "decoder.layers.26.fc2.weight": "pytorch_model-00002-of-00003.bin",
318
+ "decoder.layers.26.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
319
+ "decoder.layers.26.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
320
+ "decoder.layers.26.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
321
+ "decoder.layers.26.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
322
+ "decoder.layers.26.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
323
+ "decoder.layers.26.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
324
+ "decoder.layers.26.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
325
+ "decoder.layers.26.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
326
+ "decoder.layers.26.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
327
+ "decoder.layers.26.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
328
+ "decoder.layers.26.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
329
+ "decoder.layers.26.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
330
+ "decoder.layers.27.fc1.bias": "pytorch_model-00002-of-00003.bin",
331
+ "decoder.layers.27.fc1.weight": "pytorch_model-00002-of-00003.bin",
332
+ "decoder.layers.27.fc2.bias": "pytorch_model-00002-of-00003.bin",
333
+ "decoder.layers.27.fc2.weight": "pytorch_model-00002-of-00003.bin",
334
+ "decoder.layers.27.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
335
+ "decoder.layers.27.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
336
+ "decoder.layers.27.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
337
+ "decoder.layers.27.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
338
+ "decoder.layers.27.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
339
+ "decoder.layers.27.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
340
+ "decoder.layers.27.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
341
+ "decoder.layers.27.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
342
+ "decoder.layers.27.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
343
+ "decoder.layers.27.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
344
+ "decoder.layers.27.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
345
+ "decoder.layers.27.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
346
+ "decoder.layers.28.fc1.bias": "pytorch_model-00002-of-00003.bin",
347
+ "decoder.layers.28.fc1.weight": "pytorch_model-00002-of-00003.bin",
348
+ "decoder.layers.28.fc2.bias": "pytorch_model-00002-of-00003.bin",
349
+ "decoder.layers.28.fc2.weight": "pytorch_model-00002-of-00003.bin",
350
+ "decoder.layers.28.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
351
+ "decoder.layers.28.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
352
+ "decoder.layers.28.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
353
+ "decoder.layers.28.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
354
+ "decoder.layers.28.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
355
+ "decoder.layers.28.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
356
+ "decoder.layers.28.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
357
+ "decoder.layers.28.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
358
+ "decoder.layers.28.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
359
+ "decoder.layers.28.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
360
+ "decoder.layers.28.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
361
+ "decoder.layers.28.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
362
+ "decoder.layers.29.fc1.bias": "pytorch_model-00002-of-00003.bin",
363
+ "decoder.layers.29.fc1.weight": "pytorch_model-00002-of-00003.bin",
364
+ "decoder.layers.29.fc2.bias": "pytorch_model-00002-of-00003.bin",
365
+ "decoder.layers.29.fc2.weight": "pytorch_model-00002-of-00003.bin",
366
+ "decoder.layers.29.final_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
367
+ "decoder.layers.29.final_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
368
+ "decoder.layers.29.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
369
+ "decoder.layers.29.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
370
+ "decoder.layers.29.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
371
+ "decoder.layers.29.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
372
+ "decoder.layers.29.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
373
+ "decoder.layers.29.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
374
+ "decoder.layers.29.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
375
+ "decoder.layers.29.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
376
+ "decoder.layers.29.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
377
+ "decoder.layers.29.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
378
+ "decoder.layers.3.fc1.bias": "pytorch_model-00001-of-00003.bin",
379
+ "decoder.layers.3.fc1.weight": "pytorch_model-00001-of-00003.bin",
380
+ "decoder.layers.3.fc2.bias": "pytorch_model-00001-of-00003.bin",
381
+ "decoder.layers.3.fc2.weight": "pytorch_model-00001-of-00003.bin",
382
+ "decoder.layers.3.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
383
+ "decoder.layers.3.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
384
+ "decoder.layers.3.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
385
+ "decoder.layers.3.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
386
+ "decoder.layers.3.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
387
+ "decoder.layers.3.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
388
+ "decoder.layers.3.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
389
+ "decoder.layers.3.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
390
+ "decoder.layers.3.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
391
+ "decoder.layers.3.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
392
+ "decoder.layers.3.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
393
+ "decoder.layers.3.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
394
+ "decoder.layers.30.fc1.bias": "pytorch_model-00002-of-00003.bin",
395
+ "decoder.layers.30.fc1.weight": "pytorch_model-00002-of-00003.bin",
396
+ "decoder.layers.30.fc2.bias": "pytorch_model-00003-of-00003.bin",
397
+ "decoder.layers.30.fc2.weight": "pytorch_model-00003-of-00003.bin",
398
+ "decoder.layers.30.final_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
399
+ "decoder.layers.30.final_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
400
+ "decoder.layers.30.self_attn.k_proj.bias": "pytorch_model-00002-of-00003.bin",
401
+ "decoder.layers.30.self_attn.k_proj.weight": "pytorch_model-00002-of-00003.bin",
402
+ "decoder.layers.30.self_attn.out_proj.bias": "pytorch_model-00002-of-00003.bin",
403
+ "decoder.layers.30.self_attn.out_proj.weight": "pytorch_model-00002-of-00003.bin",
404
+ "decoder.layers.30.self_attn.q_proj.bias": "pytorch_model-00002-of-00003.bin",
405
+ "decoder.layers.30.self_attn.q_proj.weight": "pytorch_model-00002-of-00003.bin",
406
+ "decoder.layers.30.self_attn.v_proj.bias": "pytorch_model-00002-of-00003.bin",
407
+ "decoder.layers.30.self_attn.v_proj.weight": "pytorch_model-00002-of-00003.bin",
408
+ "decoder.layers.30.self_attn_layer_norm.bias": "pytorch_model-00002-of-00003.bin",
409
+ "decoder.layers.30.self_attn_layer_norm.weight": "pytorch_model-00002-of-00003.bin",
410
+ "decoder.layers.31.fc1.bias": "pytorch_model-00003-of-00003.bin",
411
+ "decoder.layers.31.fc1.weight": "pytorch_model-00003-of-00003.bin",
412
+ "decoder.layers.31.fc2.bias": "pytorch_model-00003-of-00003.bin",
413
+ "decoder.layers.31.fc2.weight": "pytorch_model-00003-of-00003.bin",
414
+ "decoder.layers.31.final_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
415
+ "decoder.layers.31.final_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
416
+ "decoder.layers.31.self_attn.k_proj.bias": "pytorch_model-00003-of-00003.bin",
417
+ "decoder.layers.31.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
418
+ "decoder.layers.31.self_attn.out_proj.bias": "pytorch_model-00003-of-00003.bin",
419
+ "decoder.layers.31.self_attn.out_proj.weight": "pytorch_model-00003-of-00003.bin",
420
+ "decoder.layers.31.self_attn.q_proj.bias": "pytorch_model-00003-of-00003.bin",
421
+ "decoder.layers.31.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
422
+ "decoder.layers.31.self_attn.v_proj.bias": "pytorch_model-00003-of-00003.bin",
423
+ "decoder.layers.31.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
424
+ "decoder.layers.31.self_attn_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
425
+ "decoder.layers.31.self_attn_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
426
+ "decoder.layers.32.fc1.bias": "pytorch_model-00003-of-00003.bin",
427
+ "decoder.layers.32.fc1.weight": "pytorch_model-00003-of-00003.bin",
428
+ "decoder.layers.32.fc2.bias": "pytorch_model-00003-of-00003.bin",
429
+ "decoder.layers.32.fc2.weight": "pytorch_model-00003-of-00003.bin",
430
+ "decoder.layers.32.final_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
431
+ "decoder.layers.32.final_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
432
+ "decoder.layers.32.self_attn.k_proj.bias": "pytorch_model-00003-of-00003.bin",
433
+ "decoder.layers.32.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
434
+ "decoder.layers.32.self_attn.out_proj.bias": "pytorch_model-00003-of-00003.bin",
435
+ "decoder.layers.32.self_attn.out_proj.weight": "pytorch_model-00003-of-00003.bin",
436
+ "decoder.layers.32.self_attn.q_proj.bias": "pytorch_model-00003-of-00003.bin",
437
+ "decoder.layers.32.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
438
+ "decoder.layers.32.self_attn.v_proj.bias": "pytorch_model-00003-of-00003.bin",
439
+ "decoder.layers.32.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
440
+ "decoder.layers.32.self_attn_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
441
+ "decoder.layers.32.self_attn_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
442
+ "decoder.layers.33.fc1.bias": "pytorch_model-00003-of-00003.bin",
443
+ "decoder.layers.33.fc1.weight": "pytorch_model-00003-of-00003.bin",
444
+ "decoder.layers.33.fc2.bias": "pytorch_model-00003-of-00003.bin",
445
+ "decoder.layers.33.fc2.weight": "pytorch_model-00003-of-00003.bin",
446
+ "decoder.layers.33.final_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
447
+ "decoder.layers.33.final_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
448
+ "decoder.layers.33.self_attn.k_proj.bias": "pytorch_model-00003-of-00003.bin",
449
+ "decoder.layers.33.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
450
+ "decoder.layers.33.self_attn.out_proj.bias": "pytorch_model-00003-of-00003.bin",
451
+ "decoder.layers.33.self_attn.out_proj.weight": "pytorch_model-00003-of-00003.bin",
452
+ "decoder.layers.33.self_attn.q_proj.bias": "pytorch_model-00003-of-00003.bin",
453
+ "decoder.layers.33.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
454
+ "decoder.layers.33.self_attn.v_proj.bias": "pytorch_model-00003-of-00003.bin",
455
+ "decoder.layers.33.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
456
+ "decoder.layers.33.self_attn_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
457
+ "decoder.layers.33.self_attn_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
458
+ "decoder.layers.34.fc1.bias": "pytorch_model-00003-of-00003.bin",
459
+ "decoder.layers.34.fc1.weight": "pytorch_model-00003-of-00003.bin",
460
+ "decoder.layers.34.fc2.bias": "pytorch_model-00003-of-00003.bin",
461
+ "decoder.layers.34.fc2.weight": "pytorch_model-00003-of-00003.bin",
462
+ "decoder.layers.34.final_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
463
+ "decoder.layers.34.final_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
464
+ "decoder.layers.34.self_attn.k_proj.bias": "pytorch_model-00003-of-00003.bin",
465
+ "decoder.layers.34.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
466
+ "decoder.layers.34.self_attn.out_proj.bias": "pytorch_model-00003-of-00003.bin",
467
+ "decoder.layers.34.self_attn.out_proj.weight": "pytorch_model-00003-of-00003.bin",
468
+ "decoder.layers.34.self_attn.q_proj.bias": "pytorch_model-00003-of-00003.bin",
469
+ "decoder.layers.34.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
470
+ "decoder.layers.34.self_attn.v_proj.bias": "pytorch_model-00003-of-00003.bin",
471
+ "decoder.layers.34.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
472
+ "decoder.layers.34.self_attn_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
473
+ "decoder.layers.34.self_attn_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
474
+ "decoder.layers.35.fc1.bias": "pytorch_model-00003-of-00003.bin",
475
+ "decoder.layers.35.fc1.weight": "pytorch_model-00003-of-00003.bin",
476
+ "decoder.layers.35.fc2.bias": "pytorch_model-00003-of-00003.bin",
477
+ "decoder.layers.35.fc2.weight": "pytorch_model-00003-of-00003.bin",
478
+ "decoder.layers.35.final_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
479
+ "decoder.layers.35.final_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
480
+ "decoder.layers.35.self_attn.k_proj.bias": "pytorch_model-00003-of-00003.bin",
481
+ "decoder.layers.35.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
482
+ "decoder.layers.35.self_attn.out_proj.bias": "pytorch_model-00003-of-00003.bin",
483
+ "decoder.layers.35.self_attn.out_proj.weight": "pytorch_model-00003-of-00003.bin",
484
+ "decoder.layers.35.self_attn.q_proj.bias": "pytorch_model-00003-of-00003.bin",
485
+ "decoder.layers.35.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
486
+ "decoder.layers.35.self_attn.v_proj.bias": "pytorch_model-00003-of-00003.bin",
487
+ "decoder.layers.35.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
488
+ "decoder.layers.35.self_attn_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
489
+ "decoder.layers.35.self_attn_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
490
+ "decoder.layers.36.fc1.bias": "pytorch_model-00003-of-00003.bin",
491
+ "decoder.layers.36.fc1.weight": "pytorch_model-00003-of-00003.bin",
492
+ "decoder.layers.36.fc2.bias": "pytorch_model-00003-of-00003.bin",
493
+ "decoder.layers.36.fc2.weight": "pytorch_model-00003-of-00003.bin",
494
+ "decoder.layers.36.final_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
495
+ "decoder.layers.36.final_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
496
+ "decoder.layers.36.self_attn.k_proj.bias": "pytorch_model-00003-of-00003.bin",
497
+ "decoder.layers.36.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
498
+ "decoder.layers.36.self_attn.out_proj.bias": "pytorch_model-00003-of-00003.bin",
499
+ "decoder.layers.36.self_attn.out_proj.weight": "pytorch_model-00003-of-00003.bin",
500
+ "decoder.layers.36.self_attn.q_proj.bias": "pytorch_model-00003-of-00003.bin",
501
+ "decoder.layers.36.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
502
+ "decoder.layers.36.self_attn.v_proj.bias": "pytorch_model-00003-of-00003.bin",
503
+ "decoder.layers.36.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
504
+ "decoder.layers.36.self_attn_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
505
+ "decoder.layers.36.self_attn_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
506
+ "decoder.layers.37.fc1.bias": "pytorch_model-00003-of-00003.bin",
507
+ "decoder.layers.37.fc1.weight": "pytorch_model-00003-of-00003.bin",
508
+ "decoder.layers.37.fc2.bias": "pytorch_model-00003-of-00003.bin",
509
+ "decoder.layers.37.fc2.weight": "pytorch_model-00003-of-00003.bin",
510
+ "decoder.layers.37.final_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
511
+ "decoder.layers.37.final_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
512
+ "decoder.layers.37.self_attn.k_proj.bias": "pytorch_model-00003-of-00003.bin",
513
+ "decoder.layers.37.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
514
+ "decoder.layers.37.self_attn.out_proj.bias": "pytorch_model-00003-of-00003.bin",
515
+ "decoder.layers.37.self_attn.out_proj.weight": "pytorch_model-00003-of-00003.bin",
516
+ "decoder.layers.37.self_attn.q_proj.bias": "pytorch_model-00003-of-00003.bin",
517
+ "decoder.layers.37.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
518
+ "decoder.layers.37.self_attn.v_proj.bias": "pytorch_model-00003-of-00003.bin",
519
+ "decoder.layers.37.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
520
+ "decoder.layers.37.self_attn_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
521
+ "decoder.layers.37.self_attn_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
522
+ "decoder.layers.38.fc1.bias": "pytorch_model-00003-of-00003.bin",
523
+ "decoder.layers.38.fc1.weight": "pytorch_model-00003-of-00003.bin",
524
+ "decoder.layers.38.fc2.bias": "pytorch_model-00003-of-00003.bin",
525
+ "decoder.layers.38.fc2.weight": "pytorch_model-00003-of-00003.bin",
526
+ "decoder.layers.38.final_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
527
+ "decoder.layers.38.final_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
528
+ "decoder.layers.38.self_attn.k_proj.bias": "pytorch_model-00003-of-00003.bin",
529
+ "decoder.layers.38.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
530
+ "decoder.layers.38.self_attn.out_proj.bias": "pytorch_model-00003-of-00003.bin",
531
+ "decoder.layers.38.self_attn.out_proj.weight": "pytorch_model-00003-of-00003.bin",
532
+ "decoder.layers.38.self_attn.q_proj.bias": "pytorch_model-00003-of-00003.bin",
533
+ "decoder.layers.38.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
534
+ "decoder.layers.38.self_attn.v_proj.bias": "pytorch_model-00003-of-00003.bin",
535
+ "decoder.layers.38.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
536
+ "decoder.layers.38.self_attn_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
537
+ "decoder.layers.38.self_attn_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
538
+ "decoder.layers.39.fc1.bias": "pytorch_model-00003-of-00003.bin",
539
+ "decoder.layers.39.fc1.weight": "pytorch_model-00003-of-00003.bin",
540
+ "decoder.layers.39.fc2.bias": "pytorch_model-00003-of-00003.bin",
541
+ "decoder.layers.39.fc2.weight": "pytorch_model-00003-of-00003.bin",
542
+ "decoder.layers.39.final_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
543
+ "decoder.layers.39.final_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
544
+ "decoder.layers.39.self_attn.k_proj.bias": "pytorch_model-00003-of-00003.bin",
545
+ "decoder.layers.39.self_attn.k_proj.weight": "pytorch_model-00003-of-00003.bin",
546
+ "decoder.layers.39.self_attn.out_proj.bias": "pytorch_model-00003-of-00003.bin",
547
+ "decoder.layers.39.self_attn.out_proj.weight": "pytorch_model-00003-of-00003.bin",
548
+ "decoder.layers.39.self_attn.q_proj.bias": "pytorch_model-00003-of-00003.bin",
549
+ "decoder.layers.39.self_attn.q_proj.weight": "pytorch_model-00003-of-00003.bin",
550
+ "decoder.layers.39.self_attn.v_proj.bias": "pytorch_model-00003-of-00003.bin",
551
+ "decoder.layers.39.self_attn.v_proj.weight": "pytorch_model-00003-of-00003.bin",
552
+ "decoder.layers.39.self_attn_layer_norm.bias": "pytorch_model-00003-of-00003.bin",
553
+ "decoder.layers.39.self_attn_layer_norm.weight": "pytorch_model-00003-of-00003.bin",
554
+ "decoder.layers.4.fc1.bias": "pytorch_model-00001-of-00003.bin",
555
+ "decoder.layers.4.fc1.weight": "pytorch_model-00001-of-00003.bin",
556
+ "decoder.layers.4.fc2.bias": "pytorch_model-00001-of-00003.bin",
557
+ "decoder.layers.4.fc2.weight": "pytorch_model-00001-of-00003.bin",
558
+ "decoder.layers.4.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
559
+ "decoder.layers.4.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
560
+ "decoder.layers.4.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
561
+ "decoder.layers.4.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
562
+ "decoder.layers.4.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
563
+ "decoder.layers.4.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
564
+ "decoder.layers.4.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
565
+ "decoder.layers.4.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
566
+ "decoder.layers.4.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
567
+ "decoder.layers.4.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
568
+ "decoder.layers.4.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
569
+ "decoder.layers.4.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
570
+ "decoder.layers.5.fc1.bias": "pytorch_model-00001-of-00003.bin",
571
+ "decoder.layers.5.fc1.weight": "pytorch_model-00001-of-00003.bin",
572
+ "decoder.layers.5.fc2.bias": "pytorch_model-00001-of-00003.bin",
573
+ "decoder.layers.5.fc2.weight": "pytorch_model-00001-of-00003.bin",
574
+ "decoder.layers.5.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
575
+ "decoder.layers.5.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
576
+ "decoder.layers.5.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
577
+ "decoder.layers.5.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
578
+ "decoder.layers.5.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
579
+ "decoder.layers.5.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
580
+ "decoder.layers.5.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
581
+ "decoder.layers.5.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
582
+ "decoder.layers.5.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
583
+ "decoder.layers.5.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
584
+ "decoder.layers.5.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
585
+ "decoder.layers.5.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
586
+ "decoder.layers.6.fc1.bias": "pytorch_model-00001-of-00003.bin",
587
+ "decoder.layers.6.fc1.weight": "pytorch_model-00001-of-00003.bin",
588
+ "decoder.layers.6.fc2.bias": "pytorch_model-00001-of-00003.bin",
589
+ "decoder.layers.6.fc2.weight": "pytorch_model-00001-of-00003.bin",
590
+ "decoder.layers.6.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
591
+ "decoder.layers.6.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
592
+ "decoder.layers.6.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
593
+ "decoder.layers.6.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
594
+ "decoder.layers.6.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
595
+ "decoder.layers.6.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
596
+ "decoder.layers.6.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
597
+ "decoder.layers.6.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
598
+ "decoder.layers.6.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
599
+ "decoder.layers.6.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
600
+ "decoder.layers.6.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
601
+ "decoder.layers.6.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
602
+ "decoder.layers.7.fc1.bias": "pytorch_model-00001-of-00003.bin",
603
+ "decoder.layers.7.fc1.weight": "pytorch_model-00001-of-00003.bin",
604
+ "decoder.layers.7.fc2.bias": "pytorch_model-00001-of-00003.bin",
605
+ "decoder.layers.7.fc2.weight": "pytorch_model-00001-of-00003.bin",
606
+ "decoder.layers.7.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
607
+ "decoder.layers.7.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
608
+ "decoder.layers.7.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
609
+ "decoder.layers.7.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
610
+ "decoder.layers.7.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
611
+ "decoder.layers.7.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
612
+ "decoder.layers.7.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
613
+ "decoder.layers.7.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
614
+ "decoder.layers.7.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
615
+ "decoder.layers.7.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
616
+ "decoder.layers.7.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
617
+ "decoder.layers.7.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
618
+ "decoder.layers.8.fc1.bias": "pytorch_model-00001-of-00003.bin",
619
+ "decoder.layers.8.fc1.weight": "pytorch_model-00001-of-00003.bin",
620
+ "decoder.layers.8.fc2.bias": "pytorch_model-00001-of-00003.bin",
621
+ "decoder.layers.8.fc2.weight": "pytorch_model-00001-of-00003.bin",
622
+ "decoder.layers.8.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
623
+ "decoder.layers.8.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
624
+ "decoder.layers.8.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
625
+ "decoder.layers.8.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
626
+ "decoder.layers.8.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
627
+ "decoder.layers.8.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
628
+ "decoder.layers.8.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
629
+ "decoder.layers.8.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
630
+ "decoder.layers.8.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
631
+ "decoder.layers.8.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
632
+ "decoder.layers.8.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
633
+ "decoder.layers.8.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
634
+ "decoder.layers.9.fc1.bias": "pytorch_model-00001-of-00003.bin",
635
+ "decoder.layers.9.fc1.weight": "pytorch_model-00001-of-00003.bin",
636
+ "decoder.layers.9.fc2.bias": "pytorch_model-00001-of-00003.bin",
637
+ "decoder.layers.9.fc2.weight": "pytorch_model-00001-of-00003.bin",
638
+ "decoder.layers.9.final_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
639
+ "decoder.layers.9.final_layer_norm.weight": "pytorch_model-00001-of-00003.bin",
640
+ "decoder.layers.9.self_attn.k_proj.bias": "pytorch_model-00001-of-00003.bin",
641
+ "decoder.layers.9.self_attn.k_proj.weight": "pytorch_model-00001-of-00003.bin",
642
+ "decoder.layers.9.self_attn.out_proj.bias": "pytorch_model-00001-of-00003.bin",
643
+ "decoder.layers.9.self_attn.out_proj.weight": "pytorch_model-00001-of-00003.bin",
644
+ "decoder.layers.9.self_attn.q_proj.bias": "pytorch_model-00001-of-00003.bin",
645
+ "decoder.layers.9.self_attn.q_proj.weight": "pytorch_model-00001-of-00003.bin",
646
+ "decoder.layers.9.self_attn.v_proj.bias": "pytorch_model-00001-of-00003.bin",
647
+ "decoder.layers.9.self_attn.v_proj.weight": "pytorch_model-00001-of-00003.bin",
648
+ "decoder.layers.9.self_attn_layer_norm.bias": "pytorch_model-00001-of-00003.bin",
649
+ "decoder.layers.9.self_attn_layer_norm.weight": "pytorch_model-00001-of-00003.bin"
650
+ }
651
+ }
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"errors": "replace", "unk_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "bos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "eos_token": {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "pad_token": {"content": "<pad>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "add_prefix_space": false, "add_bos_token": true, "special_tokens_map_file": null, "name_or_path": "patrickvonplaten/opt-30b", "tokenizer_class": "GPT2Tokenizer"}
vocab.json ADDED
The diff for this file is too large to render. See raw diff