.gitattributes CHANGED
@@ -1,34 +1,35 @@
1
  *.7z filter=lfs diff=lfs merge=lfs -text
2
  *.arrow filter=lfs diff=lfs merge=lfs -text
3
  *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bin.* filter=lfs diff=lfs merge=lfs -text
5
  *.bz2 filter=lfs diff=lfs merge=lfs -text
 
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
7
  *.gz filter=lfs diff=lfs merge=lfs -text
8
  *.h5 filter=lfs diff=lfs merge=lfs -text
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
10
  *.lfs.* filter=lfs diff=lfs merge=lfs -text
 
11
  *.model filter=lfs diff=lfs merge=lfs -text
12
  *.msgpack filter=lfs diff=lfs merge=lfs -text
 
 
13
  *.onnx filter=lfs diff=lfs merge=lfs -text
14
  *.ot filter=lfs diff=lfs merge=lfs -text
15
  *.parquet filter=lfs diff=lfs merge=lfs -text
16
  *.pb filter=lfs diff=lfs merge=lfs -text
 
 
17
  *.pt filter=lfs diff=lfs merge=lfs -text
18
  *.pth filter=lfs diff=lfs merge=lfs -text
19
  *.rar filter=lfs diff=lfs merge=lfs -text
 
20
  saved_model/**/* filter=lfs diff=lfs merge=lfs -text
21
  *.tar.* filter=lfs diff=lfs merge=lfs -text
 
22
  *.tflite filter=lfs diff=lfs merge=lfs -text
23
  *.tgz filter=lfs diff=lfs merge=lfs -text
 
24
  *.xz filter=lfs diff=lfs merge=lfs -text
25
  *.zip filter=lfs diff=lfs merge=lfs -text
26
- *.zstandard filter=lfs diff=lfs merge=lfs -text
27
- *.tfevents* filter=lfs diff=lfs merge=lfs -text
28
- *.db* filter=lfs diff=lfs merge=lfs -text
29
- *.ark* filter=lfs diff=lfs merge=lfs -text
30
- **/*ckpt*data* filter=lfs diff=lfs merge=lfs -text
31
- **/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text
32
- **/*ckpt*.index filter=lfs diff=lfs merge=lfs -text
33
- *.safetensors filter=lfs diff=lfs merge=lfs -text
34
- *.ckpt filter=lfs diff=lfs merge=lfs -text
 
1
  *.7z filter=lfs diff=lfs merge=lfs -text
2
  *.arrow filter=lfs diff=lfs merge=lfs -text
3
  *.bin filter=lfs diff=lfs merge=lfs -text
 
4
  *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
7
  *.gz filter=lfs diff=lfs merge=lfs -text
8
  *.h5 filter=lfs diff=lfs merge=lfs -text
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
10
  *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
  *.model filter=lfs diff=lfs merge=lfs -text
13
  *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
  *.onnx filter=lfs diff=lfs merge=lfs -text
17
  *.ot filter=lfs diff=lfs merge=lfs -text
18
  *.parquet filter=lfs diff=lfs merge=lfs -text
19
  *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
  *.pt filter=lfs diff=lfs merge=lfs -text
23
  *.pth filter=lfs diff=lfs merge=lfs -text
24
  *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
  saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
  *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
  *.tflite filter=lfs diff=lfs merge=lfs -text
30
  *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
  *.xz filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
LOGO.png CHANGED
README.md CHANGED
@@ -1,10 +1,8 @@
1
  ---
2
  license: other
3
- tasks:
4
- - code-generation
5
  ---
6
  # Model Card for CodeFuse-CodeLlama-34B
7
- ![logo](LOGO.png)
8
 
9
  [[中文]](#chinese) [[English]](#english)
10
 
@@ -20,38 +18,17 @@ The context length of finetuning is 4K while it is able to be finetuned by 16k c
20
 
21
  ## News and Updates
22
 
23
- 🔥🔥🔥 2023-09-26 We are pleased to announce the release of the [4-bit quantized version](https://huggingface.co/codefuse-ai/CodeFuse-CodeLlama-34B-4bits) of CodeFuse-CodeLlama-34B. Despite the quantization process, the model still achieves a remarkable 73.8% accuracy (greedy decoding) on the HumanEval pass@1 metric.
24
-
25
- 🔥🔥🔥 2023-09-11 CodeFuse-CodeLlama34B has achieved 74.4% of pass@1 (greedy decoding) on HumanEval, which is SOTA results for openspurced LLMs at present.
26
 
27
  <br>
28
 
29
- ## Code Community
30
-
31
- **Homepage**: 🏡 https://github.com/codefuse-ai (**Please give us your support with a Star🌟 + Fork🚀 + Watch👀**)
32
-
33
- + If you wish to fine-tune the model yourself, you can visit ✨[MFTCoder](https://github.com/codefuse-ai/MFTCoder)✨✨
34
-
35
- + If you wish to deploy the model yourself, you can visit ✨[FasterTransformer4CodeFuse](https://github.com/codefuse-ai/FasterTransformer4CodeFuse)✨✨
36
-
37
- + If you wish to see a demo of the model, you can visit ✨[CodeFuse Demo](https://github.com/codefuse-ai/codefuse)✨✨
38
-
39
-
40
  ## Performance
41
 
42
-
43
- | Model | HumanEval(pass@1) | Date |
44
- |:----------------------------|:-----------------:|:-------:|
45
- | **CodeFuse-CodeLlama-34B** | **74.4%** | 2023.9 |
46
- | WizardCoder-Python-34B-V1.0 | 73.2% | 2023.8 |
47
- | GPT-4(zero-shot) | 67.0% | 2023.3 |
48
- | PanGu-Coder2 15B | 61.6% | 2023.8 |
49
- | CodeLlama-34b-Python | 53.7% | 2023.8 |
50
- | CodeLlama-34b | 48.8% | 2023.8 |
51
- | GPT-3.5(zero-shot) | 48.1% | 2022.11 |
52
- | OctoCoder | 46.2% | 2023.8 |
53
- | StarCoder-15B | 33.6% | 2023.5 |
54
- | LLaMA 2 70B(zero-shot) | 29.9% | 2023.7 |
55
 
56
  <br>
57
 
@@ -93,7 +70,6 @@ pip install -r requirements.txt
93
  ```
94
 
95
  ```python
96
- import torch
97
  from transformers import (
98
  AutoTokenizer,
99
  AutoModelForCausalLM,
@@ -102,13 +78,7 @@ tokenizer = AutoTokenizer.from_pretrained(mode_name_or_path, trust_remote_code=T
102
  tokenizer.padding_side = "left"
103
  tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids("<unk>")
104
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids("</s>")
105
- # try 4bit loading if cuda memory not enough
106
- model = AutoModelForCausalLM.from_pretrained(mode_name_or_path,
107
- trust_remote_code=True,
108
- load_in_4bit=False,
109
- device_map="auto",
110
- torch_dtype=torch.bfloat16)
111
- model.eval()
112
 
113
  HUMAN_ROLE_START_TAG = "<|role_start|>human<|role_end|>"
114
  BOT_ROLE_START_TAG = "<|role_start|>bot<|role_end|>"
@@ -129,32 +99,17 @@ gen_text = tokenizer.batch_decode(outputs[:, inputs["input_ids"].shape[1]:], ski
129
  print(gen_text)
130
  ```
131
 
132
- ## MD5
133
- We notice that the file may be corrupted during transfer process. Please check MD5 value before use.
134
 
135
- | Model File | MD5 Value |
136
- |:---------------------------------|:--------------------------------:|
137
- | pytorch_model-00001-of-00007.bin | 8d544b1bcb3449934184d4141137329c |
138
- | pytorch_model-00002-of-00007.bin | 9d5dbb30911e48a42fb6d0fcabb322a4 |
139
- | pytorch_model-00003-of-00007.bin | b0d4aecee0457d9332005a187e1fffed |
140
- | pytorch_model-00004-of-00007.bin | 5c7e002de5eab77d0194a2b0f6de0c24 |
141
- | pytorch_model-00005-of-00007.bin | d22a511aa26b5b17117b665a877490ab |
142
- | pytorch_model-00006-of-00007.bin | a5c28ac277fac07d16dd66537e54d109 |
143
- | pytorch_model-00007-of-00007.bin | a967e2c6195477b7407089c0bffa2d53 |
144
 
145
 
146
- ## Citation
147
- If you find our [work](https://arxiv.org/abs/2311.02303) useful or helpful for your R&D works, please feel free to cite our paper as below.
148
- ```
149
- @article{mftcoder2023,
150
- title={MFTCoder: Boosting Code LLMs with Multitask Fine-Tuning},
151
- author={Bingchang Liu and Chaoyu Chen and Cong Liao and Zi Gong and Huan Wang and Zhichao Lei and Ming Liang and Dajun Chen and Min Shen and Hailian Zhou and Hang Yu and Jianguo Li},
152
- year={2023},
153
- journal={arXiv preprint arXiv},
154
- archivePrefix={arXiv},
155
- eprint={2311.02303}
156
- }
157
- ```
158
 
159
  <a id="chinese"></a>
160
 
@@ -169,31 +124,13 @@ CodeFuse-CodeLlama34B-MFT 是一个通过QLoRA对基座模型CodeLlama-34b-Pytho
169
 
170
  <br>
171
 
172
- ## 代码社区
173
- **大本营**: 🏡 https://github.com/codefuse-ai (**欢迎为我们的项目一键三连 Star🌟 + Fork🚀 + Watch👀**)
174
-
175
- + 如果您想自己微调该模型,可以访问 ✨[MFTCoder](https://github.com/codefuse-ai/MFTCoder)✨✨
176
-
177
- + 如果您想自己部署该模型,可以访问 ✨[FasterTransformer4CodeFuse](https://github.com/codefuse-ai/FasterTransformer4CodeFuse)✨✨
178
-
179
- + 如果您想观看该模型示例,可以访问 ✨[CodeFuse Demo](https://github.com/codefuse-ai/codefuse)✨✨
180
-
181
-
182
  ## 评测表现(代码)
183
 
184
- | 模型 | HumanEval(pass@1) | 日期 |
185
- |:----------------------------|:-----------------:|:-------:|
186
- | **CodeFuse-CodeLlama-34B** | **74.4%** | 2023.9 |
187
- | WizardCoder-Python-34B-V1.0 | 73.2% | 2023.8 |
188
- | GPT-4(zero-shot) | 67.0% | 2023.3 |
189
- | PanGu-Coder2 15B | 61.6% | 2023.8 |
190
- | CodeLlama-34b-Python | 53.7% | 2023.8 |
191
- | CodeLlama-34b | 48.8% | 2023.8 |
192
- | GPT-3.5(zero-shot) | 48.1% | 2022.11 |
193
- | OctoCoder | 46.2% | 2023.8 |
194
- | StarCoder-15B | 33.6% | 2023.5 |
195
- | LLaMA 2 70B(zero-shot) | 29.9% | 2023.7 |
196
-
197
  <br>
198
 
199
  ## Requirements
@@ -236,18 +173,12 @@ tokenizer = AutoTokenizer.from_pretrained(mode_name_or_path, trust_remote_code=T
236
  tokenizer.padding_side = "left"
237
  tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids("<unk>")
238
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids("</s>")
239
- # 如果显存不够,可以考虑量化加载
240
- model = AutoModelForCausalLM.from_pretrained(mode_name_or_path,
241
- trust_remote_code=True,
242
- load_in_4bit=False,
243
- device_map="auto",
244
- torch_dtype=torch.bfloat16)
245
- model.eval()
246
 
247
  HUMAN_ROLE_START_TAG = "<|role_start|>human<|role_end|>"
248
  BOT_ROLE_START_TAG = "<|role_start|>bot<|role_end|>"
249
 
250
- text = f"{HUMAN_ROLE_START_TAG}请用C++实现求解第n个斐波那契数{BOT_ROLE_START_TAG}"
251
  inputs = tokenizer(text, return_tensors='pt', padding=True, add_special_tokens=False).to("cuda")
252
  outputs = model.generate(
253
  inputs=inputs["input_ids"],
@@ -261,18 +192,4 @@ outputs = model.generate(
261
  )
262
  gen_text = tokenizer.batch_decode(outputs[:, inputs["input_ids"].shape[1]:], skip_special_tokens=True)
263
  print(gen_text)
264
- ```
265
-
266
-
267
- ## MD5
268
- 我们发现模型文件可能会在传输过程中损坏,使用前请检查文件MD5值。
269
-
270
- | 模型文件 | MD5值 |
271
- |:---------------------------------|:--------------------------------:|
272
- | pytorch_model-00001-of-00007.bin | 8d544b1bcb3449934184d4141137329c |
273
- | pytorch_model-00002-of-00007.bin | 9d5dbb30911e48a42fb6d0fcabb322a4 |
274
- | pytorch_model-00003-of-00007.bin | b0d4aecee0457d9332005a187e1fffed |
275
- | pytorch_model-00004-of-00007.bin | 5c7e002de5eab77d0194a2b0f6de0c24 |
276
- | pytorch_model-00005-of-00007.bin | d22a511aa26b5b17117b665a877490ab |
277
- | pytorch_model-00006-of-00007.bin | a5c28ac277fac07d16dd66537e54d109 |
278
- | pytorch_model-00007-of-00007.bin | a967e2c6195477b7407089c0bffa2d53 |
 
1
  ---
2
  license: other
 
 
3
  ---
4
  # Model Card for CodeFuse-CodeLlama-34B
5
+ ![Creation Approach](LOGO.png)
6
 
7
  [[中文]](#chinese) [[English]](#english)
8
 
 
18
 
19
  ## News and Updates
20
 
21
+ 🔥🔥🔥 CodeFuse-CodeLlama34B-MFT has achived 74.4% of pass@1 on HumanEval, which is SOTA at present.
 
 
22
 
23
  <br>
24
 
 
 
 
 
 
 
 
 
 
 
 
25
  ## Performance
26
 
27
+ | Model | HumanEval(pass@1) |
28
+ | :---------------------------- | :---------------: |
29
+ | CodeLlama-34b | 48.8%(greedy decoding) |
30
+ | CodeLlama-34b-Python | 53.7%(greedy decoding) |
31
+ | **CodeFuse-CodeLlama-34B** | **74.4%**(greedy decoding) |
 
 
 
 
 
 
 
 
32
 
33
  <br>
34
 
 
70
  ```
71
 
72
  ```python
 
73
  from transformers import (
74
  AutoTokenizer,
75
  AutoModelForCausalLM,
 
78
  tokenizer.padding_side = "left"
79
  tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids("<unk>")
80
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids("</s>")
81
+ model = AutoModelForCausalLM.from_pretrained(mode_name_or_path, trust_remote_code=True)
 
 
 
 
 
 
82
 
83
  HUMAN_ROLE_START_TAG = "<|role_start|>human<|role_end|>"
84
  BOT_ROLE_START_TAG = "<|role_start|>bot<|role_end|>"
 
99
  print(gen_text)
100
  ```
101
 
 
 
102
 
 
 
 
 
 
 
 
 
 
103
 
104
 
105
+
106
+
107
+
108
+
109
+
110
+
111
+
112
+
 
 
 
 
113
 
114
  <a id="chinese"></a>
115
 
 
124
 
125
  <br>
126
 
 
 
 
 
 
 
 
 
 
 
127
  ## 评测表现(代码)
128
 
129
+ | 模型 | HumanEval(pass@1) |
130
+ | :---------------------------- | :---------------: |
131
+ | CodeLlama-34b | 48.8%(greedy decoding) |
132
+ | CodeLlama-34b-Python | 53.7%(greedy decoding) |
133
+ | **CodeFuse-CodeLlama-34B** | **74.4%**(greedy decoding) |
 
 
 
 
 
 
 
 
134
  <br>
135
 
136
  ## Requirements
 
173
  tokenizer.padding_side = "left"
174
  tokenizer.pad_token_id = tokenizer.convert_tokens_to_ids("<unk>")
175
  tokenizer.eos_token_id = tokenizer.convert_tokens_to_ids("</s>")
176
+ model = AutoModelForCausalLM.from_pretrained(mode_name_or_path, trust_remote_code=True)
 
 
 
 
 
 
177
 
178
  HUMAN_ROLE_START_TAG = "<|role_start|>human<|role_end|>"
179
  BOT_ROLE_START_TAG = "<|role_start|>bot<|role_end|>"
180
 
181
+ text = f"{HUMAN_ROLE_START_TAG}write a python function of quick sort.{BOT_ROLE_START_TAG}"
182
  inputs = tokenizer(text, return_tensors='pt', padding=True, add_special_tokens=False).to("cuda")
183
  outputs = model.generate(
184
  inputs=inputs["input_ids"],
 
192
  )
193
  gen_text = tokenizer.batch_decode(outputs[:, inputs["input_ids"].shape[1]:], skip_special_tokens=True)
194
  print(gen_text)
195
+ ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json DELETED
@@ -1,29 +0,0 @@
1
- {
2
- "_name_or_path": "/mnt/user/qumu/download_models/codellama/CodeLlama-34b-Python-hf",
3
- "architectures": [
4
- "LlamaForCausalLM"
5
- ],
6
- "bos_token_id": 1,
7
- "eos_token": "</s>",
8
- "eos_token_id": 2,
9
- "hidden_act": "silu",
10
- "hidden_size": 8192,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 22016,
13
- "max_position_embeddings": 16384,
14
- "model_type": "llama",
15
- "num_attention_heads": 64,
16
- "num_hidden_layers": 48,
17
- "num_key_value_heads": 8,
18
- "pad_token": "<unk>",
19
- "pad_token_id": 0,
20
- "pretraining_tp": 1,
21
- "rms_norm_eps": 1e-05,
22
- "rope_scaling": null,
23
- "rope_theta": 1000000,
24
- "tie_word_embeddings": false,
25
- "torch_dtype": "bfloat16",
26
- "transformers_version": "4.32.0",
27
- "use_cache": true,
28
- "vocab_size": 32000
29
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
generation_config.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "bos_token_id": 1,
4
- "eos_token_id": 2,
5
- "transformers_version": "4.32.0"
6
- }
 
 
 
 
 
 
 
pytorch_model-00001-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb1cd3c944b6dd2ece21c09d7378a77ccc279867dfe4b8b8b7cbfa58e4ffe4a3
3
- size 9852638393
 
 
 
 
pytorch_model-00002-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1641f834b6d3938353873d95b642d0083ca806b7fd32dc8aaea0b3c64752b358
3
- size 9689094033
 
 
 
 
pytorch_model-00003-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:38ce1d749b110332ffbfc237c3271c10229f0913ed3e424aa61c4261af033357
3
- size 9689094033
 
 
 
 
pytorch_model-00004-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5c7a36d9f48fdd4b629f44d6e238085fa727f393c94e6cdf17d4f5da19be613
3
- size 9689094033
 
 
 
 
pytorch_model-00005-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ccdaa07cce1d16ed6d8ca1ca773b075f61c36d04430f8683b6e2bf458be4a72
3
- size 9689094033
 
 
 
 
pytorch_model-00006-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a257280910a701168406cc2114a669a975176ec08a373f7c2b7ba94d22c05215
3
- size 9689094033
 
 
 
 
pytorch_model-00007-of-00007.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:aaef8ce33ccbc255d78026458d81fc9a55513a0d1ac3ef4f4f74967300af49e7
3
- size 9189986713
 
 
 
 
pytorch_model.bin.index.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a714170172282cfbcaa120af13c0df08b06d040ff24dab30229d8a010821d3d
3
- size 35838
 
 
 
 
requirements.txt DELETED
@@ -1,14 +0,0 @@
1
- numpy
2
- pandas
3
- einops
4
- sentencepiece
5
- deepspeed==0.9.3
6
- transformers==4.32.0
7
- accelerate==0.21.0
8
- peft==0.4.0
9
- BitsAndBytes==0.40.2
10
- xformers==0.0.21
11
- ujson
12
- jsonlines
13
- tiktoken
14
- transformers_stream_generator
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
special_tokens_map.json DELETED
@@ -1,23 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "</s>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "unk_token": {
17
- "content": "<unk>",
18
- "lstrip": false,
19
- "normalized": true,
20
- "rstrip": false,
21
- "single_word": false
22
- }
23
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
tokenizer.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
- size 499723
 
 
 
 
tokenizer_config.json DELETED
@@ -1,33 +0,0 @@
1
- {
2
- "bos_token": {
3
- "__type": "AddedToken",
4
- "content": "<s>",
5
- "lstrip": false,
6
- "normalized": true,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- "clean_up_tokenization_spaces": false,
11
- "eos_token": {
12
- "__type": "AddedToken",
13
- "content": "</s>",
14
- "lstrip": false,
15
- "normalized": true,
16
- "rstrip": false,
17
- "single_word": false
18
- },
19
- "legacy": null,
20
- "model_max_length": 1000000000000000019884624838656,
21
- "pad_token": null,
22
- "sp_model_kwargs": {},
23
- "tokenizer_class": "LlamaTokenizer",
24
- "unk_token": {
25
- "__type": "AddedToken",
26
- "content": "<unk>",
27
- "lstrip": false,
28
- "normalized": true,
29
- "rstrip": false,
30
- "single_word": false
31
- },
32
- "use_default_system_prompt": true
33
- }