Upload folder using huggingface_hub
Browse files- DolphinProf.code-workspace +11 -0
- README.md +1 -1
- added_tokens.json +4 -0
- config.json +29 -0
- mergekit_config.yml +47 -0
- model-00001-of-00032.safetensors +3 -0
- model-00002-of-00032.safetensors +3 -0
- model-00003-of-00032.safetensors +3 -0
- model-00004-of-00032.safetensors +3 -0
- model-00005-of-00032.safetensors +3 -0
- model-00006-of-00032.safetensors +3 -0
- model-00007-of-00032.safetensors +3 -0
- model-00008-of-00032.safetensors +3 -0
- model-00009-of-00032.safetensors +3 -0
- model-00010-of-00032.safetensors +3 -0
- model-00011-of-00032.safetensors +3 -0
- model-00012-of-00032.safetensors +3 -0
- model-00013-of-00032.safetensors +3 -0
- model-00014-of-00032.safetensors +3 -0
- model-00015-of-00032.safetensors +3 -0
- model-00016-of-00032.safetensors +3 -0
- model-00017-of-00032.safetensors +3 -0
- model-00018-of-00032.safetensors +3 -0
- model-00019-of-00032.safetensors +3 -0
- model-00020-of-00032.safetensors +3 -0
- model-00021-of-00032.safetensors +3 -0
- model-00022-of-00032.safetensors +3 -0
- model-00023-of-00032.safetensors +3 -0
- model-00024-of-00032.safetensors +3 -0
- model-00025-of-00032.safetensors +3 -0
- model-00026-of-00032.safetensors +3 -0
- model-00027-of-00032.safetensors +3 -0
- model-00028-of-00032.safetensors +3 -0
- model-00029-of-00032.safetensors +3 -0
- model-00030-of-00032.safetensors +3 -0
- model-00031-of-00032.safetensors +3 -0
- model-00032-of-00032.safetensors +3 -0
- model.safetensors.index.json +0 -0
- special_tokens_map.json +30 -0
- tokenizer.json +0 -0
- tokenizer.model +3 -0
- tokenizer_config.json +65 -0
DolphinProf.code-workspace
ADDED
@@ -0,0 +1,11 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"folders": [
|
3 |
+
{
|
4 |
+
"path": "."
|
5 |
+
},
|
6 |
+
{
|
7 |
+
"path": "../../git/mergekit"
|
8 |
+
}
|
9 |
+
],
|
10 |
+
"settings": {}
|
11 |
+
}
|
README.md
CHANGED
@@ -81,7 +81,7 @@ dtype: float16
|
|
81 |
tokenizer_source: model:cognitivecomputations/dolphin-2.2-70b # keep exact tokenizer used by dolphin - or you could use `union` if you add all of the input models to the first/last slice, but they would need to be non-zero weight or you'll get NaNs in your embeddings
|
82 |
```
|
83 |
|
84 |
-
# Example Output
|
85 |
|
86 |
> Please invent a new idea in the area of mathematics, that combines two or more papers into a new idea that has not yet been published to your knowledge
|
87 |
## Integrability and Computability of Continuous Functions on Infinite Domains: A Category-Theoretical Approach
|
|
|
81 |
tokenizer_source: model:cognitivecomputations/dolphin-2.2-70b # keep exact tokenizer used by dolphin - or you could use `union` if you add all of the input models to the first/last slice, but they would need to be non-zero weight or you'll get NaNs in your embeddings
|
82 |
```
|
83 |
|
84 |
+
# Example Output
|
85 |
|
86 |
> Please invent a new idea in the area of mathematics, that combines two or more papers into a new idea that has not yet been published to your knowledge
|
87 |
## Integrability and Computability of Continuous Functions on Infinite Domains: A Category-Theoretical Approach
|
added_tokens.json
ADDED
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"<|im_end|>": 32000,
|
3 |
+
"<|im_start|>": 32001
|
4 |
+
}
|
config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "cognitivecomputations/dolphin-2.2-70b",
|
3 |
+
"architectures": [
|
4 |
+
"LlamaForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_bias": false,
|
7 |
+
"attention_dropout": 0.0,
|
8 |
+
"bos_token_id": 1,
|
9 |
+
"eos_token_id": 32000,
|
10 |
+
"hidden_act": "silu",
|
11 |
+
"hidden_size": 8192,
|
12 |
+
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 28672,
|
14 |
+
"max_position_embeddings": 4096,
|
15 |
+
"model_type": "llama",
|
16 |
+
"num_attention_heads": 64,
|
17 |
+
"num_hidden_layers": 180,
|
18 |
+
"num_key_value_heads": 8,
|
19 |
+
"pad_token_id": 0,
|
20 |
+
"pretraining_tp": 1,
|
21 |
+
"rms_norm_eps": 1e-05,
|
22 |
+
"rope_scaling": null,
|
23 |
+
"rope_theta": 10000.0,
|
24 |
+
"tie_word_embeddings": false,
|
25 |
+
"torch_dtype": "float16",
|
26 |
+
"transformers_version": "4.36.2",
|
27 |
+
"use_cache": true,
|
28 |
+
"vocab_size": 32002
|
29 |
+
}
|
mergekit_config.yml
ADDED
@@ -0,0 +1,47 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
merge_method: linear # use linear so we can include multiple models, albeit at a zero weight
|
2 |
+
parameters:
|
3 |
+
weight: 1.0 # weight everything as 1 unless specified otherwise - linear with one model weighted at 1 is a no-op like passthrough
|
4 |
+
slices:
|
5 |
+
- sources:
|
6 |
+
- model: cognitivecomputations/dolphin-2.2-70b # embed_tokens comes along with the ride with whatever is the first layer
|
7 |
+
layer_range: [0, 1]
|
8 |
+
- model: migtissera/SynthIA-70B-v1.2b # add dummy second model with 0 weight so tokenizer-based merge routine is invoked for embed_tokens
|
9 |
+
layer_range: [0, 1]
|
10 |
+
parameters:
|
11 |
+
weight: 0
|
12 |
+
- sources:
|
13 |
+
- model: cognitivecomputations/dolphin-2.2-70b
|
14 |
+
layer_range: [1, 20]
|
15 |
+
- sources:
|
16 |
+
- model: migtissera/SynthIA-70B-v1.2b
|
17 |
+
layer_range: [10, 30]
|
18 |
+
- sources:
|
19 |
+
- model: WizardLM/WizardMath-70B-V1.0
|
20 |
+
layer_range: [20, 40]
|
21 |
+
- sources:
|
22 |
+
- model: epfl-llm/meditron-70b
|
23 |
+
layer_range: [25, 45]
|
24 |
+
- sources:
|
25 |
+
- model: cognitivecomputations/dolphin-2.2-70b
|
26 |
+
layer_range: [30, 50]
|
27 |
+
- sources:
|
28 |
+
- model: migtissera/SynthIA-70B-v1.2b
|
29 |
+
layer_range: [40, 60]
|
30 |
+
- sources:
|
31 |
+
- model: WizardLM/WizardMath-70B-V1.0
|
32 |
+
layer_range: [50, 70]
|
33 |
+
- sources:
|
34 |
+
- model: epfl-llm/meditron-70b
|
35 |
+
layer_range: [55, 75]
|
36 |
+
- sources:
|
37 |
+
- model: cognitivecomputations/dolphin-2.2-70b
|
38 |
+
layer_range: [60, 79]
|
39 |
+
- sources: # same as above, but for lm_head with the last layer
|
40 |
+
- model: cognitivecomputations/dolphin-2.2-70b
|
41 |
+
layer_range: [79, 80]
|
42 |
+
- model: migtissera/SynthIA-70B-v1.2b
|
43 |
+
layer_range: [79, 80]
|
44 |
+
parameters:
|
45 |
+
weight: 0
|
46 |
+
dtype: float16
|
47 |
+
tokenizer_source: model:cognitivecomputations/dolphin-2.2-70b # keep exact tokenizer used by dolphin - or you could use `union` if you add all of the input models to the first/last slice, but they would need to be non-zero weight or you'll get NaNs in your embeddings
|
model-00001-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dc1f515376643696b5509f8c984f6557a4f50117a2bc8e532a36bbee2fd738f1
|
3 |
+
size 9986840424
|
model-00002-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d975a85c77ed7aebba0df45caa990c725c58db0fadd781fed5d43cc8c66dcc95
|
3 |
+
size 9798081088
|
model-00003-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e347cad10be6ebcbb3273abe6a4e614ddc806a8388069750cda86698489fa87d
|
3 |
+
size 9965852200
|
model-00004-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:193eeb23b10ccbf5d2befaea8c0cc891d257a913e4226a57de3194f299d49277
|
3 |
+
size 9798080424
|
model-00005-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c7093ed30bca6efcedd46f076215aa0c7c876c9b44f6f2bb98120fd04625e942
|
3 |
+
size 9965868720
|
model-00006-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bceceb6a3b836e06656f23a715c1a3454a77170c049522301e85055f39225179
|
3 |
+
size 9965868720
|
model-00007-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2a4e533dfd31f8cb4cccf0d569c0945962cd46a5e1133e5da8a12f8dfe4a87f9
|
3 |
+
size 9798096912
|
model-00008-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dcacf3df594becf74fdd69d95e03a3429b31ef1a06752263977e840ccda33c70
|
3 |
+
size 9798096896
|
model-00009-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df5a71dff2d98ce1c34f7b485f9d98960c0ec37e92d4efc72665da7d45a2c744
|
3 |
+
size 9965852216
|
model-00010-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:93f5421e951030b5922a1bdc1b75b30254176be43d529a28b88189b8b65ff1de
|
3 |
+
size 9630275624
|
model-00011-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:54323e93a709646377f75b9dcbae9c385610fb80631096bf31792619259ccae2
|
3 |
+
size 9798096936
|
model-00012-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0f471aa32b901dcc790cb03a4e0c207cfd867a6a5697172837c76732b594ca34
|
3 |
+
size 9798096952
|
model-00013-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d4f96305e454995b053b83ae709e915a17025b1e7c65d58875795c98a399e123
|
3 |
+
size 9798096920
|
model-00014-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:150770b4314788a1b4898a2fa7fb7353fd2d9d67657b6796aeda2252c10d3d76
|
3 |
+
size 9965868752
|
model-00015-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:320e9c280b375d4010c21e3552b17efadc87c2697e7f4b3ab750ee95a259452e
|
3 |
+
size 9965885240
|
model-00016-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:63a786826334cf4cbe07c702e1e2a0a510b82a186027e23689faa9aefd15bd23
|
3 |
+
size 9798080424
|
model-00017-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c56e77028142b50084eb3a40cd3d25c4d3259fa1f604a55755032e87490f002
|
3 |
+
size 9798063944
|
model-00018-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fef28e5d83623d0e8cf1ae02dc676b9e48c581a098a89456d1627009f5e4be8b
|
3 |
+
size 9663846072
|
model-00019-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8161098b800cbc962bc5f2df1a40288b7cd51227ba193f553b3ced0e57890641
|
3 |
+
size 9798096952
|
model-00020-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:de924bb285bf6fa16525f329f6511a6c4f2ce3959790c750784729b717cc1abe
|
3 |
+
size 9798129904
|
model-00021-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6b19593608d51b4968eb21984f07ee54804b4c32a226acabda22ca4ab8fa68be
|
3 |
+
size 9630291448
|
model-00022-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8d54fe3a7a938ae6a810485fe57b9d3e68071ce1a22d869833f13515bc832a07
|
3 |
+
size 9798096936
|
model-00023-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:977577794d614156f1ae932c824ae47935fb26f8a7b913d66e5fcd5c27a86ec9
|
3 |
+
size 9965835752
|
model-00024-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e17c6b62353b784d2e2f8f8d6963a24a79bde5507d4bf02dc618b6156e14ecd7
|
3 |
+
size 9798096952
|
model-00025-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1a4297e6519fa8af018c446d32ee206d06f558336ff6c23688f62303095a724
|
3 |
+
size 9965852944
|
model-00026-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01db6dcb0b659323e8a28171a79ca25174ccd83c7756c5231fa49907cae77176
|
3 |
+
size 9798113448
|
model-00027-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74fdc329c94d413cdb28953d036693a134f033b1b57c35f760e2c6df8e3c20f1
|
3 |
+
size 9965885232
|
model-00028-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:085a0d37413e40bc61acdfb2fb2f2ddfd39a7661eb950ba62d3dfc2044c49b19
|
3 |
+
size 9550665264
|
model-00029-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:38553cc54e82a00fa231b3356339e5f0058a1878d38150938729d3216a3ef58a
|
3 |
+
size 9630274944
|
model-00030-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1cef9fa248f4468f2402225c0b4ab4d03b5c19ca8db75ebd8281850f7cd93684
|
3 |
+
size 9965885976
|
model-00031-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0e8e9db59eb0ef2aa207e34d4fc85e0c0123030e27c08c6de29154ca58dbff8b
|
3 |
+
size 9798079744
|
model-00032-of-00032.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ed036ef34236715773ce4bdd4712af2f2114c2308544e3d5c8dbd7e46f3c1d43
|
3 |
+
size 4362177048
|
model.safetensors.index.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
special_tokens_map.json
ADDED
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": false,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "<|im_end|>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": false,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"pad_token": {
|
17 |
+
"content": "</s>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": false,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
},
|
23 |
+
"unk_token": {
|
24 |
+
"content": "<unk>",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
}
|
30 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
|
3 |
+
size 499723
|
tokenizer_config.json
ADDED
@@ -0,0 +1,65 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"0": {
|
6 |
+
"content": "<unk>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"1": {
|
14 |
+
"content": "<s>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": false,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
},
|
21 |
+
"2": {
|
22 |
+
"content": "</s>",
|
23 |
+
"lstrip": false,
|
24 |
+
"normalized": false,
|
25 |
+
"rstrip": false,
|
26 |
+
"single_word": false,
|
27 |
+
"special": true
|
28 |
+
},
|
29 |
+
"32000": {
|
30 |
+
"content": "<|im_end|>",
|
31 |
+
"lstrip": false,
|
32 |
+
"normalized": false,
|
33 |
+
"rstrip": false,
|
34 |
+
"single_word": false,
|
35 |
+
"special": true
|
36 |
+
},
|
37 |
+
"32001": {
|
38 |
+
"content": "<|im_start|>",
|
39 |
+
"lstrip": false,
|
40 |
+
"normalized": false,
|
41 |
+
"rstrip": false,
|
42 |
+
"single_word": false,
|
43 |
+
"special": false
|
44 |
+
}
|
45 |
+
},
|
46 |
+
"bos_token": "<s>",
|
47 |
+
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
48 |
+
"clean_up_tokenization_spaces": false,
|
49 |
+
"eos_token": "<|im_end|>",
|
50 |
+
"legacy": false,
|
51 |
+
"max_length": 2048,
|
52 |
+
"model_max_length": 1000000000000000019884624838656,
|
53 |
+
"pad_token": "</s>",
|
54 |
+
"padding_side": "right",
|
55 |
+
"sp_model_kwargs": {},
|
56 |
+
"spaces_between_special_tokens": false,
|
57 |
+
"stride": 0,
|
58 |
+
"tokenizer_class": "LlamaTokenizer",
|
59 |
+
"truncation_side": "right",
|
60 |
+
"truncation_strategy": "longest_first",
|
61 |
+
"trust_remote_code": false,
|
62 |
+
"unk_token": "<unk>",
|
63 |
+
"use_default_system_prompt": true,
|
64 |
+
"use_fast": true
|
65 |
+
}
|