init
#1
by
hushell
- opened
- Qwen1.5-1.8B-Chat-q4f16_1-android.tar +3 -0
- merges.txt +0 -0
- mlc-chat-config.json +47 -0
- ndarray-cache.json +0 -0
- params_shard_0.bin +3 -0
- params_shard_1.bin +3 -0
- params_shard_10.bin +3 -0
- params_shard_11.bin +3 -0
- params_shard_12.bin +3 -0
- params_shard_13.bin +3 -0
- params_shard_14.bin +3 -0
- params_shard_15.bin +3 -0
- params_shard_16.bin +3 -0
- params_shard_17.bin +3 -0
- params_shard_18.bin +3 -0
- params_shard_19.bin +3 -0
- params_shard_2.bin +3 -0
- params_shard_20.bin +3 -0
- params_shard_21.bin +3 -0
- params_shard_22.bin +3 -0
- params_shard_23.bin +3 -0
- params_shard_24.bin +3 -0
- params_shard_25.bin +3 -0
- params_shard_26.bin +3 -0
- params_shard_27.bin +3 -0
- params_shard_3.bin +3 -0
- params_shard_4.bin +3 -0
- params_shard_5.bin +3 -0
- params_shard_6.bin +3 -0
- params_shard_7.bin +3 -0
- params_shard_8.bin +3 -0
- params_shard_9.bin +3 -0
- tokenizer.json +0 -0
- tokenizer_config.json +40 -0
- vocab.json +0 -0
Qwen1.5-1.8B-Chat-q4f16_1-android.tar
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6962d97579661f5cb30146370e578c44cb7c2646f5bd6cd5697d7eeb917d95d7
|
3 |
+
size 293781
|
merges.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|
mlc-chat-config.json
ADDED
@@ -0,0 +1,47 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"model_type": "qwen2",
|
3 |
+
"quantization": "q4f16_1",
|
4 |
+
"model_config": {
|
5 |
+
"hidden_act": "silu",
|
6 |
+
"hidden_size": 2048,
|
7 |
+
"intermediate_size": 5504,
|
8 |
+
"num_attention_heads": 16,
|
9 |
+
"num_hidden_layers": 24,
|
10 |
+
"num_key_value_heads": 16,
|
11 |
+
"rms_norm_eps": 1e-06,
|
12 |
+
"rope_theta": 1000000.0,
|
13 |
+
"vocab_size": 151936,
|
14 |
+
"context_window_size": 32768,
|
15 |
+
"prefill_chunk_size": 32768,
|
16 |
+
"tensor_parallel_shards": 1,
|
17 |
+
"dtype": "float32"
|
18 |
+
},
|
19 |
+
"vocab_size": 151936,
|
20 |
+
"context_window_size": 32768,
|
21 |
+
"sliding_window_size": -1,
|
22 |
+
"prefill_chunk_size": 32768,
|
23 |
+
"attention_sink_size": -1,
|
24 |
+
"tensor_parallel_shards": 1,
|
25 |
+
"mean_gen_len": 128,
|
26 |
+
"max_gen_len": 512,
|
27 |
+
"shift_fill_factor": 0.3,
|
28 |
+
"temperature": 0.7,
|
29 |
+
"presence_penalty": 0.0,
|
30 |
+
"frequency_penalty": 0.0,
|
31 |
+
"repetition_penalty": 1.1,
|
32 |
+
"top_p": 0.8,
|
33 |
+
"conv_template": "chatml",
|
34 |
+
"pad_token_id": 0,
|
35 |
+
"bos_token_id": 151643,
|
36 |
+
"eos_token_id": [
|
37 |
+
151645,
|
38 |
+
151643
|
39 |
+
],
|
40 |
+
"tokenizer_files": [
|
41 |
+
"tokenizer.json",
|
42 |
+
"vocab.json",
|
43 |
+
"merges.txt",
|
44 |
+
"tokenizer_config.json"
|
45 |
+
],
|
46 |
+
"version": "0.1.0"
|
47 |
+
}
|
ndarray-cache.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
params_shard_0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae6aa4ce63ed1cf62dfc6a457ca9ac7013289b6e10eb0e0a02ea9159e7f6684f
|
3 |
+
size 155582464
|
params_shard_1.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b8ec50cbe05eeca11a829392fc8bb941ef9c95b95fb5f5c001b9dfc5fbb5aead
|
3 |
+
size 155582464
|
params_shard_10.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53750acd7037ec4e3a4460c813d01946e68371cfd4b0c8b7c7ec2a76ffd1fe4a
|
3 |
+
size 28479488
|
params_shard_11.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:55180778e5c91d059e8784b3b12e58124d5dd0613f31d288255083f3b2c53512
|
3 |
+
size 28479488
|
params_shard_12.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:77a033ed91100b080c08af057df821e4870a634640ac085fc78302401829af24
|
3 |
+
size 28479488
|
params_shard_13.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9359698625fe747400352f70d48b604300c089e0a729855e393573912a124378
|
3 |
+
size 28479488
|
params_shard_14.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37e689a4759f3979596a3cb33406e9312d8a2ab7b6e4eaa59a95947e5a76a148
|
3 |
+
size 28479488
|
params_shard_15.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5800565941f7b0037447333837c3bd453faed915bb89acab3dedf3478756b929
|
3 |
+
size 28479488
|
params_shard_16.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5330b3094a220af833e18b29ced46d595b3fa448a903c90aa9160f351afecb74
|
3 |
+
size 28479488
|
params_shard_17.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dcb819e83d5581b85173d3d502d5731f22d76a982e6a67f1d94f5f774e82c6c9
|
3 |
+
size 28479488
|
params_shard_18.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3015e940a54b5cd12389f9313f100d49aa49992ac57eefc46195dd3ee62a215d
|
3 |
+
size 28479488
|
params_shard_19.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a18203704fc2dde723ccea35a779a5c578a45d83b4b3f264f76de6fe382f7153
|
3 |
+
size 28479488
|
params_shard_2.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:92d28e00d872e03fdca57076746f2440a921e43d383f0174808cb2319cfe5c8f
|
3 |
+
size 19447808
|
params_shard_20.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1373d6e1756f594ea446f90f3d22965b32efa6b8fe8b9da85af36c8efc89cc13
|
3 |
+
size 28479488
|
params_shard_21.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80cc3ba4f7415acbe32a019fc585045d872cb783358a9736af28e6262e62b08b
|
3 |
+
size 28479488
|
params_shard_22.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5d12b6bbb2a679c2f24105a9ab5a27f9261a5cf2bdaf63c1e3092ef5db73a6c9
|
3 |
+
size 28479488
|
params_shard_23.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:822abe2b881a3f38a2322ac50befb96901127a16c392f1a7ca191e77592e51cf
|
3 |
+
size 28479488
|
params_shard_24.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:19bd7acca55505b72c0633de14ce6c7e6146ec695ae431bd90a8768f0c468bf6
|
3 |
+
size 28479488
|
params_shard_25.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dfab2d1c0a69b55391b053691694e5fcf077b1ed39def57e9a11166afc30d921
|
3 |
+
size 28479488
|
params_shard_26.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:93fc48444d08796cd45d1d3d3a47e24d258c43baee79153a38adc5a9bc822ade
|
3 |
+
size 28479488
|
params_shard_27.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5781d8db5c3ab8a15e638fdb691cf46e49379403d305382b7977e0fed4653a13
|
3 |
+
size 22138880
|
params_shard_3.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:228a7d631f2a712e6b95a427c6e7ce8dc0cdd79c18f68e32b0467c9784d608b6
|
3 |
+
size 25792512
|
params_shard_4.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8fa7b51efaf5c0cf1987c60c5b4bf6e93942bb1eec1e4c5cd3736d243d76f8d7
|
3 |
+
size 28479488
|
params_shard_5.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bb5a5bbd348898aff64a369f18ef31a09566cd50ee9814fc00f088ca45dda657
|
3 |
+
size 28479488
|
params_shard_6.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a31d43ed3df84ca2bcc2b2b985bd844831c62a7b9dfc673019d5c42112ebdb00
|
3 |
+
size 28479488
|
params_shard_7.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5ce603434d9367601f312954c75e29782f9cb824cdb34757024e1b234e99eb21
|
3 |
+
size 28479488
|
params_shard_8.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:241eb9a89e0c10b09a8cf007b894f18ed1e152874e4fa2deddef948f421bc4a9
|
3 |
+
size 28479488
|
params_shard_9.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7161f5e71269622b5d137b69268111c17d079cfe326222cb1efdc03b0d664ff7
|
3 |
+
size 28479488
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_prefix_space": false,
|
3 |
+
"added_tokens_decoder": {
|
4 |
+
"151643": {
|
5 |
+
"content": "<|endoftext|>",
|
6 |
+
"lstrip": false,
|
7 |
+
"normalized": false,
|
8 |
+
"rstrip": false,
|
9 |
+
"single_word": false,
|
10 |
+
"special": true
|
11 |
+
},
|
12 |
+
"151644": {
|
13 |
+
"content": "<|im_start|>",
|
14 |
+
"lstrip": false,
|
15 |
+
"normalized": false,
|
16 |
+
"rstrip": false,
|
17 |
+
"single_word": false,
|
18 |
+
"special": true
|
19 |
+
},
|
20 |
+
"151645": {
|
21 |
+
"content": "<|im_end|>",
|
22 |
+
"lstrip": false,
|
23 |
+
"normalized": false,
|
24 |
+
"rstrip": false,
|
25 |
+
"single_word": false,
|
26 |
+
"special": true
|
27 |
+
}
|
28 |
+
},
|
29 |
+
"additional_special_tokens": ["<|im_start|>", "<|im_end|>"],
|
30 |
+
"bos_token": null,
|
31 |
+
"chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content']}}{% if (loop.last and add_generation_prompt) or not loop.last %}{{ '<|im_end|>' + '\n'}}{% endif %}{% endfor %}{% if add_generation_prompt and messages[-1]['role'] != 'assistant' %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
32 |
+
"clean_up_tokenization_spaces": false,
|
33 |
+
"eos_token": "<|im_end|>",
|
34 |
+
"errors": "replace",
|
35 |
+
"model_max_length": 32768,
|
36 |
+
"pad_token": "<|endoftext|>",
|
37 |
+
"split_special_tokens": false,
|
38 |
+
"tokenizer_class": "Qwen2Tokenizer",
|
39 |
+
"unk_token": null
|
40 |
+
}
|
vocab.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|