Qwen1.5-1.8B-Chat-q4f16_1-android.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fda822b347faa841780fc6cd1bb8f64d6333345dd4dadec110a62f19acee789
3
+ size 293716
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
mlc-chat-config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "qwen2",
3
+ "quantization": "q4f16_1",
4
+ "model_config": {
5
+ "hidden_act": "silu",
6
+ "hidden_size": 2048,
7
+ "intermediate_size": 5504,
8
+ "num_attention_heads": 16,
9
+ "num_hidden_layers": 24,
10
+ "num_key_value_heads": 16,
11
+ "rms_norm_eps": 1e-06,
12
+ "rope_theta": 1000000.0,
13
+ "vocab_size": 151936,
14
+ "context_window_size": 32768,
15
+ "prefill_chunk_size": 32768,
16
+ "tensor_parallel_shards": 1,
17
+ "dtype": "float32"
18
+ },
19
+ "vocab_size": 151936,
20
+ "context_window_size": 32768,
21
+ "sliding_window_size": -1,
22
+ "prefill_chunk_size": 32768,
23
+ "attention_sink_size": -1,
24
+ "tensor_parallel_shards": 1,
25
+ "mean_gen_len": 128,
26
+ "max_gen_len": 512,
27
+ "shift_fill_factor": 0.3,
28
+ "temperature": 0.7,
29
+ "presence_penalty": 0.0,
30
+ "frequency_penalty": 0.0,
31
+ "repetition_penalty": 1.1,
32
+ "top_p": 0.8,
33
+ "conv_template": "chatml",
34
+ "pad_token_id": 0,
35
+ "bos_token_id": 151643,
36
+ "eos_token_id": 151645,
37
+ "tokenizer_files": [
38
+ "tokenizer.json",
39
+ "vocab.json",
40
+ "merges.txt",
41
+ "tokenizer_config.json"
42
+ ],
43
+ "version": "0.1.0"
44
+ }
ndarray-cache.json ADDED
The diff for this file is too large to render. See raw diff
 
params_shard_0.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae6aa4ce63ed1cf62dfc6a457ca9ac7013289b6e10eb0e0a02ea9159e7f6684f
3
+ size 155582464
params_shard_1.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8ec50cbe05eeca11a829392fc8bb941ef9c95b95fb5f5c001b9dfc5fbb5aead
3
+ size 155582464
params_shard_10.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53750acd7037ec4e3a4460c813d01946e68371cfd4b0c8b7c7ec2a76ffd1fe4a
3
+ size 28479488
params_shard_11.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55180778e5c91d059e8784b3b12e58124d5dd0613f31d288255083f3b2c53512
3
+ size 28479488
params_shard_12.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77a033ed91100b080c08af057df821e4870a634640ac085fc78302401829af24
3
+ size 28479488
params_shard_13.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9359698625fe747400352f70d48b604300c089e0a729855e393573912a124378
3
+ size 28479488
params_shard_14.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37e689a4759f3979596a3cb33406e9312d8a2ab7b6e4eaa59a95947e5a76a148
3
+ size 28479488
params_shard_15.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5800565941f7b0037447333837c3bd453faed915bb89acab3dedf3478756b929
3
+ size 28479488
params_shard_16.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5330b3094a220af833e18b29ced46d595b3fa448a903c90aa9160f351afecb74
3
+ size 28479488
params_shard_17.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcb819e83d5581b85173d3d502d5731f22d76a982e6a67f1d94f5f774e82c6c9
3
+ size 28479488
params_shard_18.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3015e940a54b5cd12389f9313f100d49aa49992ac57eefc46195dd3ee62a215d
3
+ size 28479488
params_shard_19.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a18203704fc2dde723ccea35a779a5c578a45d83b4b3f264f76de6fe382f7153
3
+ size 28479488
params_shard_2.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92d28e00d872e03fdca57076746f2440a921e43d383f0174808cb2319cfe5c8f
3
+ size 19447808
params_shard_20.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1373d6e1756f594ea446f90f3d22965b32efa6b8fe8b9da85af36c8efc89cc13
3
+ size 28479488
params_shard_21.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80cc3ba4f7415acbe32a019fc585045d872cb783358a9736af28e6262e62b08b
3
+ size 28479488
params_shard_22.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d12b6bbb2a679c2f24105a9ab5a27f9261a5cf2bdaf63c1e3092ef5db73a6c9
3
+ size 28479488
params_shard_23.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:822abe2b881a3f38a2322ac50befb96901127a16c392f1a7ca191e77592e51cf
3
+ size 28479488
params_shard_24.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19bd7acca55505b72c0633de14ce6c7e6146ec695ae431bd90a8768f0c468bf6
3
+ size 28479488
params_shard_25.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfab2d1c0a69b55391b053691694e5fcf077b1ed39def57e9a11166afc30d921
3
+ size 28479488
params_shard_26.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93fc48444d08796cd45d1d3d3a47e24d258c43baee79153a38adc5a9bc822ade
3
+ size 28479488
params_shard_27.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5781d8db5c3ab8a15e638fdb691cf46e49379403d305382b7977e0fed4653a13
3
+ size 22138880
params_shard_3.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:228a7d631f2a712e6b95a427c6e7ce8dc0cdd79c18f68e32b0467c9784d608b6
3
+ size 25792512
params_shard_4.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fa7b51efaf5c0cf1987c60c5b4bf6e93942bb1eec1e4c5cd3736d243d76f8d7
3
+ size 28479488
params_shard_5.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb5a5bbd348898aff64a369f18ef31a09566cd50ee9814fc00f088ca45dda657
3
+ size 28479488
params_shard_6.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a31d43ed3df84ca2bcc2b2b985bd844831c62a7b9dfc673019d5c42112ebdb00
3
+ size 28479488
params_shard_7.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ce603434d9367601f312954c75e29782f9cb824cdb34757024e1b234e99eb21
3
+ size 28479488
params_shard_8.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:241eb9a89e0c10b09a8cf007b894f18ed1e152874e4fa2deddef948f421bc4a9
3
+ size 28479488
params_shard_9.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7161f5e71269622b5d137b69268111c17d079cfe326222cb1efdc03b0d664ff7
3
+ size 28479488
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "151643": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "151644": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "151645": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "additional_special_tokens": ["<|im_start|>", "<|im_end|>"],
30
+ "bos_token": null,
31
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content']}}{% if (loop.last and add_generation_prompt) or not loop.last %}{{ '<|im_end|>' + '\n'}}{% endif %}{% endfor %}{% if add_generation_prompt and messages[-1]['role'] != 'assistant' %}{{ '<|im_start|>assistant\n' }}{% endif %}",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "<|im_end|>",
34
+ "errors": "replace",
35
+ "model_max_length": 32768,
36
+ "pad_token": "<|endoftext|>",
37
+ "split_special_tokens": false,
38
+ "tokenizer_class": "Qwen2Tokenizer",
39
+ "unk_token": null
40
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff