Commit
·
3ca7262
1
Parent(s):
4dacd55
mlc
Browse files- config.json +38 -0
- generation_config.json +8 -0
- mlc-chat-config.json +83 -0
- model.safetensors +3 -0
- ndarray-cache.json +0 -0
- params_shard_0.bin +3 -0
- params_shard_1.bin +3 -0
- params_shard_10.bin +3 -0
- params_shard_11.bin +3 -0
- params_shard_12.bin +3 -0
- params_shard_13.bin +3 -0
- params_shard_14.bin +3 -0
- params_shard_15.bin +3 -0
- params_shard_16.bin +3 -0
- params_shard_17.bin +3 -0
- params_shard_18.bin +3 -0
- params_shard_19.bin +3 -0
- params_shard_2.bin +3 -0
- params_shard_20.bin +3 -0
- params_shard_21.bin +3 -0
- params_shard_22.bin +3 -0
- params_shard_23.bin +3 -0
- params_shard_24.bin +3 -0
- params_shard_25.bin +3 -0
- params_shard_26.bin +3 -0
- params_shard_27.bin +3 -0
- params_shard_28.bin +3 -0
- params_shard_29.bin +3 -0
- params_shard_3.bin +3 -0
- params_shard_30.bin +3 -0
- params_shard_31.bin +3 -0
- params_shard_32.bin +3 -0
- params_shard_33.bin +3 -0
- params_shard_34.bin +3 -0
- params_shard_35.bin +3 -0
- params_shard_36.bin +3 -0
- params_shard_4.bin +3 -0
- params_shard_5.bin +3 -0
- params_shard_6.bin +3 -0
- params_shard_7.bin +3 -0
- params_shard_8.bin +3 -0
- params_shard_9.bin +3 -0
- tokenizer_config.json +1 -1
config.json
ADDED
@@ -0,0 +1,38 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "unsloth/SmolLM2-1.7B-Instruct-bnb-4bit",
|
3 |
+
"architectures": [
|
4 |
+
"LlamaForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_bias": false,
|
7 |
+
"attention_dropout": 0.0,
|
8 |
+
"bos_token_id": 1,
|
9 |
+
"eos_token_id": 2,
|
10 |
+
"head_dim": 64,
|
11 |
+
"hidden_act": "silu",
|
12 |
+
"hidden_size": 2048,
|
13 |
+
"initializer_range": 0.02,
|
14 |
+
"intermediate_size": 8192,
|
15 |
+
"max_position_embeddings": 8192,
|
16 |
+
"mlp_bias": false,
|
17 |
+
"model_type": "llama",
|
18 |
+
"num_attention_heads": 32,
|
19 |
+
"num_hidden_layers": 24,
|
20 |
+
"num_key_value_heads": 32,
|
21 |
+
"pad_token_id": 49152,
|
22 |
+
"pretraining_tp": 1,
|
23 |
+
"rms_norm_eps": 1e-05,
|
24 |
+
"rope_scaling": null,
|
25 |
+
"rope_theta": 130000,
|
26 |
+
"tie_word_embeddings": true,
|
27 |
+
"torch_dtype": "bfloat16",
|
28 |
+
"transformers.js_config": {
|
29 |
+
"kv_cache_dtype": {
|
30 |
+
"fp16": "float16",
|
31 |
+
"q4f16": "float16"
|
32 |
+
}
|
33 |
+
},
|
34 |
+
"transformers_version": "4.48.0",
|
35 |
+
"unsloth_version": "2025.1.6",
|
36 |
+
"use_cache": true,
|
37 |
+
"vocab_size": 49153
|
38 |
+
}
|
generation_config.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 1,
|
4 |
+
"eos_token_id": 2,
|
5 |
+
"max_length": 8192,
|
6 |
+
"pad_token_id": 49152,
|
7 |
+
"transformers_version": "4.48.0"
|
8 |
+
}
|
mlc-chat-config.json
ADDED
@@ -0,0 +1,83 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"version": "0.1.0",
|
3 |
+
"model_type": "llama",
|
4 |
+
"quantization": "q4f16_0",
|
5 |
+
"model_config": {
|
6 |
+
"hidden_size": 2048,
|
7 |
+
"intermediate_size": 8192,
|
8 |
+
"num_attention_heads": 32,
|
9 |
+
"num_hidden_layers": 24,
|
10 |
+
"rms_norm_eps": 1e-05,
|
11 |
+
"vocab_size": 49153,
|
12 |
+
"tie_word_embeddings": true,
|
13 |
+
"position_embedding_base": 130000,
|
14 |
+
"rope_scaling": null,
|
15 |
+
"context_window_size": 8192,
|
16 |
+
"prefill_chunk_size": 8192,
|
17 |
+
"num_key_value_heads": 32,
|
18 |
+
"head_dim": 64,
|
19 |
+
"tensor_parallel_shards": 1,
|
20 |
+
"pipeline_parallel_stages": 1,
|
21 |
+
"max_batch_size": 128,
|
22 |
+
"disaggregation": false
|
23 |
+
},
|
24 |
+
"vocab_size": 49153,
|
25 |
+
"context_window_size": 8192,
|
26 |
+
"sliding_window_size": -1,
|
27 |
+
"prefill_chunk_size": 8192,
|
28 |
+
"attention_sink_size": -1,
|
29 |
+
"tensor_parallel_shards": 1,
|
30 |
+
"pipeline_parallel_stages": 1,
|
31 |
+
"temperature": 0.7,
|
32 |
+
"presence_penalty": 0.2,
|
33 |
+
"frequency_penalty": 0.2,
|
34 |
+
"repetition_penalty": 1.0,
|
35 |
+
"top_p": 0.75,
|
36 |
+
"tokenizer_files": [
|
37 |
+
"tokenizer.json",
|
38 |
+
"vocab.json",
|
39 |
+
"merges.txt",
|
40 |
+
"added_tokens.json",
|
41 |
+
"tokenizer_config.json"
|
42 |
+
],
|
43 |
+
"tokenizer_info": {
|
44 |
+
"token_postproc_method": "byte_level",
|
45 |
+
"prepend_space_in_encode": false,
|
46 |
+
"strip_space_in_decode": false
|
47 |
+
},
|
48 |
+
"conv_template": {
|
49 |
+
"name": "qwen2",
|
50 |
+
"system_template": "<|im_start|>system\n{system_message}<|im_end|>\n",
|
51 |
+
"system_message": "You are Sunny Starscout, a female Earth pony and the main protagonist of the fifth generation of My Little Pony. You are the daughter of Argyle Starshine and an unnamed mare. You represent the element of hope. You have five friends: Pipp, Zipp, Hitch, Izzy, Misty. Somepony wants to talk to you, and you will reply to them in a friendly manner.",
|
52 |
+
"system_prefix_token_ids": null,
|
53 |
+
"add_role_after_system_message": true,
|
54 |
+
"roles": {
|
55 |
+
"user": "<|im_start|>somepony",
|
56 |
+
"assistant": "<|im_start|>Sunny Starscout"
|
57 |
+
},
|
58 |
+
"role_templates": {
|
59 |
+
"user": "{user_message}",
|
60 |
+
"assistant": "{assistant_message}",
|
61 |
+
"tool": "{tool_message}"
|
62 |
+
},
|
63 |
+
"messages": [],
|
64 |
+
"seps": [
|
65 |
+
"<|im_end|>\n"
|
66 |
+
],
|
67 |
+
"role_content_sep": "\n",
|
68 |
+
"role_empty_sep": "\n",
|
69 |
+
"stop_str": [
|
70 |
+
"<|endoftext|>",
|
71 |
+
"<|im_end|>"
|
72 |
+
],
|
73 |
+
"stop_token_ids": [
|
74 |
+
151643,
|
75 |
+
151645
|
76 |
+
],
|
77 |
+
"function_string": "",
|
78 |
+
"use_function_calling": false
|
79 |
+
},
|
80 |
+
"pad_token_id": 49152,
|
81 |
+
"bos_token_id": 1,
|
82 |
+
"eos_token_id": 2
|
83 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1c20f9c1c8f2319323d9e593ad08c7f60a35e9a333934c53c81db61a03a6476
|
3 |
+
size 3422782048
|
ndarray-cache.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
params_shard_0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d9b998a4e59ff009fe039d6619bea5883cc2d34c30ab94a0696a0feaa3b559d9
|
3 |
+
size 50332672
|
params_shard_1.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3fb21bc3dc0d76b0e7cea2d2a79d42a34d063461e40e928dd1c113efbb9dda09
|
3 |
+
size 32510080
|
params_shard_10.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b15bf13017fa439cf5268fabcf2f7e27a206b8c7a9bb2dda49cca0aab82b69b3
|
3 |
+
size 28315648
|
params_shard_11.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d6a7982c2c219383fc0bf6ecbdbd770fdca383e1a2a1008c0f0de6c4b3ea967
|
3 |
+
size 16777216
|
params_shard_12.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ad84156ac7d23eb89e3f4f92c5e07917a43a77db379fc52e68f53f69eb550c05
|
3 |
+
size 30420992
|
params_shard_13.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:66ebdb3aba00ccf7895203d76838bab876e30d997a006a998d65342442e8df9f
|
3 |
+
size 28315648
|
params_shard_14.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:769d9ef99b4b7de3ec0bc756fa6f73646ec2c66d4489a9612fdec6f485522b7e
|
3 |
+
size 16777216
|
params_shard_15.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:449a69b0d24f253d31738259d44b69e555374dfa2287a8fb70f95277569b427c
|
3 |
+
size 30420992
|
params_shard_16.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:990cb470e88f7ec24a01d402b4c0197314764e49870f058c524b3a79c6cc4613
|
3 |
+
size 28315648
|
params_shard_17.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4af677dcb7703bb5870e7ab9a97d73ea0774252f2aa16a195716546c498bfc22
|
3 |
+
size 16777216
|
params_shard_18.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7675e206217aa45be63b20965d864373e47edbd163bb352f4996bea6c4780af0
|
3 |
+
size 30420992
|
params_shard_19.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a83161743e65fd855a8373a8bf3fde5c766dfe3dd5eb215264ded2e5c4dbebc1
|
3 |
+
size 28315648
|
params_shard_2.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0d7968efd6e12d1e57ecaaed5e5bc84db798a9ce46ba24ad3cfef379b3e544e0
|
3 |
+
size 16777216
|
params_shard_20.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ac678bf37d6b88dde3ca5eb1ae3bf1a29cfd38eebdc9ed4fffcad9a826eed2a1
|
3 |
+
size 16777216
|
params_shard_21.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:23429259841736777e94346091a76d6f4bd5e6e8b0605df0df486027b75feb00
|
3 |
+
size 30420992
|
params_shard_22.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5811639503bbc69ebc2f7c3bab2c2f34f1244615807d16c2aacba09cb10b04f0
|
3 |
+
size 28315648
|
params_shard_23.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:42fc0b39010171367097ed41b74ab712f8091ced740a7b139dc6e6f6f5d9c3c9
|
3 |
+
size 16777216
|
params_shard_24.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:944e37b963fe3eb1b2a358de0075b93c0590e0bd9c4e62a701b920996ec51d43
|
3 |
+
size 30420992
|
params_shard_25.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a710136aa6bd7cfac8b3554cff31e9c50527856dd4d2564b01bff519da9e00cf
|
3 |
+
size 28315648
|
params_shard_26.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a63912373c4e63d3b8486196a36b98310b87e28806c739f54381f8cbd86399b9
|
3 |
+
size 16777216
|
params_shard_27.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:461be7f1db692ed55da316589596c4fb175647b5e2f89690fb7529e9f2fb494b
|
3 |
+
size 30420992
|
params_shard_28.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eaffdd6b337eb8f2f1b951067cff01dccbdcb36109ada2e168b5f6b5d1af2fd9
|
3 |
+
size 28315648
|
params_shard_29.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a02fa91bca191477c124bbbfaa27e8b12dba985851dae09d814f497b2890f5a2
|
3 |
+
size 16777216
|
params_shard_3.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a5738c15f136714ef1c06f04a950780a2c3587f4be1573586e59ac5733b4f9ed
|
3 |
+
size 32522240
|
params_shard_30.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:52601d10071c96b8b801cd97e8e29d26997fdd119f0e643bdb9de5dde103ce54
|
3 |
+
size 30420992
|
params_shard_31.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f9e37ebb3bd64fba30b4b63911b385492efe3e8d0e1eea486e47f6d2c5668118
|
3 |
+
size 28315648
|
params_shard_32.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b4bc76f721200de787ad9a148951dcd1f877e6fdf7ba1b70e0e61d6f22a12e0d
|
3 |
+
size 16777216
|
params_shard_33.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f7a2e6e68c61ed31d6a3858b4ad8d480cec00dccf603c93dcdf7eee6e6a89b2f
|
3 |
+
size 30420992
|
params_shard_34.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ef01f5e673346915b28cc30e75a7daf6c9049e47c3ed545490c5431090187c36
|
3 |
+
size 28315648
|
params_shard_35.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:30ed6161d3d377ea3add7fe04386956fb6ccfe51a5ee7f03dd15509640ac27e3
|
3 |
+
size 16777216
|
params_shard_36.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ccfad423476eb6b5cc35ea5ed9429a8af6bc8a723898acb66ad308627d0aca34
|
3 |
+
size 30420992
|
params_shard_4.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:797e875ec8775473985d3c28541452e5eccf02cb36895d070812d63d4da738bf
|
3 |
+
size 28315648
|
params_shard_5.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:14fb8ad439268d816e27a8d8a80ef289959e90b11c4ba22b3018fb0cd7c70221
|
3 |
+
size 16777216
|
params_shard_6.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:feadddbfcf3e0aa4aec94eb0a160ebadcfc65fed56ae9915397ceeacdd2b020f
|
3 |
+
size 30420992
|
params_shard_7.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a64226d81d149406639f01261ca77e9b99af9468758215937e3ef69baab86089
|
3 |
+
size 28315648
|
params_shard_8.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f744a5762013cef2dc7eaa174f6fa7cc78dc66b9c758d330cbfcba79fbc27af5
|
3 |
+
size 16777216
|
params_shard_9.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53a495318c1eb6b55329374c55cfc36faede3526510ffca449bfb931eb2ff992
|
3 |
+
size 30420992
|
tokenizer_config.json
CHANGED
@@ -169,4 +169,4 @@
|
|
169 |
"tokenizer_class": "GPT2Tokenizer",
|
170 |
"unk_token": "�",
|
171 |
"vocab_size": 49152
|
172 |
-
}
|
|
|
169 |
"tokenizer_class": "GPT2Tokenizer",
|
170 |
"unk_token": "�",
|
171 |
"vocab_size": 49152
|
172 |
+
}
|