CharlieFRuan commited on
Commit
b9c462d
1 Parent(s): f606262

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. mlc-chat-config.json +7 -4
mlc-chat-config.json CHANGED
@@ -12,6 +12,7 @@
12
  "rms_norm_eps": 1e-06,
13
  "rope_theta": 1000000.0,
14
  "vocab_size": 152064,
 
15
  "context_window_size": 32768,
16
  "prefill_chunk_size": 2048,
17
  "tensor_parallel_shards": 1,
@@ -25,6 +26,7 @@
25
  "prefill_chunk_size": 2048,
26
  "attention_sink_size": -1,
27
  "tensor_parallel_shards": 1,
 
28
  "temperature": 0.7,
29
  "presence_penalty": 0.0,
30
  "frequency_penalty": 0.0,
@@ -42,9 +44,9 @@
42
  "strip_space_in_decode": false
43
  },
44
  "conv_template": {
45
- "name": "chatml",
46
  "system_template": "<|im_start|>system\n{system_message}<|im_end|>\n",
47
- "system_message": "A conversation between a user and an LLM-based AI assistant. The assistant gives helpful and honest answers.",
48
  "system_prefix_token_ids": null,
49
  "add_role_after_system_message": true,
50
  "roles": {
@@ -63,10 +65,11 @@
63
  "role_content_sep": "\n",
64
  "role_empty_sep": "\n",
65
  "stop_str": [
66
- "<|im_end|>"
67
  ],
68
  "stop_token_ids": [
69
- 2
 
70
  ],
71
  "function_string": "",
72
  "use_function_calling": false
 
12
  "rms_norm_eps": 1e-06,
13
  "rope_theta": 1000000.0,
14
  "vocab_size": 152064,
15
+ "tie_word_embeddings": false,
16
  "context_window_size": 32768,
17
  "prefill_chunk_size": 2048,
18
  "tensor_parallel_shards": 1,
 
26
  "prefill_chunk_size": 2048,
27
  "attention_sink_size": -1,
28
  "tensor_parallel_shards": 1,
29
+ "pipeline_parallel_stages": 1,
30
  "temperature": 0.7,
31
  "presence_penalty": 0.0,
32
  "frequency_penalty": 0.0,
 
44
  "strip_space_in_decode": false
45
  },
46
  "conv_template": {
47
+ "name": "qwen2",
48
  "system_template": "<|im_start|>system\n{system_message}<|im_end|>\n",
49
+ "system_message": "You are a helpful assistant.",
50
  "system_prefix_token_ids": null,
51
  "add_role_after_system_message": true,
52
  "roles": {
 
65
  "role_content_sep": "\n",
66
  "role_empty_sep": "\n",
67
  "stop_str": [
68
+ "<|endoftext|>, <|im_end|>"
69
  ],
70
  "stop_token_ids": [
71
+ 151643,
72
+ 151645
73
  ],
74
  "function_string": "",
75
  "use_function_calling": false