Jue Wang commited on
Commit
9995d97
1 Parent(s): af6ae45
config.json CHANGED
@@ -25,7 +25,7 @@
25
  "rope_theta": 100000,
26
  "tie_word_embeddings": false,
27
  "torch_dtype": "bfloat16",
28
- "transformers_version": "4.38.1",
29
  "use_cache": true,
30
  "vocab_size": 32256
31
  }
 
25
  "rope_theta": 100000,
26
  "tie_word_embeddings": false,
27
  "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.41.0.dev0",
29
  "use_cache": true,
30
  "vocab_size": 32256
31
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:995a68cf76915b5ad450690e43655ac6fac9e2b1f0c6b96ffc8616f6f3819b25
3
  size 4971750904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c01f2b54cbc01ee79565279165446a0a470ed66d0d3f892c368aaeffd9018898
3
  size 4971750904
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2bddb316f61128d1bba3c8fdc23b22503fe98e0d2612dc5ce4e55e9350d3313
3
- size 2235727984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fab83d3cad51065ab6d1f4874a5e825ee097a777595084c4f879ccab6c2c7ae
3
+ size 2235727976
model.safetensors.index.json CHANGED
@@ -742,7 +742,7 @@
742
  "model.layers.9.self_attn.v_proj.qzeros": "model-00001-of-00002.safetensors",
743
  "model.layers.9.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
744
  "model.norm.weight": "model-00002-of-00002.safetensors",
745
- "trtllm_bits": "model-00002-of-00002.safetensors",
746
- "trtllm_groupsize": "model-00002-of-00002.safetensors"
747
  }
748
  }
 
742
  "model.layers.9.self_attn.v_proj.qzeros": "model-00001-of-00002.safetensors",
743
  "model.layers.9.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
744
  "model.norm.weight": "model-00002-of-00002.safetensors",
745
+ "trt_bits": "model-00002-of-00002.safetensors",
746
+ "trt_groupsize": "model-00002-of-00002.safetensors"
747
  }
748
  }
tokenizer.json CHANGED
@@ -321,6 +321,7 @@
321
  "end_of_word_suffix": null,
322
  "fuse_unk": false,
323
  "byte_fallback": false,
 
324
  "vocab": {
325
  "!": 0,
326
  "\"": 1,
 
321
  "end_of_word_suffix": null,
322
  "fuse_unk": false,
323
  "byte_fallback": false,
324
+ "ignore_merges": false,
325
  "vocab": {
326
  "!": 0,
327
  "\"": 1,
tokenizer_config.json CHANGED
@@ -183,7 +183,6 @@
183
  "chat_template": "{% if not add_generation_prompt is defined %}\n{% set add_generation_prompt = false %}\n{% endif %}\n{%- set ns = namespace(found=false) -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'system' -%}\n {%- set ns.found = true -%}\n {%- endif -%}\n{%- endfor -%}\n{{bos_token}}{%- if not ns.found -%}\n{{'You are an AI programming assistant, utilizing the Deepseek Coder model, developed by Deepseek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer\\n'}}\n{%- endif %}\n{%- for message in messages %}\n {%- if message['role'] == 'system' %}\n{{ message['content'] }}\n {%- else %}\n {%- if message['role'] == 'user' %}\n{{'### Instruction:\\n' + message['content'] + '\\n'}}\n {%- else %}\n{{'### Response:\\n' + message['content'] + '\\n<|EOT|>\\n'}}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{% if add_generation_prompt %}\n{{'### Response:'}}\n{% endif %}",
184
  "clean_up_tokenization_spaces": false,
185
  "eos_token": "<|EOT|>",
186
- "legacy": true,
187
  "model_max_length": 16384,
188
  "pad_token": "<|end▁of▁sentence|>",
189
  "sp_model_kwargs": {},
 
183
  "chat_template": "{% if not add_generation_prompt is defined %}\n{% set add_generation_prompt = false %}\n{% endif %}\n{%- set ns = namespace(found=false) -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'system' -%}\n {%- set ns.found = true -%}\n {%- endif -%}\n{%- endfor -%}\n{{bos_token}}{%- if not ns.found -%}\n{{'You are an AI programming assistant, utilizing the Deepseek Coder model, developed by Deepseek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer\\n'}}\n{%- endif %}\n{%- for message in messages %}\n {%- if message['role'] == 'system' %}\n{{ message['content'] }}\n {%- else %}\n {%- if message['role'] == 'user' %}\n{{'### Instruction:\\n' + message['content'] + '\\n'}}\n {%- else %}\n{{'### Response:\\n' + message['content'] + '\\n<|EOT|>\\n'}}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{% if add_generation_prompt %}\n{{'### Response:'}}\n{% endif %}",
184
  "clean_up_tokenization_spaces": false,
185
  "eos_token": "<|EOT|>",
 
186
  "model_max_length": 16384,
187
  "pad_token": "<|end▁of▁sentence|>",
188
  "sp_model_kwargs": {},