bettercallcaleb
commited on
Commit
•
9e17add
1
Parent(s):
b5f129a
initial commit
Browse files- LICENSE +0 -0
- README.md +28 -1
- config.json +42 -0
- deep-coder-4bit-128g.safetensors +3 -0
- generation_config.json +6 -0
- quantize_config.json +10 -0
- tokenizer.json +0 -0
- tokenizer_config.json +35 -0
LICENSE
ADDED
File without changes
|
README.md
CHANGED
@@ -1,3 +1,30 @@
|
|
1 |
---
|
2 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
base_model: deepseek-ai/deepseek-coder-6.7b-instruct
|
3 |
+
inference: false
|
4 |
+
license: Apache 2.0
|
5 |
+
license_link: LICENSE
|
6 |
+
license_name: deepseek
|
7 |
+
model_creator: DeepSeek
|
8 |
+
model_name: Deepseek Coder 6.7B
|
9 |
+
model_type: deepseek
|
10 |
+
prompt_template: 'You are an AI programming assistant, utilizing the Deepseek Coder
|
11 |
+
model, developed by Deepseek Company, and you only answer questions related to computer
|
12 |
+
science. For politically sensitive questions, security and privacy issues, and other
|
13 |
+
non-computer science questions, you will refuse to answer.
|
14 |
+
|
15 |
+
### Instruction:
|
16 |
+
|
17 |
+
{prompt}
|
18 |
+
|
19 |
+
### Response:
|
20 |
+
|
21 |
+
'
|
22 |
+
quantized_by: ramgpt
|
23 |
---
|
24 |
+
|
25 |
+
# Deepseek Coder 6.7B - GPTQ
|
26 |
+
- Model creator: [DeepSeek](https://huggingface.co/deepseek-ai)
|
27 |
+
- Original model: [Deepseek Coder 6.7B](https://huggingface.co/deepseek-ai/deepseek-coder-6.7b-instruct)
|
28 |
+
|
29 |
+
|
30 |
+
|
config.json
ADDED
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "/workspace/process/deepseek-ai_deepseek-coder-6.7b-instruct/source",
|
3 |
+
"architectures": [
|
4 |
+
"LlamaForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_bias": false,
|
7 |
+
"bos_token_id": 32013,
|
8 |
+
"eos_token_id": 32021,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 4096,
|
11 |
+
"initializer_range": 0.02,
|
12 |
+
"intermediate_size": 11008,
|
13 |
+
"max_position_embeddings": 16384,
|
14 |
+
"model_type": "llama",
|
15 |
+
"num_attention_heads": 32,
|
16 |
+
"num_hidden_layers": 32,
|
17 |
+
"num_key_value_heads": 32,
|
18 |
+
"pad_token_id": 0,
|
19 |
+
"pretraining_tp": 1,
|
20 |
+
"rms_norm_eps": 1e-06,
|
21 |
+
"rope_scaling": {
|
22 |
+
"factor": 4.0,
|
23 |
+
"type": "linear"
|
24 |
+
},
|
25 |
+
"rope_theta": 100000,
|
26 |
+
"tie_word_embeddings": false,
|
27 |
+
"torch_dtype": "bfloat16",
|
28 |
+
"transformers_version": "4.35.0",
|
29 |
+
"use_cache": true,
|
30 |
+
"vocab_size": 32256,
|
31 |
+
"quantization_config": {
|
32 |
+
"bits": 4,
|
33 |
+
"group_size": 128,
|
34 |
+
"damp_percent": 0.1,
|
35 |
+
"desc_act": true,
|
36 |
+
"sym": true,
|
37 |
+
"true_sequential": true,
|
38 |
+
"model_name_or_path": null,
|
39 |
+
"model_file_base_name": "model",
|
40 |
+
"quant_method": "gptq"
|
41 |
+
}
|
42 |
+
}
|
deep-coder-4bit-128g.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ce7918b8d9a4d0f1eb97087713ef287f61e980c5eed67a7e111edbe3a5357833
|
3 |
+
size 3898164856
|
generation_config.json
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_from_model_config": true,
|
3 |
+
"bos_token_id": 32013,
|
4 |
+
"eos_token_id": 32014,
|
5 |
+
"transformers_version": "4.34.1"
|
6 |
+
}
|
quantize_config.json
ADDED
@@ -0,0 +1,10 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bits": 4,
|
3 |
+
"group_size": 128,
|
4 |
+
"damp_percent": 0.1,
|
5 |
+
"desc_act": true,
|
6 |
+
"sym": true,
|
7 |
+
"true_sequential": true,
|
8 |
+
"model_name_or_path": null,
|
9 |
+
"model_file_base_name": "model"
|
10 |
+
}
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
ADDED
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"bos_token": {
|
5 |
+
"__type": "AddedToken",
|
6 |
+
"content": "<|begin▁of▁sentence|>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": true,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false
|
11 |
+
},
|
12 |
+
"clean_up_tokenization_spaces": false,
|
13 |
+
"eos_token": {
|
14 |
+
"__type": "AddedToken",
|
15 |
+
"content": "<|EOT|>",
|
16 |
+
"lstrip": false,
|
17 |
+
"normalized": true,
|
18 |
+
"rstrip": false,
|
19 |
+
"single_word": false
|
20 |
+
},
|
21 |
+
"legacy": true,
|
22 |
+
"model_max_length": 16384,
|
23 |
+
"pad_token": {
|
24 |
+
"__type": "AddedToken",
|
25 |
+
"content": "<|end▁of▁sentence|>",
|
26 |
+
"lstrip": false,
|
27 |
+
"normalized": true,
|
28 |
+
"rstrip": false,
|
29 |
+
"single_word": false
|
30 |
+
},
|
31 |
+
"sp_model_kwargs": {},
|
32 |
+
"unk_token": null,
|
33 |
+
"tokenizer_class": "LlamaTokenizerFast",
|
34 |
+
"chat_template": "{%- set found_item = false -%}\n{%- for message in messages -%}\n {%- if message['role'] == 'system' -%}\n {%- set found_item = true -%}\n {%- endif -%}\n{%- endfor -%}\n{%- if not found_item -%}\n{{'You are an AI programming assistant, utilizing the Deepseek Coder model, developed by Deepseek Company, and you only answer questions related to computer science. For politically sensitive questions, security and privacy issues, and other non-computer science questions, you will refuse to answer.\\n'}}\n{%- endif %}\n{%- for message in messages %}\n {%- if message['role'] == 'system' %}\n{{ message['content'] }}\n {%- else %}\n {%- if message['role'] == 'user' %}\n{{'### Instruction:\\n' + message['content'] + '\\n'}}\n {%- else %}\n{{'### Response:\\n' + message['content'] + '\\n<|EOT|>\\n'}}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{{'### Response:\\n'}}\n"
|
35 |
+
}
|