sharpenb commited on
Commit
1e450c7
1 Parent(s): 510b0bb

Upload folder using huggingface_hub (#1)

Browse files

- acc789cf2ec45a3b4238f63bf644a3cf7133af0deb1e9f9b5285855b381c9cac (92bee968e4d4a8baede89171016f2922bf38cb3b)
- cef3134eb6b0e19e7675a9ab7d3bca7f53ab4079437aece2848dba3a60cf2fa8 (7ae221777ae5e3bd06d7406c95ef512c419c89bc)
- a1823acc47ddb9447ec1689c14194703dd9785355bbbaab51a8cf02c18e54484 (d858ee9061b9700854994709785f208795855616)
- a19ae5aac19222fa6ff7ef7c3b31fc6ab5ea40e1b8cbfa896fd5e26f2f371600 (215e09c51ba38a463e6a457de2558a557c634ea0)
- 0a0c0f0c815824378dce02a9cfddf5d45a82383354ec4954970688c1e594f6c1 (73736a02f3a1c5abc84800e796bc69318d665c3e)
- 6e492f439d227448bce5c6e08f5d1303e5ef9ac0b4c5d8ba3534852f86bfd53b (88d28e67fbf67d3eea14319894a3dfcee31250de)
- 479d2523ed6a857b2398817881743a755a33c67a099e54b0e5567f11c5dfc129 (ac1534c8d48b7ca715e5cc1f62d9ed43ce3c4786)
- 1a8fcb57781e7284f08d68f4a96805cefedc8da5f7568419aad1d87f8d911b6a (981649ebe9e71cb849b0c8c3d9b52fb3eb0dec51)
- 2e0d3bdda95109e9b01efbdce2a5e1285a9e9362afad0489f54be135412ed9d9 (ad5cbb2a757511a0b258127204278e3e2c3f1a4d)
- dc8a2a106533b6e36536c9a52a5b6a72c206e75f77de402aecf27351c8dc8e2a (a66cc17a11fca339a966b4ca230a5847925cd695)
- 7e72910f0bdf7172d91f85559798ce54677e97b29735043b2002c28367207b1a (27b08be5582a6d94432c41870dda57ba6fdeaa52)
- 4f3ebcac0c740918e29355bcb9cf81299033636407b6bdfb4c7fae9bdc2ebb29 (833aeb125170466cab7184f337bf5f4b6a5b5e25)
- ce2fdbb2e652915d17cfbaeef9b9faa865b343be61952c0ee29dde94bc14c572 (915b35c254a3e33cda41b8dca857f229bd10aa26)
- a3956a11d4d8216163bb2e8ae2740c7293461e56293fb3f3f803beaa2a9d80a7 (c77e6f60c259913ead5366d0eed60abc48a059e7)
- 46b466877d14fb2c7f90ec28aea889c1dcffdd6f6fdf53c10a4af001af6404cd (e611733b1f6580807ce5694141a5e8a280efb5ad)

added_tokens.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "<|im_end|>": 100279,
3
+ "<|im_start|>": 100278,
4
+ "<|pad|>": 100277
5
+ }
config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "SinclairSchneider/dbrx-instruct-quantization-fixed",
3
+ "architectures": [
4
+ "DbrxForCausalLM"
5
+ ],
6
+ "attn_config": {
7
+ "clip_qkv": 8,
8
+ "kv_n_heads": 8,
9
+ "model_type": "",
10
+ "rope_theta": 500000
11
+ },
12
+ "auto_map": {
13
+ "AutoConfig": "SinclairSchneider/dbrx-instruct-quantization-fixed--configuration_dbrx.DbrxConfig",
14
+ "AutoModelForCausalLM": "SinclairSchneider/dbrx-instruct-quantization-fixed--modeling_dbrx.DbrxForCausalLM"
15
+ },
16
+ "d_model": 6144,
17
+ "emb_pdrop": 0.0,
18
+ "ffn_config": {
19
+ "ffn_hidden_size": 10752,
20
+ "model_type": "",
21
+ "moe_jitter_eps": 0,
22
+ "moe_loss_weight": 0.05,
23
+ "moe_num_experts": 16,
24
+ "moe_top_k": 4
25
+ },
26
+ "initializer_range": 0.02,
27
+ "max_seq_len": 32768,
28
+ "model_type": "dbrx",
29
+ "n_heads": 48,
30
+ "n_layers": 40,
31
+ "output_router_logits": false,
32
+ "quantization_config": {
33
+ "_load_in_4bit": true,
34
+ "_load_in_8bit": false,
35
+ "bnb_4bit_compute_dtype": "bfloat16",
36
+ "bnb_4bit_quant_storage": "uint8",
37
+ "bnb_4bit_quant_type": "fp4",
38
+ "bnb_4bit_use_double_quant": true,
39
+ "llm_int8_enable_fp32_cpu_offload": false,
40
+ "llm_int8_has_fp16_weight": false,
41
+ "llm_int8_skip_modules": [
42
+ "lm_head"
43
+ ],
44
+ "llm_int8_threshold": 6.0,
45
+ "load_in_4bit": true,
46
+ "load_in_8bit": false,
47
+ "quant_method": "bitsandbytes"
48
+ },
49
+ "resid_pdrop": 0.0,
50
+ "router_aux_loss_coef": 0.05,
51
+ "tie_word_embeddings": false,
52
+ "torch_dtype": "bfloat16",
53
+ "transformers_version": "4.39.2",
54
+ "use_cache": true,
55
+ "vocab_size": 100352
56
+ }
generation_config.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "eos_token_id": [
4
+ 100257,
5
+ 100279
6
+ ],
7
+ "transformers_version": "4.39.2"
8
+ }
model-00001-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39a7d2e79067a4038e3d3e2e868fb4c637a79e1e12feb2780cc93c9b07b48900
3
+ size 4982218709
model-00002-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6caefabcbef424f5c164bba37f24f4c9a45ccd5ba27772956026c69fce1f1028
3
+ size 4975991632
model-00003-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b394b566d55a32eb8796ba3c4ee9fed687ca8fab8eb0c7e3b0759fdb285f190b
3
+ size 4975991636
model-00004-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:726f672000ccd78e4552920e703363efcb4e857086fb357d41df25e52e0b808f
3
+ size 4975991957
model-00005-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba3347d50ac5f64e8ee58f8145d0e0ec534e320569d8c44320ff07b4ef3452eb
3
+ size 4975992536
model-00006-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:749045e516bd5673583de3a8da1aa2a5580c4fedf369e84796b97104f62390b7
3
+ size 4975992539
model-00007-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b7e2a409b2e4586b8f07dd381db414634196f7ee167bea1b0f2967659a980fa
3
+ size 4998646326
model-00008-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2531b0fa4f0a13f2fee54ceac1a68321cb3afd8cf53f5c1cabad4dd48655ec7
3
+ size 4975980133
model-00009-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b4dfcd4fbf2358f2020f28536d38040e6cf55bf6d93f9871112e9f9609ca39e
3
+ size 4975992531
model-00010-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69e875475cae60d267509f02547c87cb27665c12d39e3677664cd298ff54068f
3
+ size 4975992531
model-00011-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba2f01c3c5e6732fd200e7f9d04855eeec00e34708cd923eecd0f93dd379100c
3
+ size 4975992531
model-00012-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c260cb7890346ed6364486b826863e3e99d45dcee9b32a0751f51e24f2c76b9a
3
+ size 4975992528
model-00013-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a09f7c97c3e9c9f2aaa8cf41d8c7d8a2c0cb13d26539e190b8fc9564970f9243
3
+ size 4975992516
model-00014-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbb846a1be7dd94c54cc7ee72cfd1ccd88d5a8ae95f3b849ded8beda2ec0ec6d
3
+ size 3771747555
model-00015-of-00015.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c09710cf02b44be76007e8ddc836d335f3e5f008bf247431d55c4afbc0a3b17
3
+ size 1233125504
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>"
5
+ ],
6
+ "bos_token": {
7
+ "content": "<|endoftext|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false
12
+ },
13
+ "eos_token": {
14
+ "content": "<|endoftext|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false
19
+ },
20
+ "pad_token": {
21
+ "content": "<|pad|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false
26
+ },
27
+ "unk_token": {
28
+ "content": "<|endoftext|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false
33
+ }
34
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": false,
5
+ "added_tokens_decoder": {
6
+ "100257": {
7
+ "content": "<|endoftext|>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "100277": {
15
+ "content": "<|pad|>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "100278": {
23
+ "content": "<|im_start|>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "100279": {
31
+ "content": "<|im_end|>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ }
38
+ },
39
+ "additional_special_tokens": [
40
+ "<|im_start|>",
41
+ "<|im_end|>"
42
+ ],
43
+ "auto_map": {
44
+ "AutoTokenizer": [
45
+ "SinclairSchneider/dbrx-instruct-quantization-fixed--tiktoken.TiktokenTokenizerWrapper",
46
+ null
47
+ ]
48
+ },
49
+ "bos_token": "<|endoftext|>",
50
+ "clean_up_tokenization_spaces": true,
51
+ "encoding_name": null,
52
+ "eos_token": "<|endoftext|>",
53
+ "errors": "replace",
54
+ "model_max_length": 1000000000000000019884624838656,
55
+ "model_name": "gpt-4",
56
+ "pad_token": "<|pad|>",
57
+ "tokenizer_class": "TiktokenTokenizerWrapper",
58
+ "unk_token": "<|endoftext|>",
59
+ "use_default_system_prompt": true
60
+ }