mlabonne commited on
Commit
04a9c72
1 Parent(s): 44f439e

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -37,28 +37,13 @@
37
  "num_key_value_heads": 8,
38
  "output_router_logits": false,
39
  "pad_token_id": 0,
40
- "quantization_config": {
41
- "_load_in_4bit": true,
42
- "_load_in_8bit": false,
43
- "bnb_4bit_compute_dtype": "bfloat16",
44
- "bnb_4bit_quant_storage": "bfloat16",
45
- "bnb_4bit_quant_type": "nf4",
46
- "bnb_4bit_use_double_quant": true,
47
- "llm_int8_enable_fp32_cpu_offload": false,
48
- "llm_int8_has_fp16_weight": false,
49
- "llm_int8_skip_modules": null,
50
- "llm_int8_threshold": 6.0,
51
- "load_in_4bit": true,
52
- "load_in_8bit": false,
53
- "quant_method": "bitsandbytes"
54
- },
55
  "rms_norm_eps": 1e-06,
56
  "router_aux_loss_coef": 0.001,
57
  "sliding_window": null,
58
  "tie_word_embeddings": false,
59
- "torch_dtype": "bfloat16",
60
  "transformers_version": "4.40.0.dev0",
61
- "use_cache": false,
62
  "use_mamba_kernels": true,
63
  "vocab_size": 65536
64
  }
 
37
  "num_key_value_heads": 8,
38
  "output_router_logits": false,
39
  "pad_token_id": 0,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  "rms_norm_eps": 1e-06,
41
  "router_aux_loss_coef": 0.001,
42
  "sliding_window": null,
43
  "tie_word_embeddings": false,
44
+ "torch_dtype": "float16",
45
  "transformers_version": "4.40.0.dev0",
46
+ "use_cache": true,
47
  "use_mamba_kernels": true,
48
  "vocab_size": 65536
49
  }
model-00001-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:622418d163a118a30330eea9eb61909e2fa3c1dbd74295f6a2b21b06a2b401ab
3
+ size 4951761080
model-00002-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77b5da2a63ff9dcb3f10da92508c83f381563b14438978c4938fc69152acc121
3
+ size 4884669232
model-00003-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58a28fc90973738dd058a8a5da8a2f25280bc2f3f4f07443ba30562e350194d5
3
+ size 4992556712
model-00004-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb3bca81a159a3766fea38ec1f8e1318dd6a06c40b8c2259bf88d0562fa3edbc
3
+ size 4958853128
model-00005-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f401c4605312e71678fbc95e7d751e090fe694def948dad264588f3aa6c7d9d7
3
+ size 4975763392
model-00006-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c5d7d9773bacdfd92a57eaceb9919729b87f4909ca6df57102d295a34c3947b
3
+ size 4884669224
model-00007-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c154e3b1347371e0b104bfaaf7571daace54f280d78ba295c62e2862b685e34
3
+ size 4884669208
model-00008-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59bf9c1b4d76d639b72fb21fdad5532503033fef982ac299d070283098478cdc
3
+ size 4992556768
model-00009-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f60a5f4132d099961e5bcd4185c6a3b6dbea9d113aa32591e12f7114a9dc46db
3
+ size 4932506760
model-00010-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8945b43073b4dab1c7c6f5cbfc347bb4da7f359ae378aea8a384b747434aec4d
3
+ size 4884669272
model-00011-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15b2ed16c60900a41cca661d6b895a48c1ede763c1b38748ad0fd01bbf5f68aa
3
+ size 4884669304
model-00012-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34a4db6e322aa31af2c4468ce09e5795cf0cb9dd7cc272f574830e92fde612f1
3
+ size 4884669304
model-00013-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3073e26bf19ac8d81cbee6869390c3bc61ad24c89c34fd1ee02ef882e4f5b0b
3
+ size 4932506760
model-00014-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4771e2bd565eb8cebe8bdf971926988001da5c8b3489f9fc2fb07e59c2d0ab1
3
+ size 4992556728
model-00015-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26ccb35b8ecd7978eaf84193ae9535b5b1308cb0d3fc9a48ed096ade53c18569
3
+ size 4884669296
model-00016-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:284dacd546497b0bbc09db3bc810ad117f449436567d97523f68c3ea4cdf0a1d
3
+ size 4884669304
model-00017-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58f180229dbd624b5eeb6a6784c38240efb87ee2eb85215cb710df014c517ff5
3
+ size 4908522432
model-00018-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c9a724f67bb22dc9c3d6f11aac52a906ed8cb6276638b1f249fa5bb235cb07d
3
+ size 4908653584
model-00019-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ac9f4476b113c88fb7267f41ee21a88f8f176716fb89b559ff1d4687968e165
3
+ size 4992556768
model-00020-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:613d6390cb3cc0beb1e1e8646673ad59b31d8e0481cd24f59946326875bdf472
3
+ size 4884669304
model-00021-of-00021.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32ac8666509dfcc1b7081b4e2668055d9f504c18719a158a01af1ce0b5f80c77
3
+ size 4647318216
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -36,7 +36,6 @@
36
  }
37
  },
38
  "bos_token": "<|startoftext|>",
39
- "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
40
  "clean_up_tokenization_spaces": false,
41
  "eos_token": "<|endoftext|>",
42
  "model_max_length": 1000000000000000019884624838656,
 
36
  }
37
  },
38
  "bos_token": "<|startoftext|>",
 
39
  "clean_up_tokenization_spaces": false,
40
  "eos_token": "<|endoftext|>",
41
  "model_max_length": 1000000000000000019884624838656,