mlabonne commited on
Commit
55056a1
1 Parent(s): 414e0ac

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -40,13 +40,15 @@
40
  "quantization_config": {
41
  "_load_in_4bit": true,
42
  "_load_in_8bit": false,
43
- "bnb_4bit_compute_dtype": "bfloat16",
44
- "bnb_4bit_quant_storage": "float32",
45
- "bnb_4bit_quant_type": "nf4",
46
- "bnb_4bit_use_double_quant": true,
47
  "llm_int8_enable_fp32_cpu_offload": false,
48
  "llm_int8_has_fp16_weight": false,
49
- "llm_int8_skip_modules": null,
 
 
50
  "llm_int8_threshold": 6.0,
51
  "load_in_4bit": true,
52
  "load_in_8bit": false,
@@ -56,9 +58,9 @@
56
  "router_aux_loss_coef": 0.001,
57
  "sliding_window": null,
58
  "tie_word_embeddings": false,
59
- "torch_dtype": "bfloat16",
60
  "transformers_version": "4.40.0.dev0",
61
- "use_cache": false,
62
  "use_mamba_kernels": true,
63
  "vocab_size": 65536
64
  }
 
40
  "quantization_config": {
41
  "_load_in_4bit": true,
42
  "_load_in_8bit": false,
43
+ "bnb_4bit_compute_dtype": "float32",
44
+ "bnb_4bit_quant_storage": "uint8",
45
+ "bnb_4bit_quant_type": "fp4",
46
+ "bnb_4bit_use_double_quant": false,
47
  "llm_int8_enable_fp32_cpu_offload": false,
48
  "llm_int8_has_fp16_weight": false,
49
+ "llm_int8_skip_modules": [
50
+ "mamba"
51
+ ],
52
  "llm_int8_threshold": 6.0,
53
  "load_in_4bit": true,
54
  "load_in_8bit": false,
 
58
  "router_aux_loss_coef": 0.001,
59
  "sliding_window": null,
60
  "tie_word_embeddings": false,
61
+ "torch_dtype": "float16",
62
  "transformers_version": "4.40.0.dev0",
63
+ "use_cache": true,
64
  "use_mamba_kernels": true,
65
  "vocab_size": 65536
66
  }
model-00001-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a439b2047a16c590b63a7a8f759ea4bb7a8f5205eb07f8ed2f83d39b71a2f55
3
  size 4873003324
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e92c801b498d391703c2d2aa7ecfd5251c43cb8fcea5469a35cba13e6e91fb8
3
  size 4873003324
model-00002-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9769f0709625f7c4596eb1be8e2baacfd8f388092746d4406b725a8e577144e
3
  size 4984623243
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5931dc5bfcf2db3a287940e23861ffb9a9661e76e824cedb0a73d8cc359c2f0
3
  size 4984623243
model-00003-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2a9537e3542d8016fd4daeab9b3ef4058ceefdc35f24ed3fc3ace13948310c4
3
  size 4983413752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec106651959c1fc61482e954d6c1dee67d82d387f8c7edf366a1b89f9d14e027
3
  size 4983413752
model-00004-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f1c6103bbf6db840d25d6748ee80d7df09d0fbd0e94ebeb57bb9806eb8f4715f
3
  size 4997975503
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f5178778f3ee8501126fe41ba37b8d7fd3837fcc5b2e902779ddba7e801ed42
3
  size 4997975503
model-00005-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:afb102dd8aac845bba0f9d495ca5c8095a1f6be80c22f65cb08bc295a23680aa
3
  size 4983413840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea765123548c90cc441e70a0c8f6d766208994730bf03f41c6b4f18be1f5ed9a
3
  size 4983413840
model-00006-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c0e9210edb8e25246c3eb7ff9bd7f27edefc012c96e35027352d0a07eeef41b7
3
  size 4975844741
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f25d10f17ca3965392e5c876a59d901d89a18bed479e8fa790c375ac3b40a9a3
3
  size 4975844741
model-00007-of-00007.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aaaae336c946d1f5ab14a699c1543b424533f718895dc5305638e55329802df6
3
  size 3842901815
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24d0e9a3fb0a6f8c61489778097d6debaf5da9ccc77299d1fef5fbb19b02ae9c
3
  size 3842901815
tokenizer_config.json CHANGED
@@ -36,7 +36,6 @@
36
  }
37
  },
38
  "bos_token": "<|startoftext|>",
39
- "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
40
  "clean_up_tokenization_spaces": false,
41
  "eos_token": "<|endoftext|>",
42
  "model_max_length": 1000000000000000019884624838656,
 
36
  }
37
  },
38
  "bos_token": "<|startoftext|>",
 
39
  "clean_up_tokenization_spaces": false,
40
  "eos_token": "<|endoftext|>",
41
  "model_max_length": 1000000000000000019884624838656,