schroneko commited on
Commit
385e99a
·
verified ·
1 Parent(s): 7e41f68

10bf2ece3fc58a3cbba789fd2962d0f98d86f4d872dea20841b67e49443f932d

Browse files
README.md CHANGED
@@ -1,17 +1,17 @@
1
  ---
 
 
 
 
2
  language:
3
  - en
4
  - ja
5
- license: apache-2.0
6
  library_name: transformers
 
7
  tags:
8
  - mixtral
9
  - steerlm
10
  - mlx
11
- base_model: tokyotech-llm/Swallow-MX-8x7b-NVE-v0.1
12
- datasets:
13
- - OpenAssistant/oasst2
14
- - nvidia/HelpSteer
15
  model-index:
16
  - name: karakuri-ai/karakuri-lm-8x7b-chat-v0.1
17
  results:
@@ -33,8 +33,9 @@ model-index:
33
  ---
34
 
35
  # mlx-community/karakuri-lm-8x7b-chat-v0.1-8bit
36
- This model was converted to MLX format from [`karakuri-ai/karakuri-lm-8x7b-chat-v0.1`]() using mlx-lm version **0.12.1**.
37
- Refer to the [original model card](https://huggingface.co/karakuri-ai/karakuri-lm-8x7b-chat-v0.1) for more details on the model.
 
38
  ## Use with mlx
39
 
40
  ```bash
@@ -45,5 +46,14 @@ pip install mlx-lm
45
  from mlx_lm import load, generate
46
 
47
  model, tokenizer = load("mlx-community/karakuri-lm-8x7b-chat-v0.1-8bit")
48
- response = generate(model, tokenizer, prompt="hello", verbose=True)
 
 
 
 
 
 
 
 
 
49
  ```
 
1
  ---
2
+ base_model: karakuri-ai/karakuri-lm-8x7b-chat-v0.1
3
+ datasets:
4
+ - OpenAssistant/oasst2
5
+ - nvidia/HelpSteer
6
  language:
7
  - en
8
  - ja
 
9
  library_name: transformers
10
+ license: apache-2.0
11
  tags:
12
  - mixtral
13
  - steerlm
14
  - mlx
 
 
 
 
15
  model-index:
16
  - name: karakuri-ai/karakuri-lm-8x7b-chat-v0.1
17
  results:
 
33
  ---
34
 
35
  # mlx-community/karakuri-lm-8x7b-chat-v0.1-8bit
36
+
37
+ The Model [mlx-community/karakuri-lm-8x7b-chat-v0.1-8bit](https://huggingface.co/mlx-community/karakuri-lm-8x7b-chat-v0.1-8bit) was converted to MLX format from [karakuri-ai/karakuri-lm-8x7b-chat-v0.1](https://huggingface.co/karakuri-ai/karakuri-lm-8x7b-chat-v0.1) using mlx-lm version **0.19.0**.
38
+
39
  ## Use with mlx
40
 
41
  ```bash
 
46
  from mlx_lm import load, generate
47
 
48
  model, tokenizer = load("mlx-community/karakuri-lm-8x7b-chat-v0.1-8bit")
49
+
50
+ prompt="hello"
51
+
52
+ if hasattr(tokenizer, "apply_chat_template") and tokenizer.chat_template is not None:
53
+ messages = [{"role": "user", "content": prompt}]
54
+ prompt = tokenizer.apply_chat_template(
55
+ messages, tokenize=False, add_generation_prompt=True
56
+ )
57
+
58
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
59
  ```
config.json CHANGED
@@ -21,6 +21,10 @@
21
  "group_size": 64,
22
  "bits": 8
23
  },
 
 
 
 
24
  "rms_norm_eps": 1e-05,
25
  "rope_theta": 1000000.0,
26
  "router_aux_loss_coef": 0.02,
 
21
  "group_size": 64,
22
  "bits": 8
23
  },
24
+ "quantization_config": {
25
+ "group_size": 64,
26
+ "bits": 8
27
+ },
28
  "rms_norm_eps": 1e-05,
29
  "rope_theta": 1000000.0,
30
  "router_aux_loss_coef": 0.02,
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
 
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
 
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
4
+ "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",