thomadev0 commited on
Commit
f1d96ff
1 Parent(s): 695f3cb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -37,5 +37,6 @@ response = generate(model, tokenizer, prompt="hello", verbose=True)
37
  ## Use with mlx_lm cli
38
 
39
  ```bash
 
40
  python3 -m mlx_lm.generate --model mlx-community/Nous-Hermes-2-Mixtral-8x7B-DPO-4bit --prompt "<|im_start|>system\nYou are an accurate, educational, and helpful information assistant<|im_end|>\n<|im_start|>user\nWhat is the difference between awq vs gptq quantitization?<|im_end|>\n<|im_start|>assistant\n" --max-tokens 2048
41
  ```
 
37
  ## Use with mlx_lm cli
38
 
39
  ```bash
40
+ pip install -U mlx-lm
41
  python3 -m mlx_lm.generate --model mlx-community/Nous-Hermes-2-Mixtral-8x7B-DPO-4bit --prompt "<|im_start|>system\nYou are an accurate, educational, and helpful information assistant<|im_end|>\n<|im_start|>user\nWhat is the difference between awq vs gptq quantitization?<|im_end|>\n<|im_start|>assistant\n" --max-tokens 2048
42
  ```