StopTryharding commited on
Commit
3feef72
1 Parent(s): 5eadf18

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -20
README.md CHANGED
@@ -52,24 +52,10 @@ experts:
52
 
53
  ## 💻 Usage
54
 
55
- ```python
56
- !pip install -qU transformers bitsandbytes accelerate
57
-
58
- from transformers import AutoTokenizer
59
- import transformers
60
- import torch
61
-
62
- model = "shadowml/Beyonder-2x7B-v2"
63
-
64
- tokenizer = AutoTokenizer.from_pretrained(model)
65
- pipeline = transformers.pipeline(
66
- "text-generation",
67
- model=model,
68
- model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True},
69
- )
70
-
71
- messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}]
72
- prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
73
- outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
74
- print(outputs[0]["generated_text"])
75
  ```
 
52
 
53
  ## 💻 Usage
54
 
55
+ ```
56
+ Load in Kobold.cpp or whatever. I found Alpaca (and Alpaca-ish) prompts worked well. Settings that worked good for me are:
57
+ Min P - 0.1
58
+ Dynamic Temperature Min 0 Max 3
59
+ Rep Pen 1.03
60
+ Rep Pen Range 1000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
61
  ```