Alignment-Lab-AI commited on
Commit
cfd2343
1 Parent(s): 1bb132d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -55,7 +55,7 @@ By combining high quality data, iterative fine-tuning with carefully selected "g
55
 
56
  ## Chat Template and Inference
57
 
58
- To use the Buzz-8b-Medium model for chat-based tasks, you can utilize the provided chat template. Here's an example of how to format the chat template and perform inference using the Hugging Face Transformers library:
59
  ```python
60
  from transformers import AutoTokenizer, AutoModelForCausalLM
61
 
@@ -89,6 +89,10 @@ response = tokenizer.decode(output[0], skip_special_tokens=True)
89
  print("Input:", prompt)
90
  print("Response:", response)
91
  ``````
 
 
 
 
92
  ## Conclusion
93
 
94
  We intend to focus on *updating* and improving the performance of these models, and surrounding open sourced infrastructure. Our next effort will focus on context and implementing the research currently being conducted by [Wing-Lian](https://github.com/winglian), the lead developer of the [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) training framework that underpins these experiments. We encourage the community to explore Wing-Lian's work, such as the [Llama-3-8b-64k-PoSE](https://huggingface.co/winglian/Llama-3-8b-64k-PoSE) and [llama-3-8b-256k-PoSE](https://huggingface.co/winglian/llama-3-8b-256k-PoSE) models, which showcase the potential for further advancements in language modeling.
 
55
 
56
  ## Chat Template and Inference
57
 
58
+ To use the Buzz-8b-Medium model for chat-based tasks, you can utilize the provided chat template. Here's an example of how to perform inference using the Hugging Face Transformers library:
59
  ```python
60
  from transformers import AutoTokenizer, AutoModelForCausalLM
61
 
 
89
  print("Input:", prompt)
90
  print("Response:", response)
91
  ``````
92
+ NOTE: this model is a COMPLETIONS model, it will generate text by default, which completes the text you send it, it only has a *start* <|begin_of_text|> and a *stop* token <|end_of_text|>
93
+ if you want it to have conversations reliably, append <|end_of_text|>\n<|begin_of_text|>assistant: to the end of your prompt, [the speaker 'assistant' is flexible, and can be tooled to the type of response you want, for example "Mathematician:"" will give you a different type of response, than "felon:"]
94
+
95
+ later iterations of the model will likely have formatting similar to *openchat*
96
  ## Conclusion
97
 
98
  We intend to focus on *updating* and improving the performance of these models, and surrounding open sourced infrastructure. Our next effort will focus on context and implementing the research currently being conducted by [Wing-Lian](https://github.com/winglian), the lead developer of the [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) training framework that underpins these experiments. We encourage the community to explore Wing-Lian's work, such as the [Llama-3-8b-64k-PoSE](https://huggingface.co/winglian/Llama-3-8b-64k-PoSE) and [llama-3-8b-256k-PoSE](https://huggingface.co/winglian/llama-3-8b-256k-PoSE) models, which showcase the potential for further advancements in language modeling.