--- license: other language: - en pipeline_tag: text-generation inference: false tags: - transformers - gguf - imatrix - openchat-3.6-8b-20240522 --- Quantizations of https://huggingface.co/openchat/openchat-3.6-8b-20240522 # From original readme ### Conversation templates 💡 **Default Mode**: Best for coding, chat and general tasks ``` GPT4 Correct User: Hello<|end_of_turn|>GPT4 Correct Assistant: Hi<|end_of_turn|>GPT4 Correct User: How are you today?<|end_of_turn|>GPT4 Correct Assistant: ``` ⚠️ **Notice:** Remember to set `<|end_of_turn|>` as end of generation token. The default template is also available as the integrated `tokenizer.chat_template`, which can be used instead of manually specifying the template: ```python messages = [ {"role": "user", "content": "Hello"}, {"role": "assistant", "content": "Hi"}, {"role": "user", "content": "How are you today?"} ] tokens = tokenizer.apply_chat_template(messages, add_generation_prompt=True) ``` ### Inference using Transformers ```python from transformers import AutoTokenizer, AutoModelForCausalLM import torch model_id = "openchat/openchat-3.6-8b-20240522" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16, device_map="auto") messages = [ {"role": "user", "content": "Explain how large language models work in detail."}, ] input_ids = tokenizer.apply_chat_template(messages, add_generation_prompt=True, return_tensors="pt").to(model.device) outputs = model.generate(input_ids, do_sample=True, temperature=0.5, max_new_tokens=1024 ) response = outputs[0][input_ids.shape[-1]:] print(tokenizer.decode(response, skip_special_tokens=True)) ```