File size: 250 Bytes
728e5f6
 
 
 
 
 
 
 
a99ec58
1
2
3
4
5
6
7
8
9
from ctransformers import AutoModelForCausalLM

llm_model = AutoModelForCausalLM.from_pretrained(
    "TheBloke/zephyr-7B-alpha-GGUF",
    model_file="zephyr-7b-alpha.Q4_K_M.gguf",
    model_type="llama",
    max_new_tokens=256,
    temperature=0.7
)