finetuned8b-GGUF / README.md
Pavan178's picture
Update README.md
73e2cc9 verified
CSGO Coach Mia, Finetuned on mistralai/Mistral-7B-Instruct-v0.2
Sample usage :
from huggingface_hub import hf_hub_download
from llama_cpp import Llama
import torch
# Specify the path to your .gguf file
model_path = '/content/finetuned8b/finetuned8b.Q5_K_M.gguf'
# Instantiate the Llama model
llm = Llama(model_path=model_path)
prompt = "Coach Mia, help me with aiming "
## Generation kwargs
generation_kwargs = {
"max_tokens":200,
"stop":'[INST]',
"echo":False, # Echo the prompt in the output
"top_k":1 # This is essentially greedy decoding, since the model will always return the highest-probability token. Set this value > 1 for sampling decoding
}
res = llm(prompt, **generation_kwargs)
## Unpack and the generated text from the LLM response dictionary and print it
print(res["choices"][0]["text"])
# res is short for result
#output
100% accuracy. [/INST] Aiming is a crucial aspect of CS:GO. Let's start by analyzing your sensitivity settings and crosshair placement. We can also run some aim training drills to improve your precision.