finetuned8b-GGUF / README.md
Pavan178's picture
Update README.md
73e2cc9 verified

CSGO Coach Mia, Finetuned on mistralai/Mistral-7B-Instruct-v0.2

Sample usage :

from huggingface_hub import hf_hub_download from llama_cpp import Llama import torch

Specify the path to your .gguf file

model_path = '/content/finetuned8b/finetuned8b.Q5_K_M.gguf'

Instantiate the Llama model

llm = Llama(model_path=model_path)

prompt = "Coach Mia, help me with aiming "

Generation kwargs

generation_kwargs = { "max_tokens":200, "stop":'[INST]', "echo":False, # Echo the prompt in the output "top_k":1 # This is essentially greedy decoding, since the model will always return the highest-probability token. Set this value > 1 for sampling decoding }

res = llm(prompt, **generation_kwargs)

Unpack and the generated text from the LLM response dictionary and print it

print(res["choices"][0]["text"])

res is short for result

#output

100% accuracy. [/INST] Aiming is a crucial aspect of CS:GO. Let's start by analyzing your sensitivity settings and crosshair placement. We can also run some aim training drills to improve your precision.