CSGO Coach Mia, Finetuned on mistralai/Mistral-7B-Instruct-v0.2 Sample usage : from huggingface_hub import hf_hub_download from llama_cpp import Llama import torch # Specify the path to your .gguf file model_path = '/content/finetuned8b/finetuned8b.Q5_K_M.gguf' # Instantiate the Llama model llm = Llama(model_path=model_path) prompt = "Coach Mia, help me with aiming " ## Generation kwargs generation_kwargs = { "max_tokens":200, "stop":'[INST]', "echo":False, # Echo the prompt in the output "top_k":1 # This is essentially greedy decoding, since the model will always return the highest-probability token. Set this value > 1 for sampling decoding } res = llm(prompt, **generation_kwargs) ## Unpack and the generated text from the LLM response dictionary and print it print(res["choices"][0]["text"]) # res is short for result #output 100% accuracy. [/INST] Aiming is a crucial aspect of CS:GO. Let's start by analyzing your sensitivity settings and crosshair placement. We can also run some aim training drills to improve your precision.