LLM_Finetuning / config.py
farhananis005's picture
LLM finetuning demo
ec0af28 verified
import os
from dotenv import load_dotenv
from datasets import load_dataset
load_dotenv()
# --- CONFIGURATION ---
MODAL_BASE_URL = "https://mohdfanis--unsloth-model-arena-backend" # Base URL
MY_AUTH_TOKEN = os.environ.get("ARENA_AUTH_TOKEN")
# --- DATASETS ---
print("Loading Hugging Face datasets...")
try:
math_dataset = load_dataset("microsoft/orca-math-word-problems-200k", split="train")
bio_dataset = load_dataset("bio-nlp-umass/bioinstruct", split="train")
print("βœ… Datasets loaded successfully.")
except Exception as e:
print(f"❌ Failed to load datasets: {e}")
math_dataset, bio_dataset = [], []
DATASET_CONFIG = {
"Math": {
"dataset": math_dataset,
"question_col": "question",
"answer_col": "answer"
},
"Bio": {
"dataset": bio_dataset,
"instruction_col": "instruction",
"input_col": "input",
"answer_col": "output"
}
}
# --- MODEL DEFINITIONS ---
BASE_MODELS = {
"Base Llama-3.1 8B Instruct": "unsloth/llama-3.1-8b-instruct-bnb-4bit",
"Base Llama-3 8B Instruct": "unsloth/llama-3-8b-instruct-bnb-4bit",
"Base Llama-2 7B Chat": "unsloth/llama-2-7b-chat-bnb-4bit",
"Base Mistral 7B Instruct": "unsloth/mistral-7b-v0.3-instruct-bnb-4bit",
"Base Qwen-2 7B Instruct": "unsloth/qwen2-7B-instruct-bnb-4bit",
"Base Gemma-2 9B Instruct": "unsloth/gemma-2-9b-it-bnb-4bit",
"Base Gemma 7B Instruct": "unsloth/gemma-7b-it-bnb-4bit",
}
FINETUNED_MATH = {
"Finetuned Llama-3.1 8B (e3) - MATH": "farhananis005/lora-llama-3.1-8b-Math-e3",
"Finetuned Llama-3.1 8B (e1) - MATH": "farhananis005/lora-llama-3.1-8b-Math-e1",
"Finetuned Llama-3 8B (e3) - MATH": "farhananis005/lora-llama-3-8b-Math-e3",
"Finetuned Llama-3 8B (e1) - MATH": "farhananis005/lora-llama-3-8b-Math-e1",
"Finetuned Llama-2 7B (e3) - MATH": "farhananis005/lora-llama-2-7b-Math-e3",
"Finetuned Llama-2 7B (e1) - MATH": "farhananis005/lora-llama-2-7b-Math-e1",
"Finetuned Mistral 7B (e3) - MATH": "farhananis005/lora-mistral-7b-v0.3-Math-e3",
"Finetuned Mistral 7B (e1) - MATH": "farhananis005/lora-mistral-7b-v0.3-Math-e1",
"Finetuned Qwen-2 7B (e3) - MATH": "farhananis005/lora-qwen-2-7b-Math-e3",
"Finetuned Qwen-2 7B (e1) - MATH": "farhananis005/lora-qwen-2-7b-Math-e1",
"Finetuned Gemma-2 9B (e3) - MATH": "farhananis005/lora-gemma-2-9b-Math-e3",
"Finetuned Gemma-2 9B (e1) - MATH": "farhananis005/lora-gemma-2-9b-Math-e1",
"Finetuned Gemma 7B (e3) - MATH": "farhananis005/lora-gemma-7b-Math-e3",
"Finetuned Gemma 7B (e1) - MATH": "farhananis005/lora-gemma-7b-Math-e1",
}
FINETUNED_BIO = {
"Finetuned Llama-3.1 8B (e3) - BIO": "farhananis005/lora-llama-3.1-8b-Bio-e3",
"Finetuned Llama-3.1 8B (e1) - BIO": "farhananis005/lora-llama-3.1-8b-Bio-e1",
"Finetuned Llama-3 8B (e3) - BIO": "farhananis005/lora-llama-3-8b-Bio-e3",
"Finetuned Llama-3 8B (e1) - BIO": "farhananis005/lora-llama-3-8b-Bio-e1",
"Finetuned Llama-2 7B (e3) - BIO": "farhananis005/lora-llama-2-7b-Bio-e3",
"Finetuned Llama-2 7B (e1) - BIO": "farhananis005/lora-llama-2-7b-Bio-e1",
"Finetuned Mistral 7B (e3) - BIO": "farhananis005/lora-mistral-7b-v0.3-Bio-e3",
"Finetuned Mistral 7B (e1) - BIO": "farhananis005/lora-mistral-7b-v0.3-Bio-e1",
"Finetuned Qwen-2 7B (e3) - BIO": "farhananis005/lora-qwen-2-7b-Bio-e3",
"Finetuned Qwen-2 7B (e1) - BIO": "farhananis005/lora-qwen-2-7b-Bio-e1",
"Finetuned Gemma-2 9B (e3) - BIO": "farhananis005/lora-gemma-2-9b-Bio-e3",
"Finetuned Gemma-2 9B (e1) - BIO": "farhananis005/lora-gemma-2-9b-Bio-e1",
"Finetuned Gemma 7B (e3) - BIO": "farhananis005/lora-gemma-7b-Bio-e3",
"Finetuned Gemma 7B (e1) - BIO": "farhananis005/lora-gemma-7b-Bio-e1",
}
ALL_MODELS = {
"Math": {"-- Select Math Model --": None, **BASE_MODELS, **FINETUNED_MATH},
"Bio": {"-- Select Bio Model --": None, **BASE_MODELS, **FINETUNED_BIO}
}