Model Details
I finetuned PygmalionAI/pygmalion-6b with QLora for 24 hours on 250k samples. Collected from SODA and Teacher GPT dataset. My first attempt on making LLM model as an entry to Chai competition.
Model Description
- Model type: Chatbot
- Finetuned from model : PygmalionAI/pygmalion-6b
Model Sources
Pygmalion-6b: https://huggingface.co/PygmalionAI/pygmalion-6b
Training Details
Training Data
For the training data I use 20% of SODA dadtaset mixed with TeacherGPT roleplay dataset.
Training Procedure
The model was trained for 24 hours on RTX4090.
Training Hyperparameters
Training param
batch_size = 128,
micro_batch_size = 4,
num_epochs = 1,
learning_rate = 3e-4,
cutoff_len = 512,
val_set_size = 0finetune method
finetune_method = "qlora"
prefix tuning hyperparams
num_virtual_tokens = 32
lora hyperparams
lora_r = 16,
lora_alpha = 16,
lora_dropout = 0.05,
lora_target_modules = "q_proj k_proj v_proj"llm hyperparams
bf16 = False,
load_in_8bit = False,
group_by_length = False ,
resume_from_checkpoint = None
Results
Me: Hi Nathan, how are you doing today
Nathan: I'm fine...
Me: Then tell me about your day.
Nathan:
It was good. We had a lot of fun in school and then we went to the park afterwards.