|
--- |
|
## Model Details |
|
|
|
I finetuned PygmalionAI/pygmalion-6b with QLora for 24 hours on 250k samples. Collected from SODA and Teacher GPT dataset. My first attempt on making LLM model as an entry to Chai competition. |
|
|
|
### Model Description |
|
|
|
- **Model type:** Chatbot |
|
- **Finetuned from model :** PygmalionAI/pygmalion-6b |
|
|
|
### Model Sources |
|
|
|
Pygmalion-6b: https://huggingface.co/PygmalionAI/pygmalion-6b |
|
|
|
## Training Details |
|
|
|
### Training Data |
|
|
|
For the training data I use 20% of SODA dadtaset mixed with TeacherGPT roleplay dataset. |
|
|
|
### Training Procedure |
|
|
|
The model was trained for 24 hours on RTX4090. |
|
|
|
#### Training Hyperparameters |
|
|
|
- Training param |
|
>batch_size = 128, |
|
>micro_batch_size = 4, |
|
>num_epochs = 1, |
|
>learning_rate = 3e-4, |
|
>cutoff_len = 512, |
|
>val_set_size = 0 |
|
|
|
- finetune method |
|
>finetune_method = "qlora" |
|
|
|
- prefix tuning hyperparams |
|
>num_virtual_tokens = 32 |
|
|
|
- lora hyperparams |
|
>lora_r = 16, |
|
>lora_alpha = 16, |
|
>lora_dropout = 0.05, |
|
>lora_target_modules = "q_proj k_proj v_proj" |
|
|
|
- llm hyperparams |
|
>bf16 = False, |
|
>load_in_8bit = False, |
|
>group_by_length = False , |
|
>resume_from_checkpoint = None |
|
|
|
### Results |
|
|
|
Me: Hi Nathan, how are you doing today |
|
Nathan: I'm fine... |
|
Me: Then tell me about your day. |
|
Nathan: |
|
-- |
|
It was good. We had a lot of fun in school and then we went to the park afterwards. |