|
--- |
|
license: other |
|
license_name: phi-3 |
|
license_link: https://huggingface.co/microsoft/Phi-3-mini-128k-instruct/raw/main/LICENSE |
|
datasets: |
|
- m-a-p/CodeFeedback-Filtered-Instruction |
|
tags: |
|
- phi |
|
- phi-3 |
|
- '3' |
|
- code |
|
--- |
|
Finetune of [microsoft/Phi-3-mini-128k-instruct](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct) on [m-a-p/CodeFeedback-Filtered-Instruction](https://huggingface.co/datasets/m-a-p/CodeFeedback-Filtered-Instruction) for ~9-10h using a single 3090 24GB. |
|
|
|
Due to limited resources and time, the training was only on half (0.5136) of the epoch. |
|
|
|
``` |
|
train_loss: 0.43311 |
|
``` |
|
|
|
``` |
|
learning_rate=5e-5, |
|
lr_scheduler_type="cosine", |
|
max_length=1024, |
|
max_prompt_length=512, |
|
overwrite_output_dir=True, |
|
beta=0.1, |
|
gradient_accumulation_steps=8, |
|
optim="adamw_torch", |
|
num_train_epochs=1, |
|
evaluation_strategy="steps", |
|
eval_steps=0.2, |
|
logging_steps=1, |
|
warmup_steps=50, |
|
fp16=True, |
|
save_steps=50 |
|
``` |