Text Generation
Transformers
Safetensors
llama
conversational
text-generation-inference
Inference Endpoints

finetune reproduction

#9
by ZhangAI - opened

Thank you for your open source, I am replicating your fine-tuning process according to the code on github. Do the results of train loss=0.16 and eval_loss=0.21 I trained on the 75k dataset match yours? I will continue training on the 110k dataset.
I trained for 4 epochs and indeed started overfitting after the second epoch.

Sign up or log in to comment