Text Generation
Transformers
Safetensors
llama
conversational
Inference Endpoints
text-generation-inference

How may gpu and gpu time used for this training?

#3
by aisensiy - opened

This is a really cool model, may I know which kind of gpu do you use for this and how many hours does it take. Thanks.

Hi @jondurbin Your DPO training is based on Lora but not full parameters, right? Does DPO training on full parameters will make difference? better performance?

Sign up or log in to comment