Text Generation
Transformers
Safetensors
English
mistral
axolotl
generated_from_trainer
Mistral
instruct
finetune
chatml
gpt4
synthetic data
science
physics
chemistry
biology
math
conversational
Eval Results
Inference Endpoints
text-generation-inference

Very Nice. I Hope Someone Tries Fine-Tuning Phi-2 Like This.

#4
by Phil337 - opened

I like the response style of this LLM more than any other I've tried.

They're short and to the point, but unlike others with succinct responses, this LLM gives longer responses when the circumstances or wording of the prompt call for it.

I can't find a usable Phi-2 because people are trying to make it too wordy (e.g. orange2), resulting in long strings of hallucinations, weird tangents, gratuitous moralizing... But without fine-tuning with a diverse dataset Phi-2 struggles to perform basic tasks.

Owner

Hi @Phil337

Thanks for the feedback :)

As you mentioned, the model tends to provide longer responses compared to other LLMs available. I ensured this by filtering datasets to generate longer responses and by creating multiple-choice datasets using Mixtral (rest assured, I provided the correct answers to it, don't worry :D) to explain the answers in depth rather than just providing them.

Regarding the phi-2 request, I may actually do it. I'll tag you if I do :)

Owner

Hi @Phil337

The Phi-2 version of the Einstein v4 is on the way. I hope it turns out to be something good!

It will be in this repo when it's finished: https://huggingface.co/Weyaxi/Einstein-v4-phi2

image.png

@Weyaxi Awesome! Looking forward to trying it out.

Sign up or log in to comment