Text Generation
Transformers
Safetensors
English
olmoe
conversational
Inference Endpoints

Excellent for its size!

#1
by MurphyGM - opened

There's a distinct lack of sparse MoE models in the current generation of "lite" LLMs. OLMoE is on par with 7B models while being many times faster, just what the doctor ordered.

I would really like a 3B/21B variant! If there was one and the quality extrapolated, I'd probably use it over any other local model.

MurphyGM changed discussion title from Good for its size to Excellent for its size!

Sign up or log in to comment