Excellent for its size!
#1
by
MurphyGM
- opened
There's a distinct lack of sparse MoE models in the current generation of "lite" LLMs. OLMoE is on par with 7B models while being many times faster, just what the doctor ordered.
I would really like a 3B/21B variant! If there was one and the quality extrapolated, I'd probably use it over any other local model.
MurphyGM
changed discussion title from
Good for its size
to Excellent for its size!