-
BlackMamba: Mixture of Experts for State-Space Models
Paper • 2402.01771 • Published • 26 -
OpenMoE: An Early Effort on Open Mixture-of-Experts Language Models
Paper • 2402.01739 • Published • 29 -
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
Paper • 2401.06066 • Published • 55
David Samuel
Davidsamuel101
AI & ML interests
NLP, Computer Vision
Recent Activity
updated
a model
18 days ago
bookbot/wav2vec2-xls-r-300m-swahili-cv-fleurs-alffa-alphabets-phonemes-bookbot
updated
a model
18 days ago
bookbot/wav2vec2-xls-r-300m-swahili-cv-fleurs-alffa-alphabets-phonemes-bookbot
updated
a model
18 days ago
bookbot/wav2vec2-xls-r-300m-swahili-cv-fleurs-alffa-alphabets-phonemes-bookbot