Dmitry Ryumin's picture

Dmitry Ryumin

DmitryRyumin

AI & ML interests

Machine Learning and Applications, Multi-Modal Understanding

Recent Activity

Organizations

Gradio-Themes-Party's profile picture Gradio-Blocks-Party's profile picture Blog-explorers's profile picture New Era Artificial Intelligence's profile picture ICCV2023's profile picture ZeroGPU Explorers's profile picture Journalists on Hugging Face's profile picture Social Post Explorers's profile picture Dev Mode Explorers's profile picture

DmitryRyumin's activity

upvoted an article 3 days ago
view article
Article

Welcome Llama 4 Maverick & Scout on Hugging Face!

ā€¢ 137
reacted to AdinaY's post with šŸ”„ 10 days ago
view post
Post
1914
AReal-Boba šŸ”„ a fully open RL Frameworks released by AntGroup, an affiliate company of Alibaba.
inclusionAI/areal-boba-67e9f3fa5aeb74b76dcf5f0a
āœØ 7B/32B - Apache2.0
āœØ Outperform on math reasoning
āœØ Replicating QwQ-32B with 200 data under $200
āœØ All-in-one: weights, datasets, code & tech report
  • 1 reply
Ā·
reacted to KaiChen1998's post with šŸ‘ 25 days ago
view post
Post
4815
šŸ“¢ Our EMOVA paper has been accepted by CVPR 2025, and we are glad to release all resources, including code (training & inference), datasets (training & evaluation), and checkpoints (EMOVA-3B/7B/72B)!

šŸ¤— EMOVA is a novel end-to-end omni-modal LLM that can see, hear and speak. Given omni-modal (i.e., textual, visual and speech) inputs, EMOVA can generate both textual and speech responses with vivid emotional controls by utilizing the speech decoder and a style controller.

āœØ EMOVA Highlights
āœ… State-of-the-art omni-modality: EMOVA achieves SoTA comparable results on both vision-language and speech benchmarks simultaneously.
āœ… Device adaptation: our codebase supports training/inference on both NVIDIA GPUs (e.g., A800 & H20) and Ascend NPUs (e.g., 910B3)!
āœ… Modular design: we integrate multiple implementations of vision encoder, vision projector, and language model, even including the most recent DeepSeekMoE-tiny!

šŸ”„ You are all welcome to try and star!
- Project page: https://emova-ollm.github.io/
- Github: https://github.com/emova-ollm/EMOVA
- Demo: Emova-ollm/EMOVA-demo