DreamCraft3D: Hierarchical 3D Generation with Bootstrapped Diffusion Prior Paper • 2310.16818 • Published Oct 25, 2023 • 32
Janus-Pro: Unified Multimodal Understanding and Generation with Data and Model Scaling Paper • 2501.17811 • Published 21 days ago • 5
Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention Paper • 2502.11089 • Published 3 days ago • 108
Llasa Collection TTS foundation model compatible with Llama framework (160k hours tokenized speech data released) • 11 items • Updated 6 days ago • 6
Step-Audio Collection Step-Audio model family, including Audio-Tokenizer, Audio-Chat and TTS • 3 items • Updated 2 days ago • 20
OLMoE (January 2025) Collection Improved OLMoE for iOS app. Read more: https://allenai.org/blog/olmoe-app • 10 items • Updated 8 days ago • 9
Ovis2 Collection Our latest advancement in multi-modal large language models (MLLMs) • 8 items • Updated 3 days ago • 38
Granite Experiments Collection Experimental projects under consideration for the Granite family. • 6 items • Updated 33 minutes ago • 9
Hibiki fr-en Collection Hibiki is a model for streaming speech translation , which can run on device! See https://github.com/kyutai-labs/hibiki. • 5 items • Updated 13 days ago • 48
QLIP Collection QLIP is a family of image tokenizers with SOTA reconstruction quality and zero-shot image understanding. • 3 items • Updated 13 days ago • 7
Tulu 3 Models Collection All models released with Tulu 3 -- state of the art open post-training recipes. • 11 items • Updated 7 days ago • 90
TinySwallow Collection Compact Japanese models trained with "TAID: Temporally Adaptive Interpolated Distillation for Efficient Knowledge Transfer in Language Models" • 5 items • Updated 21 days ago • 16
YuE Collection YuE is a groundbreaking series of open-source foundation models designed for music generation, led by HKUST. • 7 items • Updated 21 days ago • 2
Qwen2.5-1M Collection The long-context version of Qwen2.5, supporting 1M-token context lengths • 2 items • Updated 24 days ago • 100