Gemma 3 QAT Collection Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory • 8 items • Updated 7 days ago • 108
HoT: Highlighted Chain of Thought for Referencing Supporting Facts from Inputs Paper • 2503.02003 • Published Mar 3 • 46
LivePortrait: Efficient Portrait Animation with Stitching and Retargeting Control Paper • 2407.03168 • Published Jul 3, 2024 • 3
Qwen 2.5 Coder Llamafiles (<50B) Collection Llamafiles for the smaller Qwen 2.5 Coder models • 6 items • Updated Feb 25 • 1
Qwen 2.5 Llamafiles (<50B) Collection Llamafiles for the smaller Qwen 2.5 text only models • 6 items • Updated Feb 25 • 1
Deepseek Distilled Llamafiles (<50B) Collection Llamafiles for the smaller Deepseek Distilled Models • 5 items • Updated Feb 25 • 2
DeepHermes Collection Preview models of hybrid reasoner Hermes series • 6 items • Updated 27 days ago • 27
DeepSeek R1 (All Versions) Collection DeepSeek R1 - the most powerful reasoning open-source model - available in GGUF, original & 4-bit formats. Includes Llama & Qwen distilled models. • 29 items • Updated 4 days ago • 216
Gemma 3 Collection All versions of Google's new multimodal models in 1B, 4B, 12B, and 27B sizes. In GGUF, dynamic 4-bit and 16-bit formats. • 29 items • Updated 4 days ago • 50
LLMs Can Easily Learn to Reason from Demonstrations Structure, not content, is what matters! Paper • 2502.07374 • Published Feb 11 • 39
Hibiki fr-en Collection Hibiki is a model for streaming speech translation , which can run on device! See https://github.com/kyutai-labs/hibiki. • 5 items • Updated Feb 6 • 52
SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training Paper • 2501.17161 • Published Jan 28 • 120