Qwen2.5-1M Collection The long-context version of Qwen2.5, supporting 1M-token context lengths • 2 items • Updated 2 days ago • 80
AI4Privacy_v2 Collection Collection for AI4Privacy Version 2 trained on PII200k • 6 items • Updated Sep 25, 2024 • 4
Demons in the Detail: On Implementing Load Balancing Loss for Training Specialized Mixture-of-Expert Models Paper • 2501.11873 • Published 8 days ago • 61
HALO: Hadamard-Assisted Lossless Optimization for Efficient Low-Precision LLM Training and Fine-Tuning Paper • 2501.02625 • Published 23 days ago • 1
QTIP Quantized Models Collection See https://github.com/Cornell-RelaxML/qtip • 30 items • Updated Dec 9, 2024 • 11
Dolphin 3.0 Collection Dolphin 3.0 is the next generation of the Dolphin series of instruct-tuned models. Designed to be the ultimate general purpose local model. • 7 items • Updated 23 days ago • 60
Falcon3 Collection Falcon3 family of Open Foundation Models is a set of pretrained and instruct LLMs ranging from 1B to 10B parameters. • 40 items • Updated 21 days ago • 80
FP8 LLMs for vLLM Collection Accurate FP8 quantized models by Neural Magic, ready for use with vLLM! • 44 items • Updated Oct 17, 2024 • 62