Reward-aware Preference Optimization: A Unified Mathematical Framework for Model Alignment Paper • 2502.00203 • Published Jan 31
Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models Paper • 2504.03624 • Published 19 days ago • 13
Nemotron-H: A Family of Accurate and Efficient Hybrid Mamba-Transformer Models Paper • 2504.03624 • Published 19 days ago • 13
Minitron Collection A family of compressed models obtained via pruning and knowledge distillation • 12 items • Updated about 5 hours ago • 61
DevQuasar/nvidia.Llama-3_1-Nemotron-Ultra-253B-v1-GGUF Text Generation • Updated 7 days ago • 3.52k • 7
Llama Nemotron Collection Open, Production-ready Enterprise Models • 4 items • Updated about 5 hours ago • 37
NeMo-Aligner: Scalable Toolkit for Efficient Model Alignment Paper • 2405.01481 • Published May 2, 2024 • 31