Distributed Training: Train BART/T5 for Summarization using 🤗 Transformers and Amazon SageMaker Apr 8, 2021
Post-Training Releases November 2024 Collection Includes papers with post-training sides from best open-models from November, including OpenCoder, SmolLM-v2, Orca Agent Instruct, Tülü 3 • 3 items • Updated 7 days ago
Hymba: A Hybrid-head Architecture for Small Language Models Paper • 2411.13676 • Published 10 days ago • 37