🪐 SmolLM Collection A series of smol LLMs: 135M, 360M and 1.7B. We release base and Instruct models as well as the training corpus and some WebGPU demos • 12 items • Updated 5 days ago • 97
view article Article How I train a LoRA: m3lt style training overview By alvdansen • 20 days ago • 36
view article Article Recommendation to Revisit the Diffuser Default LoRA Parameters By alvdansen • about 1 month ago • 9
view article Article Training and Finetuning Embedding Models with Sentence Transformers v3 May 28 • 123
Direct Nash Optimization: Teaching Language Models to Self-Improve with General Preferences Paper • 2404.03715 • Published Apr 4 • 58
LLM Leaderboard best models ❤️🔥 Collection A daily uploaded list of models with best evaluations on the LLM leaderboard: • 264 items • Updated 29 days ago • 352