You Do Not Fully Utilize Transformer's Representation Capacity Paper • 2502.09245 • Published 14 days ago • 33
Train Small, Infer Large: Memory-Efficient LoRA Training for Large Language Models Paper • 2502.13533 • Published 8 days ago • 8
Enhancing Cognition and Explainability of Multimodal Foundation Models with Self-Synthesized Data Paper • 2502.14044 • Published 7 days ago • 7
From RAG to Memory: Non-Parametric Continual Learning for Large Language Models Paper • 2502.14802 • Published 7 days ago • 11
Discovering highly efficient low-weight quantum error-correcting codes with reinforcement learning Paper • 2502.14372 • Published 7 days ago • 35
How Much Knowledge Can You Pack into a LoRA Adapter without Harming LLM? Paper • 2502.14502 • Published 7 days ago • 77
TAG: A Decentralized Framework for Multi-Agent Hierarchical Reinforcement Learning Paper • 2502.15425 • Published 6 days ago • 7
Reflective Planning: Vision-Language Models for Multi-Stage Long-Horizon Robotic Manipulation Paper • 2502.16707 • Published 3 days ago • 10
Curie: Toward Rigorous and Automated Scientific Experimentation with AI Agents Paper • 2502.16069 • Published 5 days ago • 14
R18-Novels-galgame Collection Novels; galgame; visual novels; 小说; 剧本; roleplay; sq; ghs; hentai; R18; NSFW; 涩情; 涩涩; 瑟瑟; 色色; 可爱; 美少女 • 61 items • Updated 9 days ago • 36
Model Merging Collection Model Merging is a very popular technique nowadays in LLM. Here is a chronological list of papers on the space that will help you get started with it! • 30 items • Updated Jun 12, 2024 • 231
Chameleon: Mixed-Modal Early-Fusion Foundation Models Paper • 2405.09818 • Published May 16, 2024 • 131
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits Paper • 2402.17764 • Published Feb 27, 2024 • 609