BOLT: Bootstrap Long Chain-of-Thought in Language Models without Distillation Paper • 2502.03860 • Published Feb 6 • 24
Continual Training of Language Models for Few-Shot Learning Paper • 2210.05549 • Published Oct 11, 2022
Demystifying Domain-adaptive Post-training for Financial LLMs Paper • 2501.04961 • Published Jan 9 • 11
OmniFlow: Any-to-Any Generation with Multi-Modal Rectified Flows Paper • 2412.01169 • Published Dec 2, 2024 • 13
MobileAIBench: Benchmarking LLMs and LMMs for On-Device Use Cases Paper • 2406.10290 • Published Jun 12, 2024
ULIP: Learning a Unified Representation of Language, Images, and Point Clouds for 3D Understanding Paper • 2212.05171 • Published Dec 10, 2022
APIGen: Automated Pipeline for Generating Verifiable and Diverse Function-Calling Datasets Paper • 2406.18518 • Published Jun 26, 2024 • 24
xGen-MM (BLIP-3): A Family of Open Large Multimodal Models Paper • 2408.08872 • Published Aug 16, 2024 • 99
ThinK: Thinner Key Cache by Query-Driven Pruning Paper • 2407.21018 • Published Jul 30, 2024 • 32
CHAMP: A Competition-level Dataset for Fine-Grained Analyses of LLMs' Mathematical Reasoning Capabilities Paper • 2401.06961 • Published Jan 13, 2024
Towards Understanding the Behaviors of Optimal Deep Active Learning Algorithms Paper • 2101.00977 • Published Dec 29, 2020
Can Large Language Models Explain Themselves? A Study of LLM-Generated Self-Explanations Paper • 2310.11207 • Published Oct 17, 2023