-
AutoNumerics-Zero: Automated Discovery of State-of-the-Art Mathematical Functions
Paper • 2312.08472 • Published • 2 -
MathVerse: Does Your Multi-modal LLM Truly See the Diagrams in Visual Math Problems?
Paper • 2403.14624 • Published • 50 -
ChatGLM-Math: Improving Math Problem-Solving in Large Language Models with a Self-Critique Pipeline
Paper • 2404.02893 • Published • 19 -
Rho-1: Not All Tokens Are What You Need
Paper • 2404.07965 • Published • 80
Collections
Discover the best community collections!
Collections including paper arxiv:2404.07965
-
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Paper • 2403.03507 • Published • 177 -
RAFT: Adapting Language Model to Domain Specific RAG
Paper • 2403.10131 • Published • 65 -
LlamaFactory: Unified Efficient Fine-Tuning of 100+ Language Models
Paper • 2403.13372 • Published • 58 -
InternLM2 Technical Report
Paper • 2403.17297 • Published • 26
-
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
Paper • 2402.17764 • Published • 571 -
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Paper • 2403.03507 • Published • 177 -
Griffin: Mixing Gated Linear Recurrences with Local Attention for Efficient Language Models
Paper • 2402.19427 • Published • 50 -
ResLoRA: Identity Residual Mapping in Low-Rank Adaption
Paper • 2402.18039 • Published • 10
-
Ultralytics/YOLOv8
Object Detection • Updated • 57 -
YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information
Paper • 2402.13616 • Published • 44 -
Gradio: Hassle-Free Sharing and Testing of ML Models in the Wild
Paper • 1906.02569 • Published -
SpeechAlign: a Framework for Speech Translation Alignment Evaluation
Paper • 2309.11585 • Published
-
How to Train Data-Efficient LLMs
Paper • 2402.09668 • Published • 34 -
LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement
Paper • 2403.15042 • Published • 24 -
MAGID: An Automated Pipeline for Generating Synthetic Multi-modal Datasets
Paper • 2403.03194 • Published • 11 -
Orca-Math: Unlocking the potential of SLMs in Grade School Math
Paper • 2402.14830 • Published • 23
-
World Model on Million-Length Video And Language With RingAttention
Paper • 2402.08268 • Published • 33 -
Improving Text Embeddings with Large Language Models
Paper • 2401.00368 • Published • 75 -
Chain-of-Thought Reasoning Without Prompting
Paper • 2402.10200 • Published • 91 -
FiT: Flexible Vision Transformer for Diffusion Model
Paper • 2402.12376 • Published • 47
-
Medusa: Simple LLM Inference Acceleration Framework with Multiple Decoding Heads
Paper • 2401.10774 • Published • 50 -
APAR: LLMs Can Do Auto-Parallel Auto-Regressive Decoding
Paper • 2401.06761 • Published • 1 -
Infinite-LLM: Efficient LLM Service for Long Context with DistAttention and Distributed KVCache
Paper • 2401.02669 • Published • 11 -
MambaByte: Token-free Selective State Space Model
Paper • 2401.13660 • Published • 47
-
Lost in the Middle: How Language Models Use Long Contexts
Paper • 2307.03172 • Published • 33 -
Efficient Estimation of Word Representations in Vector Space
Paper • 1301.3781 • Published • 6 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 12 -
Attention Is All You Need
Paper • 1706.03762 • Published • 36
-
SOLAR 10.7B: Scaling Large Language Models with Simple yet Effective Depth Up-Scaling
Paper • 2312.15166 • Published • 55 -
PowerInfer: Fast Large Language Model Serving with a Consumer-grade GPU
Paper • 2312.12456 • Published • 40 -
Cached Transformers: Improving Transformers with Differentiable Memory Cache
Paper • 2312.12742 • Published • 11 -
Mini-GPTs: Efficient Large Language Models through Contextual Pruning
Paper • 2312.12682 • Published • 7