-
ReAct: Synergizing Reasoning and Acting in Language Models
Paper • 2210.03629 • Published • 11 -
Attention Is All You Need
Paper • 1706.03762 • Published • 34 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 11 -
Jamba: A Hybrid Transformer-Mamba Language Model
Paper • 2403.19887 • Published • 98
Collections
Discover the best community collections!
Collections including paper arxiv:2310.06825
-
Attention Is All You Need
Paper • 1706.03762 • Published • 34 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 11 -
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
Paper • 1910.01108 • Published • 9 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 9
-
Mistral 7B
Paper • 2310.06825 • Published • 41 -
Instruction Tuning with Human Curriculum
Paper • 2310.09518 • Published • 3 -
RAFT: Adapting Language Model to Domain Specific RAG
Paper • 2403.10131 • Published • 58 -
Instruction-tuned Language Models are Better Knowledge Learners
Paper • 2402.12847 • Published • 25
-
Nemotron-4 15B Technical Report
Paper • 2402.16819 • Published • 40 -
Griffin: Mixing Gated Linear Recurrences with Local Attention for Efficient Language Models
Paper • 2402.19427 • Published • 48 -
RWKV: Reinventing RNNs for the Transformer Era
Paper • 2305.13048 • Published • 10 -
Reformer: The Efficient Transformer
Paper • 2001.04451 • Published
-
LoRA: Low-Rank Adaptation of Large Language Models
Paper • 2106.09685 • Published • 24 -
Attention Is All You Need
Paper • 1706.03762 • Published • 34 -
Direct Preference Optimization: Your Language Model is Secretly a Reward Model
Paper • 2305.18290 • Published • 37 -
Lost in the Middle: How Language Models Use Long Contexts
Paper • 2307.03172 • Published • 31
-
Mistral 7B
Paper • 2310.06825 • Published • 41 -
BloombergGPT: A Large Language Model for Finance
Paper • 2303.17564 • Published • 16 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 11 -
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
Paper • 1910.01108 • Published • 9
-
DeepSeek LLM: Scaling Open-Source Language Models with Longtermism
Paper • 2401.02954 • Published • 38 -
Qwen Technical Report
Paper • 2309.16609 • Published • 30 -
GPT-4 Technical Report
Paper • 2303.08774 • Published • 3 -
Gemini: A Family of Highly Capable Multimodal Models
Paper • 2312.11805 • Published • 44