-
Fast Chain-of-Thought: A Glance of Future from Parallel Decoding Leads to Answers Faster
Paper • 2311.08263 • Published • 15 -
Exponentially Faster Language Modelling
Paper • 2311.10770 • Published • 118 -
microsoft/Orca-2-13b
Text Generation • Updated • 18.8k • 663 -
Memory Augmented Language Models through Mixture of Word Experts
Paper • 2311.10768 • Published • 16
Collections
Discover the best community collections!
Collections including paper arxiv:2403.11901
-
Ultra-Long Sequence Distributed Transformer
Paper • 2311.02382 • Published • 2 -
Ziya2: Data-centric Learning is All LLMs Need
Paper • 2311.03301 • Published • 16 -
Relax: Composable Abstractions for End-to-End Dynamic Machine Learning
Paper • 2311.02103 • Published • 16 -
Extending Context Window of Large Language Models via Semantic Compression
Paper • 2312.09571 • Published • 12