Collections
Discover the best community collections!
Collections including paper arxiv:2402.19173
-
StarCoder 2 and The Stack v2: The Next Generation
Paper • 2402.19173 • Published • 124 -
Griffin: Mixing Gated Linear Recurrences with Local Attention for Efficient Language Models
Paper • 2402.19427 • Published • 49 -
Simple linear attention language models balance the recall-throughput tradeoff
Paper • 2402.18668 • Published • 17 -
Priority Sampling of Large Language Models for Compilers
Paper • 2402.18734 • Published • 15
-
A Survey on Data Selection for Language Models
Paper • 2402.16827 • Published • 3 -
Instruction Tuning with Human Curriculum
Paper • 2310.09518 • Published • 3 -
Fine-Tuning or Retrieval? Comparing Knowledge Injection in LLMs
Paper • 2312.05934 • Published • 1 -
Language Models as Agent Models
Paper • 2212.01681 • Published
-
Griffin: Mixing Gated Linear Recurrences with Local Attention for Efficient Language Models
Paper • 2402.19427 • Published • 49 -
Beyond Language Models: Byte Models are Digital World Simulators
Paper • 2402.19155 • Published • 44 -
StarCoder 2 and The Stack v2: The Next Generation
Paper • 2402.19173 • Published • 124 -
Simple linear attention language models balance the recall-throughput tradeoff
Paper • 2402.18668 • Published • 17
-
LoRA+: Efficient Low Rank Adaptation of Large Models
Paper • 2402.12354 • Published • 5 -
The FinBen: An Holistic Financial Benchmark for Large Language Models
Paper • 2402.12659 • Published • 13 -
TofuEval: Evaluating Hallucinations of LLMs on Topic-Focused Dialogue Summarization
Paper • 2402.13249 • Published • 10 -
TrustLLM: Trustworthiness in Large Language Models
Paper • 2401.05561 • Published • 62