Empowering Large Language Models for Textual Data Augmentation Paper • 2404.17642 • Published 27 days ago • 2
ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings Paper • 2305.11554 • Published May 19, 2023 • 1 • 2
SynthesizRR: Generating Diverse Datasets with Retrieval Augmentation Paper • 2405.10040 • Published 7 days ago • 2
MuMath-Code: Combining Tool-Use Large Language Models with Multi-perspective Data Augmentation for Mathematical Reasoning Paper • 2405.07551 • Published 10 days ago • 2
MathGenie: Generating Synthetic Data with Question Back-translation for Enhancing Mathematical Reasoning of LLMs Paper • 2402.16352 • Published Feb 26 • 1 • 2
DUQGen: Effective Unsupervised Domain Adaptation of Neural Rankers by Diversifying Synthetic Query Generation Paper • 2404.02489 • Published Apr 3 • 2
Better Synthetic Data by Retrieving and Transforming Existing Datasets Paper • 2404.14361 • Published Apr 22 • 1 • 2
Automated Data Curation for Robust Language Model Fine-Tuning Paper • 2403.12776 • Published Mar 19 • 2
Enhancing Pre-Trained Generative Language Models with Question Attended Span Extraction on Machine Reading Comprehension Paper • 2404.17991 • Published 26 days ago • 3
Improving Retrieval Augmented Open-Domain Question-Answering with Vectorized Contexts Paper • 2404.02022 • Published Apr 2 • 2
Learning to Generate Instruction Tuning Datasets for Zero-Shot Task Adaptation Paper • 2402.18334 • Published Feb 28 • 12 • 2
FlashBack:Efficient Retrieval-Augmented Language Modeling for Long Context Inference Paper • 2405.04065 • Published 17 days ago • 2
On Speculative Decoding for Multimodal Large Language Models Paper • 2404.08856 • Published Apr 13 • 10 • 2
SDSAT: Accelerating LLM Inference through Speculative Decoding with Semantic Adaptive Tokens Paper • 2403.18647 • Published Mar 27 • 2
GliDe with a CaPE: A Low-Hassle Method to Accelerate Speculative Decoding Paper • 2402.02082 • Published Feb 3 • 1 • 2
Hydra: Sequentially-Dependent Draft Heads for Medusa Decoding Paper • 2402.05109 • Published Feb 7 • 2
PRANC: Pseudo RAndom Networks for Compacting deep models Paper • 2206.08464 • Published Jun 16, 2022 • 2
Orchid: Flexible and Data-Dependent Convolution for Sequence Modeling Paper • 2402.18508 • Published Feb 28 • 2
Better & Faster Large Language Models via Multi-token Prediction Paper • 2404.19737 • Published 23 days ago • 61 • 2
IntactKV: Improving Large Language Model Quantization by Keeping Pivot Tokens Intact Paper • 2403.01241 • Published Mar 2 • 2
GEAR: An Efficient KV Cache Compression Recipefor Near-Lossless Generative Inference of LLM Paper • 2403.05527 • Published Mar 8 • 2
Dynamic Memory Compression: Retrofitting LLMs for Accelerated Inference Paper • 2403.09636 • Published Mar 14 • 2 • 2
Long-Context Language Modeling with Parallel Context Encoding Paper • 2402.16617 • Published Feb 26 • 2
LexC-Gen: Generating Data for Extremely Low-Resource Languages with Large Language Models and Bilingual Lexicons Paper • 2402.14086 • Published Feb 21 • 9 • 2
Translation Errors Significantly Impact Low-Resource Languages in Cross-Lingual Learning Paper • 2402.02080 • Published Feb 3 • 3
A Morphologically-Aware Dictionary-based Data Augmentation Technique for Machine Translation of Under-Represented Languages Paper • 2402.01939 • Published Feb 2 • 1 • 1
Generative AI for Math: Part I -- MathPile: A Billion-Token-Scale Pretraining Corpus for Math Paper • 2312.17120 • Published Dec 28, 2023 • 24 • 11
Task-Specific Expert Pruning for Sparse Mixture-of-Experts Paper • 2206.00277 • Published Jun 1, 2022 • 1 • 1
LongNet: Scaling Transformers to 1,000,000,000 Tokens Paper • 2307.02486 • Published Jul 5, 2023 • 79 • 14