hugoleeShanda
's Collections
paper
updated
LLMLingua-2: Data Distillation for Efficient and Faithful Task-Agnostic
Prompt Compression
Paper
•
2403.12968
•
Published
•
24
PERL: Parameter Efficient Reinforcement Learning from Human Feedback
Paper
•
2403.10704
•
Published
•
57
Alignment Studio: Aligning Large Language Models to Particular
Contextual Regulations
Paper
•
2403.09704
•
Published
•
31
RAFT: Adapting Language Model to Domain Specific RAG
Paper
•
2403.10131
•
Published
•
67
Agent-FLAN: Designing Data and Methods of Effective Agent Tuning for
Large Language Models
Paper
•
2403.12881
•
Published
•
16
TnT-LLM: Text Mining at Scale with Large Language Models
Paper
•
2403.12173
•
Published
•
19
RewardBench: Evaluating Reward Models for Language Modeling
Paper
•
2403.13787
•
Published
•
21
LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement
Paper
•
2403.15042
•
Published
•
25
Can large language models explore in-context?
Paper
•
2403.15371
•
Published
•
32
Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient
LLMs Under Compression
Paper
•
2403.15447
•
Published
•
16
InternLM2 Technical Report
Paper
•
2403.17297
•
Published
•
30
Long-form factuality in large language models
Paper
•
2403.18802
•
Published
•
24
Localizing Paragraph Memorization in Language Models
Paper
•
2403.19851
•
Published
•
13
sDPO: Don't Use Your Data All at Once
Paper
•
2403.19270
•
Published
•
40
LLM-ABR: Designing Adaptive Bitrate Algorithms via Large Language Models
Paper
•
2404.01617
•
Published
•
6
Advancing LLM Reasoning Generalists with Preference Trees
Paper
•
2404.02078
•
Published
•
44
Long-context LLMs Struggle with Long In-context Learning
Paper
•
2404.02060
•
Published
•
36
Compact Language Models via Pruning and Knowledge Distillation
Paper
•
2407.14679
•
Published
•
39
DDK: Distilling Domain Knowledge for Efficient Large Language Models
Paper
•
2407.16154
•
Published
•
21