-
Rho-1: Not All Tokens Are What You Need
Paper • 2404.07965 • Published • 79 -
VASA-1: Lifelike Audio-Driven Talking Faces Generated in Real Time
Paper • 2404.10667 • Published • 12 -
Instruction-tuned Language Models are Better Knowledge Learners
Paper • 2402.12847 • Published • 25 -
DoRA: Weight-Decomposed Low-Rank Adaptation
Paper • 2402.09353 • Published • 18
Collections
Discover the best community collections!
Collections including paper arxiv:2404.05829
-
Continual Pre-Training for Cross-Lingual LLM Adaptation: Enhancing Japanese Language Capabilities
Paper • 2404.17790 • Published • 1 -
Construction of Domain-specified Japanese Large Language Model for Finance through Continual Pre-training
Paper • 2404.10555 • Published • 2 -
Pretraining and Updating Language- and Domain-specific Large Language Model: A Case Study in Japanese Business Domain
Paper • 2404.08262 • Published • 1 -
Heron-Bench: A Benchmark for Evaluating Vision Language Models in Japanese
Paper • 2404.07824 • Published • 2