ReMoE: Fully Differentiable Mixture-of-Experts with ReLU Routing Paper • 2412.14711 • Published 6 days ago • 6
Diving into Self-Evolving Training for Multimodal Reasoning Paper • 2412.17451 • Published 2 days ago • 27
X-Prompt: Towards Universal In-Context Image Generation in Auto-Regressive Vision Language Foundation Models Paper • 2412.01824 • Published 23 days ago • 65
CLIP-MoE: Towards Building Mixture of Experts for CLIP with Diversified Multiplet Upcycling Paper • 2409.19291 • Published Sep 28 • 19 • 2
Mirror: A Universal Framework for Various Information Extraction Tasks Paper • 2311.05419 • Published Nov 9, 2023
Enhancing Low-Resource Relation Representations through Multi-View Decoupling Paper • 2312.17267 • Published Dec 26, 2023 • 1
LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training Paper • 2406.16554 • Published Jun 24 • 1
Twin-Merging: Dynamic Integration of Modular Expertise in Model Merging Paper • 2406.15479 • Published Jun 17 • 2
On Giant's Shoulders: Effortless Weak to Strong by Dynamic Logits Fusion Paper • 2406.15480 • Published Jun 17 • 2
ConflictBank: A Benchmark for Evaluating the Influence of Knowledge Conflicts in LLM Paper • 2408.12076 • Published Aug 22 • 12
Learning the Unlearned: Mitigating Feature Suppression in Contrastive Learning Paper • 2402.11816 • Published Feb 19
CLIP-MoE: Towards Building Mixture of Experts for CLIP with Diversified Multiplet Upcycling Paper • 2409.19291 • Published Sep 28 • 19
CLIP-MoE: Towards Building Mixture of Experts for CLIP with Diversified Multiplet Upcycling Paper • 2409.19291 • Published Sep 28 • 19