LLM2CLIP Collection LLM2CLIP makes SOTA pretrained CLIP modal more SOTA ever. • 7 items • Updated 2 days ago • 35
Large Language Models Can Self-Improve in Long-context Reasoning Paper • 2411.08147 • Published 9 days ago • 58
view article Article Releasing the largest multilingual open pretraining dataset By Pclanglais • 8 days ago • 94
LLM2CLIP: Powerful Language Model Unlock Richer Visual Representation Paper • 2411.04997 • Published 14 days ago • 34
Mixture-of-Transformers: A Sparse and Scalable Architecture for Multi-Modal Foundation Models Paper • 2411.04996 • Published 14 days ago • 48
Balancing Pipeline Parallelism with Vocabulary Parallelism Paper • 2411.05288 • Published 13 days ago • 19
LBPE: Long-token-first Tokenization to Improve Large Language Models Paper • 2411.05504 • Published 13 days ago • 1
Measuring short-form factuality in large language models Paper • 2411.04368 • Published 14 days ago • 1
MagicDec: Breaking the Latency-Throughput Tradeoff for Long Context Generation with Speculative Decoding Paper • 2408.11049 • Published Aug 20 • 12
Llama 3.2 Collection This collection hosts the transformers and original repos of the Llama 3.2 and Llama Guard 3 • 15 items • Updated 28 days ago • 480
FlexAttention for Efficient High-Resolution Vision-Language Models Paper • 2407.20228 • Published Jul 29 • 1
Transformers are RNNs: Fast Autoregressive Transformers with Linear Attention Paper • 2006.16236 • Published Jun 29, 2020 • 3