Training-free Regional Prompting for Diffusion Transformers Paper β’ 2411.02395 β’ Published 21 days ago β’ 24
SG-I2V: Self-Guided Trajectory Control in Image-to-Video Generation Paper β’ 2411.04989 β’ Published 18 days ago β’ 13
TIP-I2V: A Million-Scale Real Text and Image Prompt Dataset for Image-to-Video Generation Paper β’ 2411.04709 β’ Published 20 days ago β’ 25
FasterCache: Training-Free Video Diffusion Model Acceleration with High Quality Paper β’ 2410.19355 β’ Published Oct 25 β’ 23
CCI3.0-HQ: a large-scale Chinese dataset of high quality designed for pre-training large language models Paper β’ 2410.18505 β’ Published Oct 24 β’ 8
Improve Vision Language Model Chain-of-thought Reasoning Paper β’ 2410.16198 β’ Published Oct 21 β’ 17
T2V-Turbo-v2: Enhancing Video Generation Model Post-Training through Data, Reward, and Conditional Guidance Design Paper β’ 2410.05677 β’ Published Oct 8 β’ 14
Story-Adapter: A Training-free Iterative Framework for Long Story Visualization Paper β’ 2410.06244 β’ Published Oct 8 β’ 19
GLEE: A Unified Framework and Benchmark for Language-based Economic Environments Paper β’ 2410.05254 β’ Published Oct 7 β’ 80
Agent S: An Open Agentic Framework that Uses Computers Like a Human Paper β’ 2410.08164 β’ Published Oct 10 β’ 24
EvolveDirector: Approaching Advanced Text-to-Image Generation with Large Vision-Language Models Paper β’ 2410.07133 β’ Published Oct 9 β’ 18
Meissonic: Revitalizing Masked Generative Transformers for Efficient High-Resolution Text-to-Image Synthesis Paper β’ 2410.08261 β’ Published Oct 10 β’ 49
Cavia: Camera-controllable Multi-view Video Diffusion with View-Integrated Attention Paper β’ 2410.10774 β’ Published Oct 14 β’ 24
Animate-X: Universal Character Image Animation with Enhanced Motion Representation Paper β’ 2410.10306 β’ Published Oct 14 β’ 52
Hugging Face communityβs Wikimedia datasets Collection Wikimedia datasets created by the Hugging Face community, not Wikimedia. Sorted by Wikimedia project. β’ 17 items β’ Updated Jun 7 β’ 10
view article Article makeMoE: Implement a Sparse Mixture of Experts Language Model from Scratch By AviSoori1x β’ May 7 β’ 40