OpenVLA: An Open-Source Vision-Language-Action Model Paper • 2406.09246 • Published Jun 13, 2024 • 36
CogACT: A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation Paper • 2411.19650 • Published Nov 29, 2024
Diffusion-VLA: Scaling Robot Foundation Models via Unified Diffusion and Autoregression Paper • 2412.03293 • Published Dec 4, 2024