XVLA Collection X-VLA is a soft-prompted Transformer for cross-embodiment robot learning • 6 items • Updated 2 days ago • 8
view article Article SmolVLA: Efficient Vision-Language-Action Model trained on Lerobot Community Data +7 Jun 3 • 289
view article Article π0 and π0-FAST: Vision-Language-Action Models for General Robot Control +2 Feb 4 • 184
One to rule them all: natural language to bind communication, perception and action Paper • 2411.15033 • Published Nov 22, 2024 • 3
Diffusion Policy: Visuomotor Policy Learning via Action Diffusion Paper • 2303.04137 • Published Mar 7, 2023 • 5