SmolVLA: A Vision-Language-Action Model for Affordable and Efficient Robotics Paper • 2506.01844 • Published 24 days ago • 100
3D-VLA: A 3D Vision-Language-Action Generative World Model Paper • 2403.09631 • Published Mar 14, 2024 • 10
QUAR-VLA: Vision-Language-Action Model for Quadruped Robots Paper • 2312.14457 • Published Dec 22, 2023 • 1
Diffusion-VLA: Scaling Robot Foundation Models via Unified Diffusion and Autoregression Paper • 2412.03293 • Published Dec 4, 2024
Bi-VLA: Vision-Language-Action Model-Based System for Bimanual Robotic Dexterous Manipulations Paper • 2405.06039 • Published May 9, 2024 • 1
A Dual Process VLA: Efficient Robotic Manipulation Leveraging VLM Paper • 2410.15549 • Published Oct 21, 2024
VLA-Cache: Towards Efficient Vision-Language-Action Model via Adaptive Token Caching in Robotic Manipulation Paper • 2502.02175 • Published Feb 4
VLA-OS: Structuring and Dissecting Planning Representations and Paradigms in Vision-Language-Action Models Paper • 2506.17561 • Published 5 days ago
RaceVLA: VLA-based Racing Drone Navigation with Human-like Behaviour Paper • 2503.02572 • Published Mar 4
VLA-RL: Towards Masterful and General Robotic Manipulation with Scalable Reinforcement Learning Paper • 2505.18719 • Published May 24
RoboTwin 2.0: A Scalable Data Generator and Benchmark with Strong Domain Randomization for Robust Bimanual Robotic Manipulation Paper • 2506.18088 • Published 4 days ago • 7