VLA#
开源工作#
- [2025.06] AutoVLA: A Vision-Language-Action Model for End-to-End Autonomous Driving with Adaptive Reasoning and Reinforcement Fine-Tuning
- [2025.06] Unified Vision-Language-Action Model
- [2025.05] Vision-Language-Action Models: Concepts, Progress, Applications and Challenges
- [2025.03] CoT-VLA: Visual Chain-of-Thought Reasoning for Vision-Language-Action Models
- [2024.06] OpenVLA: An Open-Source Vision-Language-Action Model
- [2024.03] Octo: An Open-Source Generalist Robot Policy
- [2023.07] RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control
- [2023.03] PaLM-E: An Embodied Multimodal Language Model
- [2023.03] Diffusion Policy: Visuomotor Policy Learning via Action Diffusion
- [2022.12] RT-1: Robotics Transformer for Real-World Control at Scale
- [2022.08] Do As I Can, Not As I Say:Grounding Language in Robotic Affordances
- [2022.05] Gato: A Generalist Agent