130 results for “topic:vla”
[ICLR 2026] SimpleVLA-RL: Scaling VLA Training via Reinforcement Learning
StarVLA: A Lego-like Codebase for Vision-Language-Action Model Developing
[RSS 2025] Learning to Act Anywhere with Task-centric Latent Actions
Dexbotic: Open-Source Vision-Language-Action Toolbox
🚀🚀🚀A collection of some awesome public projects about Large Language Model(LLM), Vision Language Model(VLM), Vision Language Action(VLA), AI Generated Content(AIGC), the related Datasets and Applications.
InternRobotics' open platform for building generalized navigation foundation models.
🔥 SpatialVLA: a spatial-enhanced vision-language-action model that is trained on 1.1 Million real robot episodes. Accepted at RSS 2025.
[NeurIPS 2025 spotlight] Official implementation for "FutureSightDrive: Thinking Visually with Spatio-Temporal CoT for Autonomous Driving"
[ICLR2026] Official implementation for "JanusVLN: Decoupling Semantics and Spatiality with Dual Implicit Memory for Vision-Language Navigation"
DeepThinkVLA: Enhancing Reasoning Capability of Vision-Language-Action Models
VLA-0: Building State-of-the-Art VLAs with Zero Modification
Running VLA at 30Hz frame rate and 480Hz trajectory frequency
[CVPR 2025, Spotlight] SimLingo (CarLLava): Vision-Only Closed-Loop Autonomous Driving with Language-Action Alignment
Control AI robots. Community-driven UI middleware for controlling robots, recording datasets, training action models. Compatible with SO-100 and SO-101
🌐 Vision-Language-Action Models for Autonomous Driving: Past, Present, and Future
EO: Open-source Unified Embodied Foundation Model Series
CLI for Physical AI Skill Training
[NeurIPS 2025] Flow x RL. "ReinFlow: Fine-tuning Flow Policy with Online Reinforcement Learning". Support VLAs e.g., pi0, pi0.5. Fully open-sourced.
1st place solution of 2025 BEHAVIOR Challenge
[ICLR'25] LLaRA: Supercharging Robot Learning Data for Vision-Language Policy
[CVPR 2026] WAM-Flow: Parallel Coarse-to-Fine Motion Planning via Discrete Flow Matching for Autonomous Driving
Unified Codebase for Advanced World Models.
WAM-Diff: A Masked Diffusion VLA Framework with MoE and Online Reinforcement Learning for Autonomous Driving
[ICLR 2026] Code of "MemoryVLA: Perceptual-Cognitive Memory in Vision-Language-Action Models for Robotic Manipulation"
The official implementation of "DynamicVLA: A Vision-Language-Action Model for Dynamic Object Manipulation". (arXiv 2601.22153)
📚这个仓库是在arxiv上收集的有关VLN,VLA,World Model,SLAM,Gaussian Splatting,非线性优化等相关论文。每天都会自动更新!issue区域是最新10篇论文
A curated list of academic papers and resources on Physical AI — focusing on Vision-Language-Action (VLA) models, world models, embodied ai, and robotic foundation models.
A collection of vision-language-action model post-training methods.
Open & Reproducible Research for Tracking VLAs
🔥This is a curated list of "A survey on Efficient Vision-Language Action Models" research. We will continue to maintain and update the repository, so follow us to keep up with the latest developments!!!