MindVLA-U1: VLA Beats VA with Unified Streaming Architecture for Autonomous Driving

1CUHK MMLab  ·   2Li Auto  ·   3Tsinghua University
*Equal contribution  ·   Project Lead  ·   Corresponding author

MindVLA-U1 is a unified streaming VLA — one backbone, one forward pass for language and continuous action — that on WOD-E2E becomes the first VLA to surpass experienced human drivers (8.20 RFS vs. 8.13 GT) while preserving the natural-language interface that prior VLAs trade away.

AD Capability Radar

MindVLA-U1 is the only architecture that delivers all six driving-VLA capabilities at once. Vision-to-Action (VA) methods are precise on continuous control but lack semantic, temporal, and language interfaces. Prior VLAs trade control precision for those interfaces. MindVLA-U1 keeps continuous control precision and exposes a measurable language → action bridge, streaming temporal context, and adaptive fast/slow compute on one set of weights.

AD capability radar (Figure 1)

Planning Performance on WOD-E2E

We report Rater Feedback Score (RFS) and ADE on Waymo Open Dataset End-to-End. On the validation split, MindVLA-U1 surpasses experienced human drivers for the first time (8.20 vs. 8.13 GT RFS). On the official test split, MindVLA-U1 reaches the highest reported RFS and lowest ADE among all VA/VLA methods.

WOD-E2E Validation Split

MethodADE 3sADE 5sRFS
Vision-to-Action (VA)
VAD3.195.814.45
UniAD6.5010.815.78
RAP-DINO0.972.207.91
Vision-Language-Action (VLA)
Poutine-Base1.272.948.12
Human Driver8.13
MindVLA-U1 (Ours)
Base0.922.147.83
+ Intent-CFG0.862.137.92
+ MoT0.892.117.92
+ RL1.012.288.20

WOD-E2E Test Split (official)

MethodADE 3sADE 5sRFS
Vision-to-Action (VA)
ViT-Adapter-GRU1.202.707.50
Swin-Trajectory1.212.817.54
DiffusionLTF1.362.987.72
Vision-Language-Action (VLA)
Open-LLaMA1.313.227.43
NaiveEMMA1.323.027.53
AutoVLA1.352.967.56
dVLM-AD1.293.027.63
HMVLM1.333.077.74
MindVLA-U1 (Ours)
Base1.162.677.77
+ RL1.092.667.87

Unified Shared Backbone & Streaming Memory

MindVLA-U1 combines a unified shared backbone that jointly performs scene understanding and continuous action generation in a single forward pass over one representation, with a streaming memory paradigm that propagates a compact memory channel across frames in whole driving sessions rather than re-attending to multi-frame video chunks. To let language guide continuous action measurably and to match compute to scene demand, we further develop a language-to-action route and a fast/slow execution scheme that runs on dense or sparse MoT variants of this shared backbone.

MindVLA-U1 unified streaming architecture overview

Overview of MindVLA-U1. A unified shared backbone decodes language and continuous action from one hidden state in a single forward pass; a streaming memory channel carries temporal context across frames, motion-aligned on read and updated after the pass.

Fast/Slow Execution & Streaming Memory Detail

What MoT brings: adaptive compute from a single checkpoint — the same weights run as a fast action-only path (~16 FPS, matched-VA throughput) or as a slow language-then-action path, with no separate models to train or maintain.

Fast/Slow systems on Sparse MoT

Fast/Slow systems on Sparse MoT. Per-modality encoders feed a shared self-attention, per-functionality decoders specialize the outputs. (Dashed: extension slots.)

What streaming memory brings: long-horizon temporal consistency at constant per-frame compute — smoothly evolving plans across frames with no chunk-boundary discontinuities, and improved ADE at every horizon up to 25 s.

Streaming memory bank with motion-aligned tokens

Streaming Memory. FIFO memory channel with motion-aligned and refreshed history tokens; gradients propagate across stream.

Intent-CFG — Language Steers Continuous Action

The unified backbone exposes a measurable language-to-action path. The language head next-token-predicts a driving intent token, whose embedding conditions the action velocity field. Training drops the intent to an unconditional token with probability 0.15, so the same action tokens learn both conditional and unconditional fields; at inference, the two passes are mixed by classifier-free guidance with scale 1.5. Trajectories fan out systematically per intent, while the unconditional pass collapses to a near-GT default.

Intent-CFG with WOD-E2E 3-class intent

(a) Deployed setting: WOD-E2E GT 3-class intent (GO_STRAIGHT / TURN_LEFT / TURN_RIGHT) + unconditional baseline.

Intent-CFG with MindLabel 20-class intent

(b) Extended: MindLabel's 20-class intent vocabulary (cruising, lane_keeping, following, lane_change_left/right, turning_left/right, u_turn, starting, stopping, waiting, accelerating, decelerating, braking, yielding, overtaking, merging, avoiding_obstacle, parking, reversing) + unconditional baseline. Released as part of MindLabel; unexploited in the main-paper.

Reading the figure. Each panel shows BEV trajectories conditioned on every intent class plus an unconditional baseline (left: overview against GT; right: per-intent subplots). Comparing (a) the 3-class deployed setting and (b) MindLabel's 20-class extension on the same scene shows that the conditioning vocabulary, not the scene, controls the modal space — multi-modality is a property of the interface.

Long-Horizon Streaming Consistency

We stitch consecutive clips into long-horizon sequences and overlay predictions against ground-truth poses. Memory propagates across clip boundaries with motion-aware alignment; planning evolves smoothly without per-chunk discontinuities. Sequence-level ADE improves at every horizon (3s/5s/10s/15s/20s/25s) when memory is added.

Case Long-horizon stitched sequence 1/4
Long-horizon sequence
 

Long-Tail Scenes

WOD-E2E targets long-tail scenarios with event frequencies below 0.003%. Across daytime urban and nighttime intersection scenes, MindVLA-U1 grounds its plans in scene-specific perception — the same backbone produces both the 360° BEV trajectory and the natural-language object-level descriptions used for VQA supervision.

Scene A — Daytime urban

Scene A — front-camera panorama Scene A — BEV with dreamed trajectories
Front-camera panorama (top) and BEV with intent-conditioned trajectory variants (bottom).

Scene B — Nighttime intersection

Scene B — front-camera panorama Scene B — object-centric annotations
Panorama (top) and object-centric annotations (bottom): 25 detected entities (11 foreground, 14 background) with natural-language descriptors.
Per-frame streaming inference: 6 consecutive frames

Per-frame streaming inference. Six consecutive frames: front view (top), 5-second BEV trajectory (middle), and per-waypoint confidence heatmaps (bottom). Plans evolve smoothly across the stream.

BibTeX

@article{huang2026mindvlau1,
  title={MindVLA-U1: VLA Beats VA with Unified Streaming Architecture for Autonomous Driving},
  author={Yuzhou Huang and Benjin Zhu and Hengtong Lu and Victor Shea-Jay Huang and Haiming Zhang and Wei Chen and Jifeng Dai and Yan Xie and Hongsheng Li},
  journal={arXiv preprint arXiv:2605.12624},
  year={2026}
}