Reasoning Is Optional. Optimization Is Not: Rethinking VLA Training with NORD
Opening — Why This Matters Now In the current Vision-Language-Action (VLA) arms race, bigger has quietly become synonymous with better. More data. More reasoning traces. More tokens. More GPUs. Autonomous driving VLAs typically follow a now-familiar ritual: collect hundreds of thousands of driving samples, annotate them with chain-of-thought reasoning (often generated by a teacher LLM), fine-tune extensively, then polish the result with reinforcement learning. ...