HiFlow: Tokenization-Free Scale-Wise Autoregressive Policy Learning via Flow Matching

arXiv cs.RO / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Hierarchical Flow Policy (HiFlow), a tokenization-free coarse-to-fine autoregressive approach for visuomotor policy learning that models raw continuous robot actions directly.
  • It avoids discrete action tokenizers that can add quantization error and previously required multi-stage training pipelines, instead generating multi-scale continuous targets via temporal pooling over action chunks.
  • HiFlow builds coarse action summaries by averaging contiguous windows and refines them at finer temporal resolutions, enabling an end-to-end single-stage training setup.
  • Experiments on MimicGen, RoboTwin 2.0, and real-world environments report consistent performance gains over both diffusion-based policies and tokenization-based autoregressive baselines.

Abstract

Coarse-to-fine autoregressive modeling has recently shown strong promise for visuomotor policy learning, combining the inference efficiency of autoregressive methods with the global trajectory coherence of diffusion-based policies. However, existing approaches rely on discrete action tokenizers that map continuous action sequences to codebook indices, a design inherited from image generation where learned compression is necessary for high-dimensional pixel data. We observe that robot actions are inherently low-dimensional continuous vectors, for which such tokenization introduces unnecessary quantization error and a multi-stage training pipeline. In this work, we propose Hierarchical Flow Policy (HiFlow), a tokenization-free coarse-to-fine autoregressive policy that operates directly on raw continuous actions. HiFlow constructs multi-scale continuous action targets from each action chunk via simple temporal pooling. Specifically, it averages contiguous action windows to produce coarse summaries that are refined at finer temporal resolutions. The entire model is trained end-to-end in a single stage, eliminating the need for a separate tokenizer. Experiments on MimicGen, RoboTwin 2.0, and real-world environments demonstrate that HiFlow consistently outperforms existing methods including diffusion-based and tokenization-based autoregressive policies.