TimeTok: Granularity-Controllable Time-Series Generation via Hierarchical Tokenization

arXiv cs.AI / 5/5/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • TimeTok (arXiv:2605.01418v1) proposes a granularity-controllable time-series generative framework (GC-TSG) that can synthesize outputs at arbitrary target temporal resolutions from coarse inputs or from scratch.
  • The method uses hierarchical tokenization to convert a time series into an ordered token sequence spanning coarse-to-fine granularities, then performs autoregressive generation across these levels to produce token blocks decoded into continuous signals.
  • By controlling the number of generated token blocks, users get explicit, direct control over output detail (temporal granularity) within a single unified generation framework.
  • Experiments indicate TimeTok performs especially well on GC-TSG tasks while also reaching state-of-the-art results on standard time-series generation benchmarks.
  • The paper demonstrates TimeTok as a foundational tokenizer by training on multiple datasets with differing temporal granularities and shows strong transferability, outperforming models trained on single-dataset settings.

Abstract

Time-series generative models often lack control over temporal granularity, forcing users to accept whatever granularity the model produces. To enable truly user-driven generation, we introduce TimeTok, a unified framework for Granularity-Controllable Time-Series Generation (GC-TSG), which generates time series at any target granularity from any coarser input (e.g., rough sketches) or from scratch. At the core of TimeTok is a hierarchical tokenization strategy that maps time series into an ordered sequence of tokens, from coarse to fine temporal granularity. Our autoregressive generation process operates across these granularity levels, producing token blocks that are decoded back into continuous time series. This design naturally enables GC-TSG - including standard generation - within a single framework, where controlling the number of token blocks provides explicit control over output detail. Experiments show that TimeTok excels at GC-TSG tasks while achieving state-of-the-art performance in standard generation. Furthermore, we showcase TimeTok's potential as a foundational tokenizer by training on multiple datasets with heterogeneous temporal granularities, verifying strong transferability that consistently outperforms models trained on individual datasets. To our knowledge, this is the first unified framework that covers the full generative spectrum for time series, offering a valuable foundation for models that benefit from diverse temporal granularities.