Mitigating Premature Discretization with Progressive Quantization for Robust Vector Tokenization
arXiv cs.LG / 2026/3/25
💬 オピニオンIdeas & Deep AnalysisModels & Research
要点
- The paper identifies a key weakness in existing vector quantization (VQ) approaches for multimodal tokenization: “Premature Discretization,” where discrete quantization is applied before the encoder has learned the data manifold.
- It introduces Progressive Quantization (ProVQ), treating quantization hardness as a training curriculum that gradually anneals from continuous latents to discrete tokens.
- Experiments show ProVQ improves reconstruction and generative performance on ImageNet-1K and ImageNet-100, indicating benefits for image generative modeling.
- The method also performs strongly on complex biological sequence modeling, setting a new state-of-the-art performance ceiling for protein structure tokenization on StrutTokenBench.
