Coarse-Guided Visual Generation via Weighted h-Transform Sampling
arXiv cs.CV / 3/13/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces a training-free method for coarse-guided visual generation by applying the h-transform to diffusion sampling, adding a drift term to steer the process toward fine samples.
- It modifies the transition probabilities at each sampling timestep and uses a noise-level-aware schedule to gradually de-weight the guidance as error increases, balancing adherence to guidance with high-quality synthesis.
- Unlike prior training-free approaches, the method does not require knowing a forward (fine-to-coarse) transformation operator, broadening applicability to image and video generation tasks.
- Extensive experiments demonstrate the method's effectiveness and generalization across diverse visual generation tasks, validating its robustness and practicality.
Related Articles

The programming passion is melting
Dev.to

Maximize Developer Revenue with Monetzly's Innovative API for AI Conversations
Dev.to
Co-Activation Pattern Detection for Prompt Injection: A Mechanistic Interpretability Approach Using Sparse Autoencoders
Reddit r/LocalLLaMA

How to Train Custom Language Models: Fine-Tuning vs Training From Scratch (2026)
Dev.to

KoboldCpp 1.110 - 3 YR Anniversary Edition, native music gen, qwen3tts voice cloning and more
Reddit r/LocalLLaMA