Post-Training Augmentation Invariance
arXiv stat.ML / 4/24/2026
💬 OpinionTools & Practical UsageModels & Research
Key Points
- The paper proposes a framework called post-training augmentation invariance that aims to add invariance to a pretrained network while keeping its behavior on the original (non-augmented) input distribution unchanged.
- It introduces “augmented encoders” (probabilistic encoders that formalize augmentation-based encoding) as the core mechanism, and defines two training objectives: Markov-Wasserstein minimization and Wasserstein correlation maximization.
- Experiments show that lightweight one-hidden-layer MLP adapter networks trained with these losses can be appended to a frozen pretrained feature extractor to achieve approximate augmentation invariance.
- On STL10 using DINOv2 features, adding a trained adapter improves rotated-image classification accuracy to 94% versus 71% without the adapter, and noise-invariant performance rises from 58% to 86%.
- The authors report the approach requires no fine-tuning of the pretrained backbone (frozen weights) and minimally perturbs original latent features, while adapters trained with alternative losses (SimCLR, HSIC maximization) fail and can corrupt the latent space.
Related Articles

Black Hat USA
AI Business

The 67th Attempt: When Your "Knowledge Management" System Becomes a Self-Fulfilling Prophecy of Excellence
Dev.to

Context Engineering for Developers: A Practical Guide (2026)
Dev.to

GPT-5.5 is here. So is DeepSeek V4. And honestly, I am tired of version numbers.
Dev.to
AI Visibility Tracking Exploded in 2026: 6 Tools Every Brand Needs Now
Dev.to