When W4A4 Breaks Camouflaged Object Detection: Token-Group Dual-Constraint Activation Quantization
arXiv cs.CV / 4/21/2026
📰 NewsDeveloper Stack & InfrastructureModels & Research
Key Points
- The paper studies post-training W4A4 (4-bit weights/4-bit activations) quantization for camouflaged object detection (COD) using Transformer-based models, showing a sharp “quantization cliff” that makes aggressive low-bit inference unusually difficult for COD.
- It identifies the cause as token-local activation bottlenecks where heavy-tailed background tokens dominate a shared activation range, increasing quantization step size and causing weak but meaningful boundary cues to be mapped into the zero bin.
- To fix this, the authors propose COD-TDQ, a COD-aware Token-group Dual-constraint activation Quantization method using Direct-Sum Token-Group (DSTG) token-group scaling and Dual-Constraint Range Projection (DCRP) to bound both the step-to-dispersion ratio and the zero-bin mass.
- Experiments on four COD benchmarks with two baseline models (CFRN and ESCNet) show COD-TDQ improves the Sα-score by more than 0.12 over the state of the art quantization approach without retraining, and the code will be released.
Related Articles

A practical guide to getting comfortable with AI coding tools
Dev.to

We built it during the NVIDIA DGX Spark Full-Stack AI Hackathon — and it ended up winning 1st place overall 🏆
Dev.to

Stop Losing Progress: Setting Up a Pro Jupyter Workflow in VS Code (No More Colab Timeouts!)
Dev.to

🚀 Major BrowserAct CLI Update
Dev.to

Building AgentOS: Why I’m Building the AWS Lambda for Insurance Claims
Dev.to