IBCapsNet: Information Bottleneck Capsule Network for Noise-Robust Representation Learning

arXiv cs.CV / 3/24/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • IBCapsNet is a new capsule network architecture that applies the Information Bottleneck principle to improve robustness and efficiency over traditional Capsule Networks (CapsNets).
  • Instead of iterative dynamic routing, it uses a one-pass variational aggregation pipeline: primary capsules are compressed into a global context, then class-specific VAEs infer noise-robust latent capsules regularized by KL divergence.
  • Experiments on MNIST, Fashion-MNIST, SVHN, and CIFAR-10 show it matches CapsNet on clean data while substantially improving performance under multiple synthetic noise types.
  • The approach reports major efficiency gains, including faster training/inference (2.54x training, 3.64x throughput) and fewer parameters (4.66% reduction) versus CapsNet.
  • The paper positions IBCapsNet as a bridge between information-theoretic representation learning and interpretable capsule models, with accompanying code released on GitHub.

Abstract

Capsule networks (CapsNets) are superior at modeling hierarchical spatial relationships but suffer from two critical limitations: high computational cost due to iterative dynamic routing and poor robustness under input corruptions. To address these issues, we propose IBCapsNet, a novel capsule architecture grounded in the Information Bottleneck (IB) principle. Instead of iterative routing, IBCapsNet employs a one-pass variational aggregation mechanism, where primary capsules are first compressed into a global context representation and then processed by class-specific variational autoencoders (VAEs) to infer latent capsules regularized by the KL divergence. This design enables efficient inference while inherently filtering out noise. Experiments on MNIST, Fashion-MNIST, SVHN and CIFAR-10 show that IBCapsNet matches CapsNet in clean-data accuracy (achieving 99.41% on MNIST and 92.01% on SVHN), yet significantly outperforms it under four types of synthetic noise - demonstrating average improvements of +17.10% and +14.54% for clamped additive and multiplicative noise, respectively. Moreover, IBCapsNet achieves 2.54x faster training and 3.64x higher inference throughput compared to CapsNet, while reducing model parameters by 4.66%. Our work bridges information-theoretic representation learning with capsule networks, offering a principled path toward robust, efficient, and interpretable deep models. Code is available at https://github.com/cxiang26/IBCapsnet