AHC: Meta-Learned Adaptive Compression for Continual Object Detection on Memory-Constrained Microcontrollers

arXiv cs.AI / 4/14/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes Adaptive Hierarchical Compression (AHC), a meta-learning framework for continual object detection on memory-constrained microcontrollers with <100KB RAM by making compression adapt per task rather than using fixed strategies.
  • AHC uses a true MAML-style inner-loop adaptation in only 5 gradient steps, along with hierarchical multi-scale compression that applies scale-aware compression ratios aligned to FPN redundancy patterns.
  • It introduces a dual-memory design with short-term and long-term feature banks plus importance-based consolidation, enforcing a hard 100KB budget to reduce catastrophic forgetting.
  • The authors provide theoretical bounds on catastrophic forgetting that depend on compression error, number of tasks, and memory size, and they validate performance on CORe50, TiROD, and PASCAL VOC.
  • Experiments show that AHC achieves competitive continual detection accuracy within a 100KB replay setting when combined with mean-pooled compressed feature replay, EWC regularization, and feature distillation.

Abstract

Deploying continual object detection on microcontrollers (MCUs) with under 100KB memory requires efficient feature compression that can adapt to evolving task distributions. Existing approaches rely on fixed compression strategies (e.g., FiLM conditioning) that cannot adapt to heterogeneous task characteristics, leading to suboptimal memory utilization and catastrophic forgetting. We introduce Adaptive Hierarchical Compression (AHC), a meta-learning framework featuring three key innovations: (1) true MAML-based compression that adapts via gradient descent to each new task in just 5 inner-loop steps, (2) hierarchical multi-scale compression with scale-aware ratios (8:1 for P3, 6.4:1 for P4, 4:1 for P5) matching FPN redundancy patterns, and (3) a dual-memory architecture combining short-term and long-term banks with importance-based consolidation under a hard 100KB budget. We provide formal theoretical guarantees bounding catastrophic forgetting as O({\epsilon}{sq.root(T)} + 1/{sq.root(M)}) where {\epsilon} is compression error, T is task count, and M is memory size. Experiments on CORe50, TiROD, and PASCAL VOC benchmarks with three standard baselines (Fine-tuning,EWC, iCaRL) demonstrate that AHC enables practical continual detection within a 100KB replay budget, achieving competitive accuracy through mean-pooled compressed feature replay combined with EWC regularization and feature distillation.