TinyNeRV: Compact Neural Video Representations via Capacity Scaling, Distillation, and Low-Precision Inference
arXiv cs.CV / 4/13/2026
💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- The paper introduces TinyNeRV, a systematic study of very compact Neural Representations for Videos (NeRV) aimed at resource-constrained and real-time deployment.
- It proposes two lightweight variants, NeRV-T and NeRV-T+, and evaluates how aggressive capacity reduction impacts reconstruction quality, computation, and decoding throughput across multiple video datasets.
- To improve fidelity without raising inference cost, the authors explore knowledge distillation using frequency-aware focal supervision for low-capacity models.
- The study also assesses robustness under low-precision inference via both post-training quantization and quantization-aware training.
- Results show that well-designed tiny NeRV architectures can substantially cut parameter count, compute cost, and memory while maintaining favorable quality-efficiency trade-offs, with an official implementation released on GitHub.




