AsyncMDE: Real-Time Monocular Depth Estimation via Asynchronous Spatial Memory
arXiv cs.CV / 3/12/2026
💬 OpinionModels & Research
Key Points
- AsyncMDE introduces an asynchronous depth perception system that splits work between a foundation model producing spatial features in the background and a lightweight foreground model that fuses memory with current observations to estimate depth.
- The system enables cross-frame feature reuse with complementary fusion and autoregressive memory updates, achieving bounded accuracy degradation across frames.
- It is compact (3.83M parameters) and delivers 237 FPS on an RTX 4090, recovering 77% of the accuracy gap to the foundation model with 25x fewer parameters; it also runs at 161 FPS on a Jetson AGX Orin with TensorRT, demonstrating edge feasibility.
- Validation on indoor static, dynamic, and synthetic extreme-motion benchmarks shows graceful degradation between refreshes and practical real-time performance.
Related Articles
Two bots, one confused server: what Nimbus revealed about AI agent identity
Dev.to
PIXIU: A Large Language Model, Instruction Data and Evaluation Benchmark forFinance
Dev.to
A Coding Implementation to Build an Uncertainty-Aware LLM System with Confidence Estimation, Self-Evaluation, and Automatic Web Research
MarkTechPost
DNA Memory: Making AI Agents Learn, Forget, and Evolve Like a Human Brain
Dev.to
Tinybox- offline AI device 120B parameters
Hacker News