AI Navigate

A Prediction-as-Perception Framework for 3D Object Detection

arXiv cs.CV / 3/16/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces the Prediction-As-Perception (PAP) framework that fuses prediction and perception for 3D object perception tasks to enhance perceptual accuracy.
  • PAP uses two main modules, prediction and perception, that operate on continuous frame information, with the prediction module forecasting future positions and guiding the perception module in the next frame.
  • The predicted positions are used as queries for the next frame's perception, and the perceived results are iteratively fed back into the predictor, creating a biomimetic loop.
  • When evaluated with the UniAD model on the nuScenes dataset, PAP improves target tracking accuracy by about 10% and increases inference speed by about 15%, indicating efficiency gains.
  • The study claims such a design reduces computational resource consumption while enhancing accuracy, potentially benefiting autonomous driving perception systems.

Abstract

Humans combine prediction and perception to observe the world. When faced with rapidly moving birds or insects, we can only perceive them clearly by predicting their next position and focusing our gaze there. Inspired by this, this paper proposes the Prediction-As-Perception (PAP) framework, integrating a prediction-perception architecture into 3D object perception tasks to enhance the model's perceptual accuracy. The PAP framework consists of two main modules: prediction and perception, primarily utilizing continuous frame information as input. Firstly, the prediction module forecasts the potential future positions of ego vehicles and surrounding traffic participants based on the perception results of the current frame. These predicted positions are then passed as queries to the perception module of the subsequent frame. The perceived results are iteratively fed back into the prediction module. We evaluated the PAP structure using the end-to-end model UniAD on the nuScenes dataset. The results demonstrate that the PAP structure improves UniAD's target tracking accuracy by 10% and increases the inference speed by 15%. This indicates that such a biomimetic design significantly enhances the efficiency and accuracy of perception models while reducing computational resource consumption.