Beyond Logit Adjustment: A Residual Decomposition Framework for Long-Tailed Reranking

arXiv cs.LG / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that fixed post-hoc logit adjustments are insufficient for long-tailed settings because the optimal correction to rerank classes can vary across inputs rather than being a constant offset per class.
  • It formulates Bayes-optimal reranking on top-k base-model candidates and shows the required residual correction decomposes into a classwise term (constant within a class) and a pairwise term that depends on the input and competing labels.
  • The authors derive conditions under which a fixed offset can recover Bayes-optimal ordering (when residuals are purely classwise) and conditions where it cannot (when the same label pair implies conflicting ordering constraints across contexts).
  • Based on the decomposition, the paper introduces REPAIR, a lightweight post-hoc reranker that combines shrinkage-stabilized classwise correction with a linear, competition-feature-driven pairwise component.
  • Experiments across five benchmarks (covering image classification, species/scene recognition, and rare disease diagnosis) support the framework by showing when pairwise correction improves performance versus when classwise correction is enough.

Abstract

Long-tailed classification, where a small number of frequent classes dominate many rare ones, remains challenging because models systematically favor frequent classes at inference time. Existing post-hoc methods such as logit adjustment address this by adding a fixed classwise offset to the base-model logits. However, the correction required to restore the relative ranking of two classes need not be constant across inputs, and a fixed offset cannot adapt to such variation. We study this problem through Bayes-optimal reranking on a base-model top-k shortlist. The gap between the optimal score and the base score, the residual correction, decomposes into a classwise component that is constant within each class, and a pairwise component that depends on the input and competing labels. When the residual is purely classwise, a fixed offset suffices to recover the Bayes-optimal ordering. We further show that when the same label pair induces incompatible ordering constraints across contexts, no fixed offset can achieve this recovery. This decomposition leads to testable predictions regarding when pairwise correction can improve performance and when cannot. We develop REPAIR (Reranking via Pairwise residual correction), a lightweight post-hoc reranker that combines a shrinkage-stabilized classwise term with a linear pairwise term driven by competition features on the shortlist. Experiments on five benchmarks spanning image classification, species recognition, scene recognition, and rare disease diagnosis confirm that the decomposition explains where pairwise correction helps and where classwise correction alone suffices.