Efficient Deconvolution in Populational Inverse Problems

arXiv stat.ML / 5/6/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses distributional inversion problems where one must infer the parameter distribution underlying a physical process from multiple noisy observation sets.
  • It tackles the key obstacle of blind deconvolution when the observational noise distribution is unknown, arguing that population-level data from repeated physical instantiations can make deconvolution feasible.
  • The authors introduce a coupled optimization framework that simultaneously estimates the parameter distribution and the unknown noise distribution by minimizing a loss that compares observed data to outputs of a parameter-dependent physical model.
  • They develop a modified gradient-descent method that exploits structure in the noise model, and they add an active-learning strategy to train a surrogate model focused on parameter regions of interest.
  • The method is evaluated on several examples—including porous medium flow, damped elastodynamics, and simplified atmospheric dynamics—showing its ability to accelerate computation and support automatic differentiation even for black-box (possibly nondifferentiable) solvers.

Abstract

This work is focussed on the inversion task of inferring the distribution over parameters of interest leading to multiple sets of observations. The potential to solve such distributional inversion problems is driven by increasing availability of data, but a major roadblock is blind deconvolution, arising when the observational noise distribution is unknown. However, when data originates from collections of physical systems, a population, it is possible to leverage this information to perform deconvolution. To this end, we propose a methodology leveraging large data sets of observations, collected from different instantiations of the same physical processes, to simultaneously deconvolve the data corrupting noise distribution, and to identify the distribution over model parameters defining the physical processes. A parameter-dependent mathematical model of the physical process is employed. A loss function characterizing the match between the observed data and the output of the mathematical model is defined; it is minimized as a function of the both the parameter inputs to the model of the physics and the parameterized observational noise. This coupled problem is addressed with a modified gradient descent algorithm that leverages specific structure in the noise model. Furthermore, a new active learning scheme is proposed, based on adaptive empirical measures, to train a surrogate model to be accurate in parameter regions of interest; this approach accelerates computation and enables automatic differentiation of black-box, potentially nondifferentiable, code computing parameter-to-solution maps. The proposed methodology is demonstrated on porous medium flow, damped elastodynamics, and simplified models of atmospheric dynamics.