Bias Inheritance in Neural-Symbolic Discovery of Constitutive Closures Under Function-Class Mismatch

arXiv cs.LG / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies data-driven discovery of constitutive closures (diffusion and reaction laws) for nonlinear reaction–diffusion PDEs from spatiotemporal observations while avoiding misleading “low residual = correct physics” conclusions.
  • It proposes a three-stage neural-symbolic pipeline: learn noise-robust weak-form numerical surrogates under physical constraints, compress them into interpretable symbolic families (polynomial/rational/saturation), and validate by explicit forward re-simulation on unseen initial conditions.
  • Numerical experiments show that with matched function libraries, classical weak polynomial baselines can already be near-correct reference estimators and neural surrogates do not automatically outperform them.
  • With function-class mismatch, the neural surrogates add necessary flexibility and can be compressed into compact symbolic laws with minimal rollout degradation.
  • The authors identify “bias inheritance,” where symbolic compression fails to correct constitutive bias; the symbolic closure’s true error tracks the neural surrogate’s error, implying the main bottleneck is the initial inverse problem rather than the symbolic step.

Abstract

We investigate the data-driven discovery of constitutive closures in nonlinear reaction-diffusion systems with known governing PDE structures. Our objective is to robustly recover diffusion and reaction laws from spatiotemporal observations while avoiding the common pitfall where low residuals or short-horizon predictions are conflated with physical recovery. We propose a three-stage neural-symbolic framework: (1) learning numerical surrogates under physical constraints using a noise-robust weak-form-driven objective; (2) compressing these surrogates into restricted interpretable symbolic families (e.g., polynomial, rational, and saturation forms); and (3) validating the symbolic closures through explicit forward re-simulation on unseen initial conditions. Extensive numerical experiments reveal two distinct regimes. Under matched-library settings, weak polynomial baselines behave as correctly specified reference estimators, showing that neural surrogates do not uniformly outperform classical bases. Conversely, under function-class mismatch, neural surrogates provide necessary flexibility and can be compressed into compact symbolic laws with minimal rollout degradation. However, we identify a critical "bias inheritance" mechanism where symbolic compression does not automatically repair constitutive bias. Across various observation regimes, the true error of the symbolic closure closely tracks that of the neural surrogate, yielding a bias inheritance ratio near one. These findings demonstrate that the primary bottleneck in neural-symbolic modeling lies in the initial numerical inverse problem rather than the subsequent symbolic compression. We underscore that constitutive claims must be rigorously supported by forward validation rather than residual minimization alone.