Local Inconsistency Resolution: The Interplay between Attention and Control in Probabilistic Models

arXiv cs.AI / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Local Inconsistency Resolution (LIR), a generic algorithm for learning and approximate inference that works by repeatedly focusing on part of a probabilistic model and fixing inconsistencies using parameters under control.
  • LIR is built on Probabilistic Dependency Graphs (PDGs), which can represent inconsistent beliefs and provide a flexible foundation for the proposed approach.
  • The authors show that LIR unifies and generalizes several well-known methods, including EM, belief propagation, adversarial training, GANs, and GFlowNets, with each recoverable as a specific LIR instance.
  • For GFlowNets, LIR yields a more natural loss function that the authors report improves GFlowNet convergence.
  • The paper includes an implementation for discrete PDGs and evaluates behavior on synthetic graphs, comparing results to global optimization on the full PDG.

Abstract

We present a generic algorithm for learning and approximate inference with an intuitive epistemic interpretation: iteratively focus on a subset of the model and resolve inconsistencies using the parameters under control. This framework, which we call Local Inconsistency Resolution (LIR) is built upon Probabilistic Dependency Graphs (PDGs), which provide a flexible representational foundation capable of capturing inconsistent beliefs. We show how LIR unifies and generalizes a wide variety of important algorithms in the literature, including the Expectation-Maximization (EM) algorithm, belief propagation, adversarial training, GANs, and GFlowNets. In the last case, LIR actually suggests a more natural loss, which we demonstrate improves GFlowNet convergence. Each method can be recovered as a specific instance of LIR by choosing a procedure to direct focus (attention and control). We implement this algorithm for discrete PDGs and study its properties on synthetically generated PDGs, comparing its behavior to the global optimization semantics of the full PDG.