Learning from Equivalence Queries, Revisited

arXiv cs.LG / 4/7/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper revisits Angluin’s (1988) learning-from-equivalence-queries model to better match real ML lifecycles like deployment and iterative updates driven by user feedback.
  • It argues that the standard fully adversarial counterexample assumption can make the learning model overly pessimistic, and therefore proposes a broader, less adversarial class of counterexample generators called “symmetric.”
  • In the symmetric setting, counterexamples depend only on the symmetric difference between the learner’s hypothesis and the target, capturing natural mechanisms such as random counterexamples and “simplest” counterexamples by complexity.
  • The authors analyze learning under both full-information feedback (seeing correct labels) and bandit-style feedback (less information), deriving tight bounds on the number of learning rounds required.
  • The technical approach blends a game-theoretic analysis of symmetric adversaries with adaptive weighting methods and minimax arguments, and outlines directions for further research.

Abstract

Modern machine learning systems, such as generative models and recommendation systems, often evolve through a cycle of deployment, user interaction, and periodic model updates. This differs from standard supervised learning frameworks, which focus on loss or regret minimization over a fixed sequence of prediction tasks. Motivated by this setting, we revisit the classical model of learning from equivalence queries, introduced by Angluin (1988). In this model, a learner repeatedly proposes hypotheses and, when a deployed hypothesis is inadequate, receives a counterexample. Under fully adversarial counterexample generation, however, the model can be overly pessimistic. In addition, most prior work assumes a \emph{full-information} setting, where the learner also observes the correct label of the counterexample, an assumption that is not always natural. We address these issues by restricting the environment to a broad class of less adversarial counterexample generators, which we call \emph{symmetric}. Informally, such generators choose counterexamples based only on the symmetric difference between the hypothesis and the target. This class captures natural mechanisms such as random counterexamples (Angluin and Dohrn, 2017; Bhatia, 2021; Chase, Freitag, and Reyzin, 2024), as well as generators that return the simplest counterexample according to a prescribed complexity measure. Within this framework, we study learning from equivalence queries under both full-information and bandit feedback. We obtain tight bounds on the number of learning rounds in both settings and highlight directions for future work. Our analysis combines a game-theoretic view of symmetric adversaries with adaptive weighting methods and minimax arguments.