Implicit Statistical Inference in Transformers: Approximating Likelihood-Ratio Tests In-Context
arXiv cs.LG / 3/12/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper treats in-context learning as a binary hypothesis-testing problem and shows the optimal policy corresponds to a likelihood-ratio test, providing a rigorous mechanistic interpretability setting where the ground truth is known.
- By training transformers on tasks with linear versus nonlinear geometry, the authors show models approximate Bayes-optimal sufficient statistics from context up to a monotone transformation, matching an ideal oracle estimator in nonlinear regimes.
- Mechanistic analysis using a logit lens and circuit alignment indicates the model does not rely on fixed kernel smoothing but instead adjusts the decision boundary, displaying a voting-style ensemble for linear tasks and deeper sequential computation for nonlinear tasks.
- The work argues that implicit in-context learning arises from task-adaptive statistical estimators rather than simple similarity matching, offering a theoretical framework for understanding ICL.




