Analytica: Soft Propositional Reasoning for Robust and Scalable LLM-Driven Analysis

arXiv cs.AI / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Analytica, an LLM agent architecture designed to make real-world analysis more stable and verifiable by using Soft Propositional Reasoning (SPR) to estimate soft truth values of outcome propositions.
  • Analytica reduces reasoning error by combining a bias-reduction phase that decomposes problems into subpropositions and uses tool-equipped “grounder” agents (including a Jupyter Notebook agent) to validate and score facts.
  • It also reduces variance via a parallel divide-and-conquer synthesis step that combines grounded outputs using robust linear models to average out stochastic noise efficiently and at scale.
  • Experiments on economic, financial, and political forecasting report an average 15.84% accuracy improvement over diverse base models, reaching 71.06% accuracy with the lowest variance (6.02%) using a Deep Research grounder.
  • The proposed Jupyter Notebook grounder is claimed to be cost-effective, achieving about 70.11% accuracy with 90.35% less cost and 52.85% less time, while maintaining near-linear growth in runtime as analysis depth increases.

Abstract

Large language model (LLM) agents are increasingly tasked with complex real-world analysis (e.g., in financial forecasting, scientific discovery), yet their reasoning suffers from stochastic instability and lacks a verifiable, compositional structure. To address this, we introduce Analytica, a novel agent architecture built on the principle of Soft Propositional Reasoning (SPR). SPR reframes complex analysis as a structured process of estimating the soft truth values of different outcome propositions, allowing us to formally model and minimize the estimation error in terms of its bias and variance. Analytica operationalizes this through a parallel, divide-and-conquer framework that systematically reduces both sources of error. To reduce bias, problems are first decomposed into a tree of subpropositions, and tool-equipped LLM grounder agents are employed, including a novel Jupyter Notebook agent for data-driven analysis, that help to validate and score facts. To reduce variance, Analytica recursively synthesizes these grounded leaves using robust linear models that average out stochastic noise with superior efficiency, scalability, and enable interactive "what-if" scenario analysis. Our theoretical and empirical results on economic, financial, and political forecasting tasks show that Analytica improves 15.84% accuracy on average over diverse base models, achieving 71.06% accuracy with the lowest variance of 6.02% when working with a Deep Research grounder. Our Jupyter Notebook grounder shows strong cost-effectiveness that achieves a close 70.11% accuracy with 90.35% less cost and 52.85% less time. Analytica also exhibits highly noise-resilient and stable performance growth as the analysis depth increases, with a near-linear time complexity, as well as good adaptivity to open-weight LLMs and scientific domains.