Beyond Local vs. External: A Game-Theoretic Framework for Trustworthy Knowledge Acquisition

arXiv cs.CL / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper highlights a core trade-off in cloud LLM use: external querying can improve reasoning and knowledge quality but may expose sensitive user intent, while local-only models protect privacy at the cost of quality.
  • It proposes GTKA (Game-theoretic Trustworthy Knowledge Acquisition), a framework that models knowledge utility vs. privacy as a strategic game.
  • GTKA uses three components: a privacy-aware sub-query generator that splits intent into low-risk fragments, an adversarial reconstruction attacker that estimates how much original intent can be recovered, and a trusted local integrator that securely combines external answers.
  • By training the generator and attacker in an alternating adversarial process, GTKA learns a sub-query policy that increases answer accuracy while reducing the reconstructability of sensitive intent.
  • Experiments on biomedical and legal benchmarks show GTKA substantially lowers intent leakage versus prior methods while preserving high-quality, high-fidelity answers.

Abstract

Cloud-hosted Large Language Models (LLMs) offer unmatched reasoning capabilities and dynamic knowledge, yet submitting raw queries to these external services risks exposing sensitive user intent. Conversely, relying exclusively on trusted local models preserves privacy but often compromises answer quality due to limited parameter scale and knowledge. To resolve this dilemma, we propose Game-theoretic Trustworthy Knowledge Acquisition (GTKA), a framework that formulates the trade-off between knowledge utility and privacy as a strategic game. GTKA consists of three components: (i) a privacy-aware sub-query generator that decomposes sensitive intent into generalized, low-risk fragments; (ii) an adversarial reconstruction attacker that attempts to infer the original query from these fragments, providing adaptive leakage signals; and (iii) a trusted local integrator that synthesizes external responses within a secure boundary. By training the generator and attacker in an alternating adversarial manner, GTKA optimizes the sub-query generation policy to maximize knowledge acquisition accuracy while minimizing the reconstructability of the original sensitive intent. To validate our approach, we construct two sensitive-domain benchmarks in the biomedical and legal fields. Extensive experiments demonstrate that GTKA significantly reduces intent leakage compared to state-of-the-art baselines while maintaining high-fidelity answer quality.