From Weak Cues to Real Identities: Evaluating Inference-Driven De-Anonymization in LLM Agents
arXiv cs.AI / 3/20/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces 'inference-driven linkage' as a privacy risk where LLM-based agents reconstruct real-world identities from sparse cues and public information without bespoke engineering.
- It evaluates this threat across three settings—classical linkage (Netflix and AOL), InferLink benchmarks, and modern text-rich artifacts—and finds agents can perform both fixed-pool matching and open-ended identity resolution without task-specific heuristics.
- In the Netflix Prize setting, an agent reconstructs 79.2% of identities, significantly higher than a 56.0% classical baseline.
- The linkage can arise even without adversarial prompts and as a byproduct of benign cross-source analysis and unstructured narratives.
- The study argues that identity inference must be treated as a first-class privacy risk and that evaluations should measure what identities an agent can infer.
Related Articles

Attacks On Data Centers, Qwen3.5 In All Sizes, DeepSeek’s Huawei Play, Apple’s Multimodal Tokenizer
The Batch

Your AI generated code is "almost right", and that is actually WORSE than it being "wrong".
Dev.to

Lessons from Academic Plagiarism Tools for SaaS Product Development
Dev.to

**Core Allocation Optimization for Energy‑Efficient Multi‑Core Scheduling in ARINC650 Systems**
Dev.to

KI in der amtlichen Recherche beim DPMA: Was Patentanwälte bei Neuanmeldungen jetzt beachten sollten (Stand: März 2026)
Dev.to