Knows: Agent-Native Structured Research Representations

arXiv cs.AI / 4/21/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • The paper introduces “Knows,” a lightweight companion format that attaches structured, verifiable research data (claims, evidence, provenance, and relations) to existing papers for easier consumption by LLM agents.
  • Knows uses a thin YAML sidecar (KnowsRecord) alongside the original PDF, requiring no changes to the publication and enforcing correctness via a deterministic schema linter.
  • In evaluations on 140 comprehension questions across 20 papers and 14 disciplines, using the sidecar substantially improves accuracy for smaller “weak” LLM agents (0.8B–2B parameters), while also reducing input token usage.
  • Re-scoring by an LLM-as-a-judge suggests that weak-model performance with the sidecar can closely approach strong-model PDF-only accuracy.
  • The project reports early adoption signals through a community sidecar hub that has indexed over ten thousand publications, supporting claims of scalability and readiness for real-world use.

Abstract

Research artifacts are distributed primarily as reader-oriented documents like PDFs. This creates a bottleneck for increasingly agent-assisted and agent-native research workflows, in which LLM agents need to infer fine-grained, task-relevant information from lengthy full documents, a process that is expensive, repetitive, and unstable at scale. We introduce Knows, a lightweight companion specification that binds structured claims, evidence, provenance, and verifiable relations to existing research artifacts in a form LLM agents can consume directly. Knows addresses the gap with a thin YAML sidecar (KnowsRecord) that coexists with the original PDF, requiring no changes to the publication itself, and validated by a deterministic schema linter. We evaluate Knows on 140 comprehension questions across 20 papers spanning 14 academic disciplines, comparing PDF-only, sidecar-only, and hybrid conditions across six LLM agents of varying capacity. Weak models (0.8B--2B parameters) improve from 19--25\% to 47--67\% accuracy (+29 to +42 percentage points) when reading sidecar instead of PDF, while consuming 29--86\% fewer input tokens; an LLM-as-judge re-scoring confirms that weak-model sidecar accuracy (75--77\%) approaches stronger-model PDF accuracy (78--83\%). Beyond this controlled evaluation, a community sidecar hub at https://knows.academy/ has already indexed over ten thousand publications and continues to grow daily, providing independent evidence that the format is adoption-ready at scale.