Compliance-by-Construction Argument Graphs: Using Generative AI to Produce Evidence-Linked Formal Arguments for Certification-Grade Accountability

arXiv cs.AI / 4/7/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses certification-grade accountability needs for high-stakes decision systems by combining formal, evidence-linked argument structures with generative AI workflows.
  • It proposes a “compliance-by-construction” architecture where each AI-assisted claim is only added to the decision record after retrieval-grounding and strict validation against explicit reasoning constraints.
  • The approach uses an argument-graph representation (inspired by assurance cases), retrieval-augmented generation for evidence-grounded drafting, and a reasoning/validation kernel enforcing completeness and admissibility.
  • To enable auditability, it adds a provenance ledger aligned with the W3C PROV standard so that justification steps can be traced and reviewed.
  • The authors outline a system design and evaluation strategy using enforceable invariants and suggest deterministic validation can block unsupported (hallucinated) claims while speeding up argument construction.

Abstract

High-stakes decision systems increasingly require structured justification, traceability, and auditability to ensure accountability and regulatory compliance. Formal arguments commonly used in the certification of safety-critical systems provide a mechanism for structuring claims, reasoning, and evidence in a verifiable manner. At the same time, generative artificial intelligence systems are increasingly integrated into decision-support workflows, assisting with drafting explanations, summarizing evidence, and generating recommendations. However, current deployments often rely on language models as loosely constrained assistants, which introduces risks such as hallucinated reasoning, unsupported claims, and weak traceability. This paper proposes a compliance-by-construction architecture that integrates Generative AI (GenAI) with structured formal argument representations. The approach treats each AI-assisted step as a claim that must be supported by verifiable evidence and validated against explicit reasoning constraints before it becomes part of an official decision record. The architecture combines four components: i) a typed Argument Graph representation inspired by assurance-case methods, ii) retrieval-augmented generation (RAG) to draft argument fragments grounded in authoritative evidence, iii) a reasoning and validation kernel enforcing completeness and admissibility constraints, and iv) a provenance ledger aligned with the W3C PROV standard to support auditability. We present a system design and an evaluation strategy based on enforceable invariants and worked examples. The analysis suggests that deterministic validation rules can prevent unsupported claims from entering the decision record while allowing GenAI to accelerate argument construction.