SciHorizon-DataEVA: An Agentic System for AI-Readiness Evaluation of Heterogeneous Scientific Data

arXiv cs.AI / 4/30/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes SciHorizon-DataEVA, an agentic system designed to evaluate the AI-readiness of heterogeneous scientific datasets at scale, addressing the lack of systematic assessment methods for AI-for-Science.
  • It introduces the Sci-TQA2 framework that structures AI-readiness into four measurable dimensions: Governance Trustworthiness, Data Quality, AI Compatibility, and Scientific Adaptability.
  • The system operationalizes Sci-TQA2 via Sci-TQA2-Eval, a hierarchical multi-agent approach using a directed cyclic workflow to iteratively generate and run dataset-aware evaluation plans.
  • It dynamically builds evaluation specifications by combining dataset profiling, applicability-aware metric selection, and knowledge-augmented planning based on domain constraints and dataset-to-paper signals.
  • Experiments across multiple scientific domains show that SciHorizon-DataEVA enables scalable, reliable, and generalizable AI-readiness evaluation.
  • It includes adaptive, tool-centric execution with built-in verification and self-correction to improve reliability of the evaluation outcomes.

Abstract

AI-for-Science (AI4Science) is increasingly transforming scientific discovery by embedding machine learning models into prediction, simulation, and hypothesis generation workflows across domains. However, the effectiveness of these models is fundamentally constrained by the AI-readiness of scientific data, for which no scalable and systematic evaluation mechanism currently exists. In this work, we propose SciHorizon-DataEVA, a novel agentic system to scalable AI-readiness evaluation of heterogeneous scientific data. At the evaluation-criteria level, we introduce the Sci-TQA2 principles, which organize AI-readiness into four complementary dimensions: Governance Trustworthiness, Data Quality, AI Compatibility, and Scientific Adaptability. Each dimension is decomposed into measurable atomic elements that enable fine-grained and executable assessment. To operationalize these principles at scale, we develop Sci-TQA2-Eval, a hierarchical multi-agent evaluation approach orchestrated through a directed, cyclic workflow. Our Sci-TQA2-Eval dynamically constructs dataset-aware evaluation specifications by combining lightweight dataset profiling, applicability-aware metric activation, and knowledge-augmented planning grounded in domain constraints and dataset-paper signals. These specifications are executed through an adaptive, tool-centric evaluation mechanism with built-in verification and self-correction, enabling scalable and reliable assessment across heterogeneous scientific data. Extensive experiments on scientific datasets spanning multiple domains demonstrate the effectiveness and generality of SciHorizon-DataEVA for principled AI-readiness evaluation.