MetaGAI: A Large-Scale and High-Quality Benchmark for Generative AI Model and Data Card Generation

arXiv cs.AI / 4/28/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • MetaGAI is a newly proposed, large-scale benchmark for evaluating automated Model Card and Data Card generation for generative AI, addressing limitations of manual documentation and prior automated methods.
  • The benchmark includes 2,541 verified document triplets built via semantic triangulation across academic papers, GitHub repositories, and Hugging Face artifacts, improving data coverage and fidelity.
  • MetaGAI uses a multi-agent pipeline (Retriever, Generator, and Editor) and validates outputs with a human-in-the-loop workflow, including human review of editor-refined ground truth.
  • The authors provide an evaluation protocol that combines automated metrics with an LLM-as-a-Judge approach, and they find that sparse Mixture-of-Experts models can offer better cost-quality efficiency, alongside a faithfulness–completeness trade-off.
  • Data and code are released publicly as a foundation for benchmarking, training, and analyzing scalable automated Model/Data Card generation systems.

Abstract

The rapid proliferation of Generative AI necessitates rigorous documentation standards for transparency and governance. However, manual creation of Model and Data Cards is not scalable, while automated approaches lack large-scale, high-fidelity benchmarks for systematic evaluation. We introduce MetaGAI, a comprehensive benchmark comprising 2,541 verified document triplets constructed through semantic triangulation of academic papers, GitHub repositories, and Hugging Face artifacts. Unlike prior single-source datasets, MetaGAI employs a multi-agent framework with specialized Retriever, Generator, and Editor agents, validated through four-dimensional human-in-the-loop assessment, including human evaluation of editor-refined ground truth. We establish a robust evaluation protocol combining automated metrics with validated LLM-as-a-Judge frameworks. Extensive analysis reveals that sparse Mixture-of-Experts architectures achieve superior cost-quality efficiency, while a fundamental trade-off exists between faithfulness and completeness. MetaGAI provides a foundational testbed for benchmarking, training, and analyzing automated Model and Data Card generation methods at scale. Our data and code are available at: https://github.com/haoxuan-unt2024/MetaGAI-Benchmark.