AI Navigate

Building Benchmarks from the Ground Up: Community-Centered Evaluation of LLMs in Healthcare Chatbot Settings

arXiv cs.CL / 3/16/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that current LLM benchmarks in healthcare often miss real-world user contexts and cultural practices, underscoring the need for contextually grounded evaluation.
  • It introduces Samiksha, a community-driven evaluation pipeline co-created with civil-society organizations and community members to guide what to evaluate, how to build benchmarks, and how outputs are scored.
  • The approach enables scalable benchmarking through automation while incorporating cultural awareness and community feedback into the evaluation process.
  • The authors demonstrate the pipeline in India's health domain, showing how multilingual LLMs handle nuanced community health queries and outlining a scalable path for inclusive LLM evaluation.

Abstract

Large Language Models (LLMs) are typically evaluated through general or domain-specific benchmarks testing capabilities that often lack grounding in the lived realities of end users. Critical domains such as healthcare require evaluations that extend beyond artificial or simulated tasks to reflect the everyday needs, cultural practices, and nuanced contexts of communities. We propose Samiksha, a community-driven evaluation pipeline co-created with civil-society organizations (CSOs) and community members. Our approach enables scalable, automated benchmarking through a culturally aware, community-driven pipeline in which community feedback informs what to evaluate, how the benchmark is built, and how outputs are scored. We demonstrate this approach in the health domain in India. Our analysis highlights how current multilingual LLMs address nuanced community health queries, while also offering a scalable pathway for contextually grounded and inclusive LLM evaluation.