LLMs can construct powerful representations and streamline sample-efficient supervised learning
arXiv cs.AI / 3/13/2026
📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- LLMs analyze a small diverse subset of input examples in-context to synthesize a global rubric that acts as a programmatic specification for extracting and organizing evidence.
- This rubric is then used to transform naive text-serializations of inputs into a standardized format for downstream models, improving representation quality and sample efficiency.
- Local rubrics provide task-conditioned summaries generated by the LLM to tailor representations for each specific task.
- Across 15 clinical tasks from the EHRSHOT benchmark, rubric-based approaches significantly outperform traditional count-feature models, naive text-serialization-based LLM baselines, and a clinical foundation model pretrained on far more data.
- Rubrics offer auditable, cost-effective deployment at scale and can be converted to tabular representations that enable a broader set of machine learning techniques.
Related Articles

Manus、AIエージェントをデスクトップ化 ローカルPC上でファイルやアプリを直接操作可能にのサムネイル画像
Ledge.ai

The programming passion is melting
Dev.to

Best AI Tools for Property Managers in 2026
Dev.to

Building “The Sentinel” – AI Parametric Insurance at Guidewire DEVTrails
Dev.to

Maximize Developer Revenue with Monetzly's Innovative API for AI Conversations
Dev.to