The Effect of Idea Elaboration on the Automatic Assessment of Idea Originality
arXiv cs.AI / 4/25/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The study examines whether large language models (LLMs) mirror human judgments when automatically assessing how original ideas are in a divergent thinking task (Alternate Uses Task).
- It compares human-rated originality from trained student raters against machine ratings from two fine-tuned specialized systems and a ChatGPT-4o setup using the same prompt instructions.
- Results show a self-preference bias: LLM-based automatic assessors tend to favor outputs that resemble their own style rather than human creativity.
- Crucially, the self-preference bias disappears when analyses control for the degree of idea elaboration, suggesting elaboration can explain the bias.
- The paper outlines theoretical and methodological implications for future research on automated creativity assessment systems.
Related Articles
Navigating WooCommerce AI Integrations: Lessons for Agencies & Developers from a Bluehost Conflict
Dev.to

One Day in Shenzhen, Seen Through an AI's Eyes
Dev.to

Underwhelming or underrated? DeepSeek V4 shows “impressive” gains
SCMP Tech

Claude Code: Hooks, Subagents, and Skills — Complete Guide
Dev.to

Finding the Gold: An AI Framework for Highlight Detection
Dev.to