CREval: An Automated Interpretable Evaluation for Creative Image Manipulation under Complex Instructions

arXiv cs.CV / 3/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces CREval, an automated, QA-based evaluation pipeline aimed at making multimodal image-manipulation model scoring more complete and interpretable than opaque MLLM-based metrics.
  • It also releases CREval-Bench, a benchmark for creative image editing under complex instructions, spanning three categories and nine creative dimensions with 800+ editing samples and 13K evaluation queries.
  • Using CREval and CREval-Bench, the authors evaluate a range of state-of-the-art open- and closed-source models and find closed-source models generally perform better on complex/creative edits.
  • Despite performance gaps, the study reports that all evaluated models still struggle to carry out such complex creative edits effectively.
  • User studies show high consistency between CREval’s automated metrics and human judgments, positioning CREval as a practical foundation for future evaluation and research.

Abstract

Instruction-based multimodal image manipulation has recently made rapid progress. However, existing evaluation methods lack a systematic and human-aligned framework for assessing model performance on complex and creative editing tasks. To address this gap, we propose CREval, a fully automated question-answer (QA)-based evaluation pipeline that overcomes the incompleteness and poor interpretability of opaque Multimodal Large Language Models (MLLMs) scoring. Simultaneously, we introduce CREval-Bench, a comprehensive benchmark specifically designed for creative image manipulation under complex instructions. CREval-Bench covers three categories and nine creative dimensions, comprising over 800 editing samples and 13K evaluation queries. Leveraging this pipeline and benchmark, we systematically evaluate a diverse set of state-of-the-art open and closed-source models. The results reveal that while closed-source models generally outperform open-source ones on complex and creative tasks, all models still struggle to complete such edits effectively. In addition, user studies demonstrate strong consistency between CREval's automated metrics and human judgments. Therefore, CREval provides a reliable foundation for evaluating image editing models on complex and creative image manipulation tasks, and highlights key challenges and opportunities for future research.