A Framework for Generating Semantically Ambiguous Images to Probe Human and Machine Perception

arXiv cs.CV / 3/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a psychophysically informed framework that generates semantically ambiguous images by interpolating between concepts in the CLIP embedding space to create continuous ambiguity spectra.
  • It uses these ambiguity probes to measure and compare where humans and machine vision classifiers place semantic boundaries between concepts (e.g., “duck” vs. “rabbit”).
  • The study finds systematic alignment differences: machine classifiers are more biased toward “rabbit,” while humans align more with the CLIP embedding used during image synthesis.
  • It reports that “guidance scale” affects human sensitivity to ambiguity more strongly than it affects machine classifiers, indicating distinct perception mechanisms under controlled conditions.
  • The framework is positioned as a diagnostic bridge between human psychophysics, classifier behavior/interpretability, and generative image synthesis for understanding alignment and robustness.

Abstract

The classic duck-rabbit illusion reveals that when visual evidence is ambiguous, the human brain must decide what it sees. But where exactly do human observers draw the line between ''duck'' and ''rabbit'', and do machine classifiers draw it in the same place? We use semantically ambiguous images as interpretability probes to expose how vision models represent the boundaries between concepts. We present a psychophysically-informed framework that interpolates between concepts in the CLIP embedding space to generate continuous spectra of ambiguous images, allowing us to precisely measure where and how humans and machine classifiers place their semantic boundaries. Using this framework, we show that machine classifiers are more biased towards seeing ''rabbit'', whereas humans are more aligned with the CLIP embedding used for synthesis, and the guidance scale seems to affect human sensitivity more strongly than machine classifiers. Our framework demonstrates how controlled ambiguity can serve as a diagnostic tool to bridge the gap between human psychophysical analysis, image classification, and generative image models, offering insight into human-model alignment, robustness, model interpretability, and image synthesis methods.