Literary Narrative as Moral Probe : A Cross-System Framework for Evaluating AI Ethical Reasoning and Refusal Behavior
arXiv cs.AI / 3/16/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper critiques current AI moral evaluation frameworks for relying on surface-level responses and proposes a novel literary-narrative probe using unresolvable moral scenarios from a published science fiction series to elicit genuine moral reasoning.
- It presents a 24-condition cross-system study spanning 13 AI systems across two series (frontier commercial systems and local/API open-source systems) with both blind and declared administrations, totaling 24 conditions.
- The study employed multiple judges (Claude as LLM judge; Gemini Pro and Copilot Pro as independent ceiling-discrimination judges) and found zero delta across 16 dimension-pair comparisons, with perfect rank-order agreement on a theological differentiator probe between Gemini Pro and Copilot Pro (rs = 1.00).
- Five qualitative reflexive failure modes were identified (including categorical self-misidentification and false positive self-attribution), supporting the claim that instrument sophistication scales with system capability and that literary narrative is an anticipatory, deployment-relevant evaluation instrument for high-stakes AI ethics.
Related Articles

The programming passion is melting
Dev.to

Maximize Developer Revenue with Monetzly's Innovative API for AI Conversations
Dev.to
Co-Activation Pattern Detection for Prompt Injection: A Mechanistic Interpretability Approach Using Sparse Autoencoders
Reddit r/LocalLLaMA

How to Train Custom Language Models: Fine-Tuning vs Training From Scratch (2026)
Dev.to

KoboldCpp 1.110 - 3 YR Anniversary Edition, native music gen, qwen3tts voice cloning and more
Reddit r/LocalLLaMA