Probing Multimodal Large Language Models on Cognitive Biases in Chinese Short-Video Misinformation
arXiv cs.CL / 5/4/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper presents a multimodal evaluation framework to test how Multimodal Large Language Models handle Chinese short-video misinformation that is intertwined with cognitive biases.
- It introduces a manually annotated dataset of 200 short videos across four health domains, with evidence-verified labels for three deceptive patterns: experimental errors, logical fallacies, and fabricated claims.
- Eight frontier MLLMs are evaluated under five modality settings, with Gemini-2.5-Pro achieving the best multimodal performance (belief score 71.5/100) and o3 the lowest (35.2).
- The study analyzes social cues in misinformation videos and finds models can form false beliefs due to biases such as authoritative channel IDs.
Related Articles
A very basic litmus test for LLMs "ok give me a python program that reads my c: and put names and folders in a sorted list from biggest to small"
Reddit r/LocalLLaMA

ALM on Power Platform: ADO + GitHub, the best of both worlds
Dev.to

Experiment: Does repeated usage influence ChatGPT 5.4 outputs in a RAG-like setup?
Dev.to

Find 12 high-volume, low-competition GEO content topics Topify.ai should rank on
Dev.to

When a memorized rule fits your bug too well: a meta-trap of agent workflows
Dev.to