Multimodal Analysis of State-Funded News Coverage of the Israel-Hamas War on YouTube Shorts
arXiv cs.CL / 4/3/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces a multimodal analysis pipeline for YouTube Shorts that combines automatic transcription, aspect-based sentiment analysis (ABSA), and semantic scene classification.
- The pipeline is evaluated for feasibility and then applied to more than 2,300 conflict-related Shorts and 94,000+ visual frames covering the Israel-Hamas war from state-funded broadcasters.
- Results show that sentiment in transcripts varies across outlets and changes over time by specific aspects, while scene-type classifications align with visual cues consistent with real-world events.
- The study finds that smaller domain-adapted models outperform large transformers and even LLMs for sentiment analysis, highlighting the effectiveness of resource-efficient approaches for humanities research.
- The authors present the pipeline as a reusable template for other short-form platforms (e.g., TikTok, Instagram) to study how algorithmic video environments encode sentiment and visual signals.
Related Articles

Black Hat Asia
AI Business

Mistral raises $830M, 9fin hits unicorn status, and new Tech.eu Summit speakers unveiled
Tech.eu

ChatGPT costs $20/month. I built an alternative for $2.99.
Dev.to

OpenAI shifts to usage-based pricing for Codex in ChatGPT business plans
THE DECODER

Why I built an AI assistant that doesn't know who you are
Dev.to