Robust Multilingual Text-to-Pictogram Mapping for Scalable Reading Rehabilitation

arXiv cs.CL / 3/26/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a multilingual, AI-powered text-to-pictogram mapping interface intended to scale one-on-one reading support for children with SEND by adding visual scaffolding to reading text.
  • The system automatically extracts key concepts from input text and links them to contextually relevant pictograms, aiming to work across typologically diverse languages.
  • Evaluation across five languages (English, French, Italian, Spanish, and Arabic) used coverage analysis, expert clinical review, and latency testing, finding high pictogram coverage and strong visual scaffolding density.
  • Expert audits rated automatically selected pictograms as semantically appropriate, reaching combined “correct + acceptable” scores above 95% for four European languages and about 90% for Arabic, even with a smaller pictogram repository for Arabic.
  • The authors report system latency that stays within interactive thresholds for real-time educational use, supporting the feasibility and acceptability of automated multimodal scaffolding for neurodiverse learners.

Abstract

Reading comprehension presents a significant challenge for children with Special Educational Needs and Disabilities (SEND), often requiring intensive one-on-one reading support. To assist therapists in scaling this support, we developed a multilingual, AI-powered interface that automatically enhances text with visual scaffolding. This system dynamically identifies key concepts and maps them to contextually relevant pictograms, supporting learners across languages. We evaluated the system across five typologically diverse languages (English, French, Italian, Spanish, and Arabic), through multilingual coverage analysis, expert clinical review by speech therapists and special education professionals, and latency assessment. Evaluation results indicate high pictogram coverage and visual scaffolding density across the five languages. Expert audits suggested that automatically selected pictograms were semantically appropriate, with combined correct and acceptable ratings exceeding 95% for the four European languages and approximately 90% for Arabic despite reduced pictogram repository coverage. System latency remained within interactive thresholds suitable for real-time educational use. These findings support the technical viability, semantic safety, and acceptability of automated multimodal scaffolding to improve accessibility for neurodiverse learners.