LaMSUM: Amplifying Voices Against Harassment through LLM Guided Extractive Summarization of User Incident Reports
arXiv cs.CL / 4/20/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces LaMSUM, a multi-level framework that uses LLMs to produce extractive summaries (selecting key excerpts) of user incident reports rather than only abstractive paraphrases.
- It is designed to handle large volumes of citizen-reported sexual harassment data, addressing practical constraints such as LLM context window limits and the need to support various code-mixed languages.
- LaMSUM combines summarization with multiple voting methods to improve the quality and reliability of the selected excerpts across large collections of reports.
- The authors evaluate the approach on incident-report summarization using four widely used LLMs (Llama, Mistral, Claude, and GPT-4o) and report that LaMSUM outperforms existing state-of-the-art extractive summarization baselines.
- The work aims to help relevant stakeholders quickly obtain comprehensive overviews of incidents, supporting better policy development to reduce unwarranted harassment.
Related Articles

From Theory to Reality: Why Most AI Agent Projects Fail (And How Mine Did Too)
Dev.to

GPT-5.4-Cyber: OpenAI's Game-Changer for AI Security and Defensive AI
Dev.to

Building Digital Souls: The Brutal Reality of Creating AI That Understands You Like Nobody Else
Dev.to
Local LLM Beginner’s Guide (Mac - Apple Silicon)
Reddit r/artificial

Is Your Skill Actually Good? Systematically Validating Agent Skills with Evals
Dev.to