Can LLMs Understand the Impact of Trauma? Costs and Benefits of LLMs Coding the Interviews of Firearm Violence Survivors

arXiv cs.CL / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The study evaluates whether open-source LLMs can perform inductive thematic coding on interview transcripts from 21 Black men who survived community firearm violence, aiming to automate a traditionally labor-intensive qualitative workflow.
  • Results show that certain LLM configurations can surface some meaningful codes, but overall relevance is low and performance is highly sensitive to how the data is processed and prepared.
  • The research finds that LLM “guardrails” can cause substantial narrative erasure, reducing the models’ ability to preserve survivors’ lived experiences.
  • The paper concludes that LLM-assisted qualitative coding offers potential but has significant technical and ethical limitations when working with vulnerable, marginalized communities.

Abstract

Firearm violence is a pressing public health issue, yet research into survivors' lived experiences remains underfunded and difficult to scale. Qualitative research, including in-depth interviews, is a valuable tool for understanding the personal and societal consequences of community firearm violence and designing effective interventions. However, manually analyzing these narratives through thematic analysis and inductive coding is time-consuming and labor-intensive. Recent advancements in large language models (LLMs) have opened the door to automating this process, though concerns remain about whether these models can accurately and ethically capture the experiences of vulnerable populations. In this study, we assess the use of open-source LLMs to inductively code interviews with 21 Black men who have survived community firearm violence. Our results demonstrate that while some configurations of LLMs can identify important codes, overall relevance remains low and is highly sensitive to data processing. Furthermore, LLM guardrails lead to substantial narrative erasure. These findings highlight both the potential and limitations of LLM-assisted qualitative coding and underscore the ethical challenges of applying AI in research involving marginalized communities.