Unmasking Algorithmic Bias in Predictive Policing: A GAN-Based Simulation Framework with Multi-City Temporal Analysis
arXiv cs.AI / 3/20/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The study presents a reproducible GAN-based simulation framework that links crime occurrence to police contact to quantify racial bias in predictive policing.
- It uses 145,000+ Baltimore Part 1 crime records (2017-2019) and 233,000+ Chicago records (2022), augmented with US Census demographics, to compute four monthly bias metrics across 264 city-year observations (DIR, Demographic Parity Gap, Gini, and a Bias Amplification Score).
- Results reveal strong, year-variant bias, with Baltimore showing extreme mean annual DIR up to 15714 in 2019 and Chicago showing under-detection of Black residents (DIR ≈ 0.22) alongside persistent Gini coefficients (0.43–0.62).
- A CTGAN-based debiasing approach partially redistributes detection rates but cannot eliminate structural disparities without accompanying policy interventions.
- The analysis finds strong correlations between neighborhood racial composition and detection likelihood (Pearson r ≈ 0.83 for percent White; r ≈ −0.81 for percent Black) and shows outcomes are most sensitive to officer deployment levels; code and data are publicly available.
Related Articles
[R] Combining Identity Anchors + Permission Hierarchies achieves 100% refusal in abliterated LLMs — system prompt only, no fine-tuning
Reddit r/MachineLearning
How I Built an AI SDR Agent That Finds Leads and Writes Personalized Cold Emails
Dev.to
Complete Guide: How To Make Money With Ai
Dev.to
I Analyzed My Portfolio with AI and Scored 53/100 — Here's How I Fixed It to 85+
Dev.to
The Demethylation
Dev.to