AI Navigate

Man and machine: artificial intelligence and judicial decision making

arXiv cs.AI / 3/20/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper surveys the integration of artificial intelligence into judicial decision-making, focusing on pretrial, sentencing, and parole contexts and highlighting concerns about transparency, reliability, and accountability.
  • It synthesizes evidence on AI performance and fairness, human judge biases, and the way judges interact with algorithmic recommendations across computer science, economics, law, criminology, and psychology.
  • Findings indicate that AI decision aids have modest or inexistent impact on pretrial and sentencing decisions, while identifying important gaps in the literature and a need for further interdisciplinary evaluation of AI tools and human decision-making.
  • The authors argue that AI-versus-human comparisons can yield new insights and advocate greater interdisciplinary integration to advance future research.

Abstract

The integration of artificial intelligence (AI) technologies into judicial decision-making - particularly in pretrial, sentencing, and parole contexts - has generated substantial concerns about transparency, reliability, and accountability. At the same time, these developments have brought the limitations of human judgment into sharper relief and underscored the importance of understanding how judges interact with AI-based decision aids. Using criminal justice risk assessment as a focal case, we conduct a synthetic review connecting three intertwined aspects of AI's role in judicial decision-making: the performance and fairness of AI tools, the strengths and biases of human judges, and the nature of AI+human interactions. Across the fields of computer science, economics, law, criminology and psychology, researchers have made significant progress in evaluating the predictive validity of automated risk assessment instruments, documenting biases in judicial decision-making, and, to a more limited extent, examining how judges use algorithmic recommendations. While the existing empirical evidence indicates that the impact of AI decision aid tools on pretrial and sentencing decisions is modest or inexistent, our review also reveals important gaps in the canvassed literatures. Further research is needed to evaluate the performance of AI risk assessment instruments, understand how judges navigate noisy decision making environments and how individual characteristics influence judges' responses to AI advice. We argue that AI vs Human comparisons have the potential to yield new insights into both algorithmic tools and human decision-makers and advocate greater interdisciplinary integration and cross-fertilization in future research.