Law & Policy

AI in the Courtroom: Can Algorithms Deliver Justice or Do They Encode Injustice?

AI risk assessment tools are already used in bail, sentencing, and parole decisions across multiple jurisdictions. Five papers examine whether these tools mitigate human bias or encode historical discrimination into the machinery of justiceโ€”and whether algorithmic justice can be democratically legitimate.

By Sean K.S. Shin
This blog summarizes research trends based on published paper abstracts. Specific numbers or findings may contain inaccuracies. For scholarly rigor, always consult the original papers cited in each post.

In courtrooms across the United States, algorithmic risk assessment tools inform decisions about who receives bail, who is sentenced to prison, and who is released on parole. COMPAS (Correctional Offender Management Profiling for Alternative Sanctions), PSA (Public Safety Assessment), and their successors analyze defendant dataโ€”criminal history, age, employment status, social tiesโ€”and generate risk scores that judges use as inputs to their decisions.

The promise is objectivity: algorithms do not have bad days, harbor racial prejudice, or vary their assessments based on when they last ate lunch (unlike human judges, whose sentencing decisions have been shown to correlate with the time since their last meal). The concern is that objectivity about a biased world merely systematizes the biasโ€”producing decisions that are consistently unfair rather than inconsistently unfair.

Can AI Mitigate or Exacerbate Bias?

Gao (2025) explores AI's dual role in criminal sentencingโ€”acknowledging that AI-driven tools can both reduce and amplify bias depending on design choices, data quality, and implementation context. As AI-driven tools increasingly integrate into global criminal justice systems, algorithmic justice has become a critical concern.

The paper identifies conditions under which AI reduces bias: when it replaces human decision-makers who exhibit demonstrable racial or socioeconomic prejudice, when it uses validated risk factors with empirical predictive power, and when its outputs are subject to meaningful human review.

It also identifies conditions under which AI amplifies bias: when training data reflects historical discrimination (e.g., higher arrest rates for Black Americans produce higher predicted risk scores), when proxy variables (zip code, employment status) correlate with protected characteristics (race, ethnicity), and when risk scores are treated as objective truth rather than probabilistic estimates.

Algorithmic Inequity

Schneider (2025) critically examines the phenomenon of algorithmic inequity within legal systems, focusing on how AI systems can perpetuate or deepen existing social inequalities when deployed in judicial contexts.

The analysis identifies a structural problem: AI systems trained on criminal justice data learn patterns from a system that has historically over-policed and over-incarcerated Black and Brown communities. A risk assessment tool that predicts recidivism based on prior arrests is, in effect, predicting future police attention rather than future criminal behavior. The prediction is accurateโ€”people who have been arrested before are more likely to be arrested againโ€”but the accuracy reflects policing patterns, not individual dangerousness.

This creates a feedback loop: algorithmic risk assessment โ†’ higher sentences for high-risk individuals โ†’ increased contact with the criminal justice system โ†’ higher future risk scores. The algorithm does not merely predict recidivism; it contributes to producing it.

Suggestive vs. Decisional Algorithms

Comoglio (2025) introduces an important distinction: between suggestive algorithms (which recommend outcomes for human judges to consider), predictive algorithms (which estimate probabilities of future events), and decisional algorithms (which autonomously determine legal outcomes). The paper argues that the legal and ethical analysis should differ depending on which type is deployed.

Suggestive algorithms preserve judicial discretion: the judge receives the risk score as one input among many and retains the authority to override it. The risk is that judges defer to algorithmic recommendationsโ€”either because they trust the technology or because they want to deflect responsibility for unpopular decisions.

Decisional algorithms eliminate judicial discretion entirely. No jurisdiction currently uses fully automated sentencing, but automated bail decisions, automated parole risk classification, and automated fine calculation are in use or under development. The democratic legitimacy concerns for decisional algorithms are qualitatively different from those for suggestive algorithms.

Fairness Verification

Zheng (2025) addresses the technical challenge of ensuring that AI criminal justice systems are fair. AI is increasingly utilized in criminal justice to support decisions related to bail, sentencing, and parole. However, these systems often perpetuate historical biases, particularly racial disparities embedded in the training data.

The paper proposes fairness verification algorithms and bias mitigation mechanisms. Many existing models lack effective mechanisms for ensuring fairness while maintaining predictive accuracy. The technical challenge is that different fairness metrics (demographic parity, equalized odds, predictive parity) are mathematically incompatibleโ€”a system cannot satisfy all fairness criteria simultaneously. This means that "fair AI" requires a choice about which kind of fairness to prioritizeโ€”a choice that is fundamentally political, not technical.

Democratic Legitimacy

Vidaki and Papakonstantinou (2025) raise a question that technical fairness analyses often neglect: can AI in judicial decision-making be democratically legitimate? The question goes beyond whether algorithmic decisions are accurate or fair to whether they are legitimateโ€”whether they carry the authority that democratic societies require of their justice systems.

Democratic legitimacy in criminal justice derives from several sources: legislation enacted by elected representatives, judicial reasoning that can be scrutinized and appealed, procedural protections that ensure defendants are heard, and the personal accountability of judges who exercise judgment on behalf of the community. Algorithmic decision-making disrupts each of these sources: the algorithm's logic may be proprietary, its reasoning may be opaque, and no individual bears personal responsibility for its outputs.

Claims and Evidence

<
ClaimEvidenceVerdict
AI can reduce human bias in sentencingGao (2025): possible under specific design and implementation conditionsโš ๏ธ Uncertain (conditional)
AI risk assessment encodes historical discriminationSchneider (2025), Zheng (2025): training data reflects discriminatory policing and sentencing patternsโœ… Supported
Suggestive algorithms preserve judicial discretionComoglio (2025): formally yes, but judicial deference to algorithms is documentedโš ๏ธ Uncertain
Technical fairness metrics can resolve algorithmic biasZheng (2025): different fairness metrics are mathematically incompatible; choice is politicalโŒ Refuted (as purely technical solution)
Algorithmic judicial decision-making is democratically legitimateVidaki & Papakonstantinou (2025): multiple sources of democratic legitimacy are disruptedโŒ Refuted (without reform)

Open Questions

  • Should defendants have a right to know their algorithmic risk score? Transparency would enable challenge but might also create self-fulfilling prophecies if individuals internalize their "risk" classification.
  • Can algorithmic risk assessment be designed for rehabilitation rather than punishment? Current tools predict recidivism. Could AI tools instead predict which interventions (education, employment support, counseling) would reduce reoffending for each individual?
  • Who is liable when an algorithm contributes to a wrongful conviction? The developer, the jurisdiction that deployed the tool, or the judge who relied on it? Current law provides no clear answer.
  • Should AI in criminal justice be subject to the same standards as medical devices? Both affect human welfare. Should algorithmic risk assessment tools undergo independent validation and regulatory approval before deployment?
  • Implications

    The research reviewed here suggests that AI in criminal justice is neither the objective oracle its advocates promise nor the discrimination machine its critics fear. It is a tool whose effects depend on design choices, implementation contexts, and governance structures. The question is not whether to use AI in criminal justice but under what conditions, with what safeguards, and with what accountability mechanisms.

    The evidence supports several design principles: training data should be audited for historical bias before use; fairness metrics should be chosen through democratic deliberation, not technical default; algorithmic outputs should be treated as recommendations, not decisions; and defendants should have meaningful rights to challenge algorithmic assessments.

    References (5)

    [1] Gao, Y. (2025). Algorithmic Justice: Can AI Mitigate or Exacerbate Bias in Criminal Sentencing?.
    [2] Schneider, J. (2025). Algorithmic Inequity in Justice: Unpacking the Societal Impact of AI in Judicial Decision-Making. IJAAIR, 2(1), 02.
    [3] Comoglio, P.M. (2025). Have Your DAI in Court: The Role of Suggestive Algorithms in Judicial Decision-Making. Proc. ACM, 3769135.
    [4] Zheng, L. (2025). Fairness Verification Algorithms and Bias Mitigation Mechanisms for AI Criminal Justice Decision Systems. International Journal of Law and Information Technology.
    [5] Vidaki, A.N. & Papakonstantinou, V. (2025). Democratic Legitimacy of AI in Judicial Decision-Making. AI and Ethics.

    Explore this topic deeper

    Search 290M+ papers, detect research gaps, and find what hasn't been studied yet.

    Click to remove unwanted keywords

    Search 8 keywords โ†’