Dutch probation algorithm found to be inaccurate, discriminatory
Dutch probation algorithm found to be inaccurate, discriminatory
Occurred: 2018-
Page published: February 2026
A Dutch government risk assessment algorithm used in probation services was found to produce inaccurate and potentially discriminatory predictions about people’s likelihood to reoffend, raising concerns about fairness and the use of automated tools in criminal justice decision-making.
Dutch justice inspectors found that Reclassering Nederland (the probation service) had used an error‑ridden algorithmic system to inform judges and prosecutors about individuals’ risk of reoffend. The system was used approximately 44,000 times a year.
In one key tool, formulas were mixed up so the model never correctly predicted recidivism risk, leading to wrongly classified risk levels in around a quarter of probation reports used by judges and parole boards.
These risk scores fed into advice on sentencing and early release, leading the tool consistently to underestimate the risk of reoffending, particularly among individuals with substance abuse issues or serious mental health conditions.
Others may have received unnecessarily harsh or restrictive conditions.
The inspection also found that the OXREC risk model used neighbourhood “poverty scores” and income data, variables known to correlate with migration background, which created a risk of indirect discrimination against poorer and racialised groups.
The Inspectorate also found that staff were pressured to ignore their professional judgment. An internal podcast even told employees their own intuition was "as reliable as flipping a coin," leading to automation bias, where the machine's errors were rarely questioned.
Following public reporting and the inspectorate’s conclusions, the probation service temporarily halted use of the systems and acknowledged serious shortcomings in how the algorithms had been designed and managed.
The core issue appears to stem from design and implementation flaws: the algorithm was not properly calibrated for the Dutch probation population and used outdated or mismatched data, while local deployment introduced errors such as applying models intended for one group to another.
These technical shortcomings, coupled with a tendency for staff to over-rely on the algorithm’s outputs rather than professional judgment, amplified the impact of inaccurate scores in judicial contexts.
Despite signals as early as 2018 about limited predictive value and a 2020 academic warning about possible discriminatory effects, the service continued using the model and did not conduct robust human‑rights or privacy impact assessments.
The algorithm included income and neighbourhood “deprivation” scores even though the Dutch Human Rights Institute had previously warned that such variables are effectively prohibited unless strong safeguards and justification are in place, which were missing here.
More broadly, the case reflects a familiar pattern in Dutch public sector algorithm use, where pressure to automate risk assessment outpaced transparency, documentation, and accountability mechanisms, similar to earlier welfare and tax‑benefit profiling scandals.
For directly affected prisoners, the system has resulted in a slew of unfair, discriminatory decisions that directly impact their lives, as well as those of their families.
For the general public, the fracas undermines trust in the Netherlands' justice system, with individuals subjected by an opaque "black box" system that neither the defendant nor the judge fully understands to harsher or more lenient treatment based on their neighbourhood or income rather than their actions. Furthermore, over-reliance on the model risks weakening public safety, with some prisoners likey to have not received targeted assistance that could reduce reoffending and improve long-term outcomes.
For policymakers and justice system administrators, it underlines the need for robust validation, transparency, and human-in-the-loop safeguards when deploying high-stakes algorithms, and may fuel calls for stronger regulation, auditing standards, and accountability mechanisms for government predictive systems.
OxRec
Developer: University of Oxford
Country: Netherlands
Sector: Govt - justice
Purpose: Assess reoffending risk
Technology: Prediction algorithm
Issue: Accountability; Accuracy/reliability; Automation bias; Bias/discrimination; Fairness; Transparency
Inspectorate of Justice and Security. Risicovol algoritmegebruik door reclassering
AIAAIC Repository ID: AIAAIC2194