Eric Loomis algorithmic risk assessment accused of denying due process
Occurred: 2016
Report incident 🔥 | Improve page 💁 | Access database 🔢
The use of a controversial algorithmic recidivism tool to sentence Eric Loomis to six years in prison for driving a car used in shooting was criticised and appealed as a denial of due process.
Eric Loomis was sentenced to six years in prison for driving a car used in a shooting by a judge partly relying on the Correctional Offender Management Profiling for Alternative Sanctions (COMPAS) tool.
The ruling led to Loomis lodging a legal appeal on the basis that the score assigned to him could not be assessed due to its protection as a trade secret, and that the use of COMPAS infringed on his right to an individualised sentence and his right to be sentenced on the basis of accurate information.
He also argued that the court unconstitutionally considered gender at sentencing by relying on a risk assessment that took gender into account.
COMPAS risk assessments are based on data gathered from a defendant's criminal file and from an interview with the defendant, and predict the risk of pretrial recidivism, general recidivism, and violent recidivism.
A registered sex offender, Loomis scored high on all three risk measures, and was sentenced to six years in prison. The judge said he had arrived at his sentencing decision in part because of Mr. Loomis’s rating on the Compas assessment. Loomis then filed a motion requesting a new sentencing hearing, which was denied by all three levels of Wisconsin courts.
Nonetheless, the court advised that judges presented with COMPAS risk scores understood that:
There was no disclosure of how factors were weighed or risk scores were determined;
That the assessment compared defendants to a national sample and no cross-validation study of Wisconsin defendants has been completed;
Some studies raised questions about whether COMPAS disproportionately gives minorities higher risk scores; and
Risk assessment tools must be updated due to changing populations.
The judge also ruled that the use of COMPAS did not deny due process as the court ultimately imposes sentences on the basis of all of its knowledge of the defendant, including their criminal histories.
The incident raised questions about the opacity, accountability, fairness and efficacy of the tool and of tools similar to it, such as the Ontario Domestic Assault Risk Assessment (ODARA) and the Virginia Non-violent Risk Assessment (NVRA).
It also prompted debate about the use of big data and algorithms to predict crime (aka 'predictive policing').
System 🤖
Documents 📃
Equivant (2016). Wisconsin vs. Loomis: Court Affirms the Use of COMPAS in Sentencing
Operator: Wisconsin Court System
Developer: Volaris Group/Equivant/Northpointe
Country: USA
Sector: Govt - justice
Purpose: Assess recidivism risk
Technology: Recidivism risk assessment system
Issue: Accountability; Bias/discrimination - race, ethnicity; Human/civil rights
Transparency: Governance; Black box
Legal, regulatory 👩🏼⚖️
George Washington University. ETI AI Litigation Database: case details
Supreme Court of Wisconsin (2016). State v Loomis (pdf)
Research, advocacy 🧮
Dressel J., Farid H. (2018). The accuracy, fairness, and limits of predicting recidivism
Kirkpatrick K. (2017). It's not the algorithm, it's the data
Angelino E., Larus-Stone N., Alabi D., Seltzer M., Rudin R. (2017) Learning Certifiably Optimal Rule Lists for Categorical Data
News, commentary, analysis 🗞️
https://www.technologyreview.com/2017/06/01/151447/secret-algorithms-threaten-the-rule-of-law/
https://www.wired.com/2017/04/courts-using-ai-sentence-criminals-must-stop-now/
https://www.theatlantic.com/technology/archive/2018/01/equivant-compas-algorithm/550646/
https://today.duke.edu/2017/07/opening-lid-criminal-sentencing-software
Page info
Type: Incident
Published: March 2023
Last updated: March 2024