Law in the Internet Society
PSRs, RAIs, and the Fight Against AI

Introduction:

Although Artificial Intelligence models have existed in some form since the 1950s, 2022 marked the beginning of what has now become known as the “AI Boom”, a term used to describe the rapid expansion of Artificial Intelligence usage into the mainstream. This technological boom, spurred by the use of large-language models like ChatGPT? ? and Meta Platforms, has become increasingly observable not only in the public sphere, but in a number of professional fields such as journalism, medicine, and, notably, law. This paper seeks to examine the potentially negative consequences of AI usage on the legal sector, specifically the judiciary. Further, it suggests some preliminary measures to limit, if not completely curb, the role AI plays in judgment.

AI and the Judiciary:

While the usage of Artificial Intelligence within the entire legal sphere has been met with rightful controversy, AI’s effect on the judiciary is especially troubling. According to the American Bar Association, numerous states have begun incorporating AI models into the judicial practice as an evaluation tool meant to aid in the generation of Pre-Sentence Reports (PSRs). Risk Assessment Tools are one specific class of AI model that rely on fact patterns and outcomes of previous cases to calculate metrics such as recidivism potential for criminal defendants. These metrics play an increasingly instrumental role in PSRs and, consequently, the sentencing outcomes of criminal cases. Sentencing courts have become increasingly reliant on these AI models to disastrous effect; already, the use of this software in PSR generation has been the subject of legal challenges on Due Process grounds. An investigative article published by ProPublica? ? highlighted one of the glaring issues with state judiciaries’ use of AI tools in criminal cases. Although limited data currently exists on these AI models, studies are beginning to show that risk assessment tools perpetuate racial bias in their assessments. The risk recidivism software COMPAS, developed by the for-profit company Equivant, serves as a shining example; Black defendants were almost twice as likely as white defendants to be wrongfully labeled as having a “high-risk” of recidivism. On the flipside, white defendants were much more likely than Black defendants to be incorrectly considered at “low-risk” of reoffense. This is far from the only problem with Artificial Intelligence models like COMPAS. Another potential issue with sentencing courts’ use of these tools is one inherent to their very nature. Artificial intelligence learns by constantly adapting its output to expanding data sets. These ever-evolving algorithms could mean skewed results for defendants as more data becomes available; the machine’s determination of a fair sentence for a defendant one day can, in theory, be completely different from its determination of a fair sentence for a future defendant with an identical fact pattern. Even further, the American Bar Association correctly posits that the use of computer-generated evaluations for determining matters such as recidivism risk removes the necessary human aspect of sentencing. Where human judges are better able to see beyond fact patterns and take more nuanced views of the defendants in front of them, AI software can only see the numbers, resulting in distressingly clinical results. With these problems in mind, it is understandable to see why the use of AI tools within the judiciary remains controversial.

Preliminary Measures:

Barring an absolute moratorium on the use of AI tools in the judiciary, which would be difficult to enforce in practice, there are mitigating measures that may be taken to minimize the negative impacts of risk assessment instruments (RAIs) on the sentencing process. For one, regulation could look like limiting what factors go into determining matters like risk recidivism in defendants. Currently, tools like COMPAS utilize information relating to a defendant’s identity when calculating risk factors– including their race, sex, and age. To avoid integrating the same biases that plague the current sentencing process into the RAI algorithms, developers should be explicitly required to exclude these demographics. Further, developing companies of RAIs should be required to publicize what considerations go into their pre-sentencing reports and risk assessments. The confidential nature of RAIs has already been the subject of legal challenge; in Loomis v. Wisconsin, a defendant raised arguments against the COMPAS software for, inter alia, not reporting what data went into the generation of his risk assessment, making it impossible to challenge the instrument’s accuracy and validity. His point was entirely valid; if pre-sentencing reports are to be made accessible to parties of a case, why should other investigative tools, like the risk assessment algorithms that help generate such reports, not be made available and open to scrutiny and potential challenge on due process grounds? Lastly, software developers should be required to analyze the algorithmic outputs of the software that they create, and publish both their process and results. In order for there to be greater transparency and scrutiny in the judiciary’s use of AI, all stakeholders need to hold equal responsibility, and accountability, for potential failings and shortcomings of the risk assessment tools. Allowing developers to gain financially from the use of their algorithms in the sentencing process without any actual stake in the outcomes will work to disincentivize them from ensuring that their models are accurate, reliable, and nondiscriminatory. While the ultimate responsibility of case outcomes should lie with the government, any party that has a stake in criminal cases should bear at least some accountability for the execution, or lack thereof, of justice. These solutions are only launching points for a longer conversation around the use of AI in the criminal justice system. There remains a larger discussion about the use of AI by police, as well as the privacy considerations that plague the integration of artificial intelligence in government as a whole. These preliminary regulations would, however, work to address the issue of AI in the judiciary pending more substantive changes. With the acceleration of the AI boom, the unregulated usage of these so-called “risk assessment tools” will only become more of a risk in-and-of-itself.

Calling everything done by statistical analysis software "AI" is surely a serious form of category error. We will then have gone from AI meaning "general artificial intelligence,"—which we don't have anywhere and which we don't have any precursor technologies for building, either—to all forms of "decision support software," which indeed by that definition I have been writing since I was fourteen.

The difference between "decision support software" and "decision-making software" is not inherent in the software. That's not an abstruse point: we all understand that human systems can rely on mechanical, mathematical or stochastic tools to make decisions. Ordeals and oracle bones are not "AI." They are, however, equally superstitious invocations of imagined divinity to perform a task which, intensely human as law-doing is, requires the people, the time and the resources necessary to its skilled accomplishment.

None of this is, as I say, difficult to put into a few hundred simple words. That both allows us to skip the "how do we regulate AI" conversation currently convulsing the class of people who believe their fundamental ignorance is a public resource, and concentrates our attention on the question we are really asking: is an illusion of precision created by reliance on "decision support" math causing avoidable social harm? Both with respect to sentencing policy and to the flawed statistical algorithms depriving public school teachers of "merit pay" under their contracts, we can probably identify dozens or hundreds of examples, each involving deprivations of life, liberty and property for large numbers of individuals and households caused by flawed automation of social, judicial and medical decision-making.

These are failures of a form of civil engineering involving our understanding (or misunderstanding) of software as a building material in the infrastructure of digital society. The arguments Zuboff makes in "The Age of Surveillance Capitalism, which I assigned but which you don't discuss, shed some important light on why failures of this sort will increase in frequency and severity in the near future, and why those harms are unequally class-distributed. It helps in thinking about remedies to have a fully-scaled understanding of the problem.

Surveillance capitalism relies on decision support for decision-making because it reduces labor's share of the resulting product at a cost either borne by labor directly or externalized to the benefit of capital. Prioritizing social outcomes over a particular division of profits results in changes in the resource management (privacy maintenance over strip-mining of human consciousness, for example) and decision support strategies of government.

When I brought one PC to Chambers in the SDNY for the first time in 1985 (and then a little tablet-sized keyboard for typing notes and saving text into the very courtroom itself) I greatly increased the productivity of two extremely hard-working law clerks. But you can be quite sure that the Judge saw to it that no machine would ever be making decisions in his court. That was why he also famously refused to read—and ostentatiously returned resealed to the Federal Probation Service in the original envelope—any presentence report that contained a numerical recommendation of any kind. You are, in that sense, pursuing an inquiry into culture rather than technology.

Sources:

Why aren't these links, anchored to the places in the text where the reader would want to reach them in the usual way, with a click, rather than being sequestered here? We are writing for the Web, where hypertext is supposed to make things easy for readers. Let's use it.

Hillman, Noel L. “The Use of Artificial Intelligence in Gauging the Risk of Recidivism.” American Bar Association, 1 Jan. 2019, www.americanbar.org/groups/judicial/publications/judges_journal/2019/winter/the-use-artificial-intelligence-gauging-risk-recidivism

Garrett, Brandon, and John Monahan. “Assessing Risk: The Use of Risk Assessment in Sentencing .” Bolch Judicial Institute at Duke Law, vol. 103, no. 2, 2019.

State v. Loomis, 371 Wis. 2d 235 (2016)

Angwin, Julia, et al. “Machine Bias.” ProPublica? ? , 23 May 2016, www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing.

18 USC Sec. 3552(d)

Navigation

Webs Webs

r2 - 12 Nov 2023 - 14:29:51 - EbenMoglen
This site is powered by the TWiki collaboration platform.
All material on this collaboration platform is the property of the contributing authors.
All material marked as authored by Eben Moglen is available under the license terms CC-BY-SA version 4.
Syndicate this site RSSATOM