Algorithms were going to save us, make our system more fair, more precise, more . . . real. Even bad algorithms were better than no algorithms, leaving it up to biased, if minimally competent, judges to make decisions they were incapable of making. Algorithms. Yeah, that’s the ticket.
There were, of course, some Luddites who didn’t appreciate the power of numbers, complained that algorithms would never be capable of taking into account the particulars of every defendant’s world. They would homogenize decision-making, reduce people to caricatures and, ultimately, end up replacing human bias with machine bias. We would feel more science-y about the law, but it would do nothing more than provide algorithmic cover for the same shallow bias it replaced.
For example, when ProPublica examined computer-generated risk scores in Broward County, Fla., in 2016, it found that black defendants were substantially more likely than whites to be rated a high risk of committing a violent crime if released. Even among defendants who ultimately were not re-arrested, blacks were more likely than whites to be deemed risky. These results elicited a visceral sense of injustice and prompted a chorus of warnings about the dangers of artificial intelligence.
And as it turned out, the algorithms used effective proxies for race and poverty, so they gave the appearance of being very high-tech while employing the same quantitative factors that humans applied instinctively. When judges did it on their own, it was called bias. Then machines did it, it was called science. And it turned out to be pretty much the same.
The next step in the process was taking the algorithm’s outcome and then introducing a post-hoc human tweak to rid it of the biased outcome. After all, if the basic premise was that algorithms would eliminate racial prejudice from the process, then produced disparate outcomes anyway, the algorithms had to be biased. Of course, the irony of human tweaking of algorithmic outcomes to align with ideology to remove bias from the system by re-introducing human bias into the system didn’t seem to bother people too much.
But now, a study shows that we’ve done no more than engage in a combination of mental and techno masturbation, albeit with the best of intentions.
The academics used a database of more than 7,000 pretrial defendants from Broward County, Florida, which included individual demographic information, age, sex, criminal history and arrest record in the two year period following the Compas scoring.
The online workers were given short descriptions that included a defendant’s sex, age, and previous criminal history and asked whether they thought they would reoffend. Using far less information than Compas (seven variables versus 137), when the results were pooled the humans were accurate in 67% of cases, compared to the 65% accuracy of Compas.
So random people did better than COMPAS? Even worse, while COMPAS used 137 measures, it turned out that only two mattered.
In a second analysis, the paper found that Compas’s accuracy at predicting recidivism could also be matched using a simple calculation involving only an offender’s age and the number of prior convictions.
“When you boil down what the software is actually doing, it comes down to two things: your age and number of prior convictions,” said Farid. “If you are young and have a lot of prior convictions you are high risk.”
Try as we might, the numbers refuse to play along. As it turns out, the same factors that judges intuitively applied all along were the best predictors of recidivism. No, they were hardly conclusive predictors, and no, they didn’t address root causes of the problem, such as why poor black defendants had a rap sheet as long as their arm in the first place, but that’s not the question for the legal system or the judge in deciding bail or sentence.
The paper also highlights the potential for racial asymmetries in the outputs of such software that can be difficult to avoid – even if the software itself is unbiased.
The analysis showed that while the accuracy of the software was the same for black and white defendants, the so-called false positive rate (when someone who does not go on to offend is classified as high risk) was higher for black than for white defendants. This kind of asymmetry is mathematically inevitable in the case where two populations have a different underlying rate of reoffending – in the Florida data set the black defendants were more likely to reoffend – but such disparities nonetheless raise thorny questions about how the fairness of an algorithm should be defined.
The reveal is that science doesn’t necessarily align with ideology. If the algorithm is unbiased, and algorithms are only as good or bad as their input, and still don’t yield the results one expects of them, is the fault in the algorithm or expectations?
Farid said the results also highlight the potential for software to magnify existing biases within the criminal justice system. For instance, if black suspects are more likely to be convicted when arrested for a crime, and if criminal history is a predictor of reoffending, then software could act to reinforce existing racial biases.
Or, as argued from the outset, the problem isn’t on the back end, where algorithms are employed in lieu of bespoke human involvement, but on the front end, with cops tossing poor black kids for kicks, because that’s where the action happens. Once a defendant has a conviction, it all goes downhill from there. Jobs, school, housing, all of it. And they end up being recidivists because what else are they going to do?
There’s nothing new about the input causes of recidivism, about how poverty means the inability to make bail, imposed for no particularly good reason, which produces pleas of convenience to get out of jail as soon as possible. How cops use ghetto kids to make their numbers, because that’s where they believe all the bad seeds can be found.
But none of this is going to be cured by algorithms. It never was. Garbage in, garbage out, and we’ve been putting garbage into the criminal justice system forever. What did we expect to be the outcome? The Sentence-O-Matic 1000 isn’t broken. It never worked in the first place, and it never will.
Discover more from Simple Justice
Subscribe to get the latest posts sent to your email.
Scott you need to stay away from me on Twitter. I know that you’re harassing me online and breaking into my accounts. I’m going to report you to every LE agency that I can tomorrow, and the bar assoc. Beth Bell is a sociopath who screws every guy she can and lies about it to her husband. Please stay away from me!
Fair enough, whoever you are. But leave dear Beth out of your delusion.
Stay away from me
Now you’re just getting pushy.
That won’t stop the tech companies creating these algos to bribe…*ahem* I mean make campaign contributions to *ahem*… politicians so that they can line their pockets with fat governmental contracts.
There remains one player in the system who can just say no to algorithms. Will they?
“Your Honor…it’s science! With statistics, and numbers, and computers. Did I mention that I’m the prosecutor?”
“Admitted.”
That’s the decision they have to make. Among my many concerns, beyond the fact that judges try hard to understand science and statistics, is that science provides cover for the same ills that pervade the system, but now it’s no longer anyone’s fault. It’s science.
Maybe after they sign the search warrant….but even then, since when is “trying” adequate?
This has nothing to do with search warrants. Not even a little bit.
WTF FUCK?! You must have reason to protect hamsters.
https://i.pinimg.com/originals/b6/13/ee/b613ee1867a83977b77864abc7286b07.jpg
anyway, don’t forget to strangle the guinea pig .
just saying!
Algorithms do have the one, saving grace that they generate a number, and do not go off the rails with editorial commentary, as the judge in the Nassar sentencing did today. Even if the algorithms are basing the number they generate, in whole or part, on somebody’s perceived need to “end today” the alleged statistical prevalence of some type of conduct in society at large, the algorithms will not say that on the record. Neither will the algorithms ever tell a defendant sent to prison for sexual misconduct, “I have just signed your death warrant.” In some courtrooms, algorithms would be a distinct improvement.
But that’s not how they work. They merely give the judge a number. The judge still pronounces sentence, along with whatever critical bits of personal thought and advice she has to offer.
But that part could be fixed. They could put an AI-equipped Harmony doll on the bench and set the algorithm to just download the years and months of the sentence to its speech center. To avoid discriminatory or biased commentary with sentence, they could use a standard format. “Hi. You have __ years, __ months to serve. Have a nice day.”
Alas, I can no longer defend the SOM. I’ve learned much from SJ…It was the wrong solution for an imagined problem when I first uttered it. The criminal justice system is not in need of brutal cold efficiency or beurocratic definition of sentencing guidelines.
However, your point does raise a problem that requires more attention: Why do black and brown skin men and women pay a heavier and more frequent toll for the same sins whites are committing? And more importantly: How do we remove that bias?
You know how much I like it when you address justice writ large. Please…Go on.
PS- I did see your offer and a candidate submission is forthcoming.
So you don’t want me to cure cancer in the comments too?
Could you?
That would be great. It might even get you that Supreme Court nomination.
The people who write the algorithms can inadvertently include their biases, albeit in attenuated form if they are careful, into the algorithms. Bias-free algorithms are impossible…until using artificial intelligence to write algorithms.
But what if, as I think would be likely, AI generated sentencing algorithms show the same allegedly disparate results as human judges?
From what I’ve been reading the algorithm was created in 1998, pretty damn old, and it’s never worked right. It really surprises me that it hasn’t been replaced or questioned seriously. Just shows how lazy the courts and judges really are at learning and thinking critically about some subjects.
The thing that gets me with these things, math can only explain the past, it’s pretty worthless at guessing the future. This idea that math can replace the hard work of thinking is an amazing urban myth that boggles my mind at just how many people buy into the bs religion of new and shiny.
You’re not allowed to ever mention math again. Strike it from your vocabulary.
If COMPAS didn’t reflect the biases of the people who make the purchasing decisions, nobody would buy it, and we wouldn’t be talking about it. If I think that the vendor is purposely tweaking COMPAS’ invisible, trade-secrety innards to make it predict just like a typical law-enforcement guy, does that make me a cynic?
Yes, that was very cynical. Reddit quality cynical.