Published in AI

AI touted as cure for faulty justice

by on23 October 2019


Robot judges could remove bias

AI is being looked at as a way of removing human bias from court cases, according to Forbes.

Computer researchers get closer to creating true Artificial Intelligence - it's predicted to eliminate most paralegal and legal research positions within the next decade.

Lawyer Tom Girardi thinks that the next step inevitably involves artificial intelligence aiding, or even completely replacing lawyers. And if we have robot lawyers, why not automated judges and juries too?

Girardi said: “It may even be considered legal malpractice not to use AI one day. It would be analogous to a lawyer in the late twentieth century still doing everything by hand when this person could use a computer. If a lawyer can use AI to win a case and do it for less than someone without AI, who do you think the client will choose to work with next time?"

Writer Rossalyn Warren points out that people are, by their nature, flawed. Flesh and blood jurors and judges will always bring their own prejudices into the courtroom.

But a robot juror, she says, "could be crammed with a far broader range of facts and figures about the nature of the crime, cases on record and the law, making it much more worthwhile than a juror who has little awareness on such matters.”

She said: “Expecting randomly selected members of the public to decide the fate of a person in a jury system is outdated because the notion of a fair and impartial jury doesn’t exist.”

Even cross examinations could be outsourced to an automated system. A thought-provoking experiment shows that people are more likely to be completely honest with an unemotional machine than a potentially judgemental human.

When researchers led by Jonathan Gratch at the Institute for Creative Technologies created an artificially intelligent robot psychologist named Ellie, they tested it on two groups of people.

Half were told Ellie was just a machine that was able to ask probing questions and understand their respondents’ emotions with 3D cameras. Those people were shown to give more honest responses to "her" while the experimental subjects that were told that Ellie was being operated by a human ‘puppeteer’ gave less direct answers.

Apart from the possibility of getting a fairer result, raw economics come into play too.

But a solid-state legal system with no humans involved isn’t necessarily more error-proof than our existing system. Former Prime Minister Theresa May, in an address last year to the Davos Forum, pointed out that we need to develop a set of laws governing Artificial Intelligence so we can “make the most of AI in a responsible way, such as by ensuring that algorithms don’t perpetuate the human biases of their developers.

 

Last modified on 23 October 2019
Rate this item
(0 votes)

Read more about: