Stay up to date with notifications from The Independent

Notifications can be managed in browser preferences.

Judge admits using ‘jolly useful’ ChatGPT to write court ruling

Appeal court judgment included words generated by controversial chatbot renowned for accuracy problems

Jane Dalton
Friday 15 September 2023 19:56 BST
Comments
(Getty Images)

Support truly
independent journalism

Our mission is to deliver unbiased, fact-based reporting that holds power to account and exposes the truth.

Whether $5 or $50, every contribution counts.

Support us to deliver journalism without an agenda.

Louise Thomas

Louise Thomas

Editor

An appeal court judge has admitted using artificial intelligence chatbot ChatGPT to help him write a court ruling.

Lord Justice Birss said the language-processing tool was “jolly useful” and the technology had “real potential”.

Scientists, writers and other professionals have previously found ChatGPT’s accuracy unreliable since it was launched last year, and it has become known for having a “hallucination problem” in which false information is generated.

The Royal Courts of Justice - the appeal court
The Royal Courts of Justice - the appeal court (Getty Images)

Earlier this year, ChatGPT falsely accused an American law professor by including him in a generated list of legal scholars who had sexually harassed someone, citing a non-existent The Washington Post report.

PCGuide.com says: “ChatGPT is not a truly reliable source. There’s no denying that it is one of the best artificial intelligence content-generator tools out there, but the accuracy on many topics is still not as good as you would want it to be.”

According to the Law Society Gazette, Lord Justice Birss spoke about AI, ChatGPT and generative large language models at a conference, saying: “I think what is of most interest is that you can ask these large language models to summarise information. 

“It is useful and it will be used and I can tell you, I have used it.

“I thought I would try it. I asked ChatGPT can you give me a summary of this area of law, and it gave me a paragraph. I know what the answer is because I was about to write a paragraph that said that, but it did it for me and I put it in my judgment.

“It’s there and it’s jolly useful. I’m taking full personal responsibility for what I put in my judgment – I am not trying to give the responsibility to somebody else.

“All it did was a task which I was about to do and which I knew the answer and could recognise an answer as being acceptable.”

Three months ago, a New York lawyer who used ChatGPT to write a legal brief and ended up citing bogus cases profusely apologised in court.

Steven Schwartz became emotional as he explained being “duped” by the artificial intelligence chatbot.

“I deeply regret my actions in this manner that led to this hearing today,” Mr Schwartz said. “I suffered both professionally and personally [because of] the widespread publicity this issue has generated. I am both embarrassed, humiliated and extremely remorseful.”

Together with a colleague, he and their law firm were fined  $5,000 (£3,935).

UK law firm Mishcon de Reya has banned lawyers from using ChatGPT because of fears they risk compromising data.

In July, two new US studies, from Stanford and UC Berkeley universities, concluded that ChatGPT appeared to be getting less accurate over time.

Join our commenting forum

Join thought-provoking conversations, follow other Independent readers and see their replies

Comments

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in