ChatGPT Use Goes Wrong in Court By Brian Simpson

This story shows the dangers in ChatGPT, for people mechanically relying on it. A New York lawyer prepared a case using ChatGPT, and it generated a brief that contained false cases and citations, pulled out of thin air! That lawyer trusted the computer. Fortunately, the court checked in the real world and found the errors, and the lawyer got a snack. The moral here is that there is a problem in how one knows that the ChatGPT output is actually reliable? How, and why, should one trust it? I don’t think one can given its present developments, so this is another reason to beware.   

https://cointelegraph.com/news/lawyer-chat-gpt-court-filing-legal-research

“The court used the word “bogus” to describe the quotes and citations the New York attorney provided, which were sourced from ChatGPT.

A New York attorney has been blasted for using ChatGPT for legal research as part of a lawsuit against a Columbian airline.

Steven Schwartz, an attorney with the New York law firm Levidow, Levidow & Oberman, was hired by Robert Mata to pursue an injury claim against Avianca Airlines.

Mata claims he sustained the injury from a serving cart during his flight with the airline in 2019, according to a May 28 report from CNN Business.

However, after a judge noticed inconsistencies and factual errors in the case documentation, Schwartz has admitted to using ChatGPT for his legal research, according to a May 24 sworn affidavit.

He claims that this was his first time using ChatGPT for legal research and “was unaware of the possibility that its content could be false.”

In an April 5 court filing, the judge presiding over the case stated:

“Six of the submitted cases appear to be bogus judicial decisions with bogus quotes and bogus internal citations.”

The judge further claimed that certain cases referenced in the submissions did not exist, and there was an instance where a docket number on a filing was mixed up with another court filing. 

Schwartz said he also regrets having trusted the artificial chatbot without conducting his own due diligence. The affidavit noted:

“[Schwartz] Greatly regrets having utilized generative artificial intelligence to supplement the legal research performed herein and will never do so in the future without absolute verification of its authenticity.”

 

 

Comments

No comments made yet. Be the first to submit a comment
Already Registered? Login Here
Tuesday, 16 April 2024

Captcha Image