ChatGPT cited “bogus” cases for a New York federal court filing. The attorneys involved may face sanctions.

The Thurgood Marshall courthouse is pictured in Manhattan in New York, October 15, 2021.

Brendan McDermid | Reuters

Roberto Mata’s lawsuit against Avianca Airlines wasn’t so different from many other personal-injury suits filed in New York federal court. Mata and his attorney, Peter LoDuca, alleged that Avianca caused Mata personal injuries when he was “struck by a metal serving cart” on board a 2019 flight bound for New York.

Avianca moved to dismiss the case. Mata’s lawyers predictably opposed the motion and cited a variety of legal decisions, as is typical in courtroom spats. Then everything fell apart.

Avianca’s attorneys told the court that it couldn’t find numerous legal cases that LoDuca had cited in his response. Federal Judge P. Kevin Castel demanded that LoDuca provide copies of nine judicial decisions that were apparently used.

In response, LoDuca filed the full text of eight cases in federal court. But the problem only deepened, Castel said in a filing, because the texts were fictitious, citing what appeared to be “bogus judicial decisions with bogus quotes and bogus internal citations.”

The culprit, it would ultimately emerge, was ChatGPT. OpenAI’s popular chatbot had “hallucinated” — a term for when artificial intelligence systems simply invent false information — and spat out cases and arguments that were entirely fiction. It appeared that LoDuca and another attorney, Steven Schwartz, had used ChatGPT to generate the motions and the subsequent legal text.

Schwartz, an associate at the law firm of Levidow, Levidow & Oberman, told the court he had been the one tooling around on ChatGPT, and that LoDuca had “no role in performing the research in question,” nor “any knowledge of how said research was conducted.”

Opposing counsel and the judge had first realized that the cases didn’t exist, providing the involved attorneys an opportunity to admit to the error.

LoDuca and his firm, though, seemed to double down on the use of ChatGPT, using it not just for the initially problematic filing but to generate false legal decisions when asked to provide them. Now, LoDuca and Schwartz may be facing judicial sanction, a move that could even lead to disbarment.

The motion from the defense was “replete with citations to non-existent cases,” according to a court filing.

“The Court is presented with an unprecedented circumstance,” Castel said. He set a hearing for June 8 hearing when both LoDuca and Schwartz will be called to explain themselves. Neither attorney responded to CNBC’s request for comment.

For all the latest Technology News Click Here 

 For the latest news and updates, follow us on Google News

Read original article here

Denial of responsibility! TheDailyCheck is an automatic aggregator around the global media. All the content are available free on Internet. We have just arranged it in one platform for educational purpose only. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, all materials to their authors. If you are the owner of the content and do not want us to publish your materials on our website, please contact us by email – [email protected] The content will be deleted within 24 hours.