Court fines US lawyers who cited fake cases produced by ChatGPT


ChatGPT: System produced six fake cases

Two lawyers who unwittingly submitted fake cases generated by ChatGPT to support their claim have been fined by a New York court because they “abandoned their responsibilities”.

As well as being fined $5,000 each, the lawyers and their firm have been ordered to inform their client and the judges whose names were wrongfully invoked in the case of the sanctions imposed on them.

Peter LoDuca, Steven A Schwartz and the firm of Levidow Levidow & Oberman attracted international attention after the brief in a personal injury claim prepared by Mr Schwartz contained six cases that ChatGPT had simply made up – the system later insisted they were real when Mr Schwartz asked it.

US District Judge P Kevin Castell in the Southern District of New York said: “In researching and drafting court submissions, good lawyers appropriately obtain assistance from junior lawyers, law students, contract lawyers, legal encyclopedias and databases such as Westlaw and LexisNexis.

“Technological advances are commonplace and there is nothing inherently improper about using a reliable artificial intelligence tool for assistance. But existing rules impose a gatekeeping role on attorneys to ensure the accuracy of their filings.”

He continued that here the lawyers “abandoned their responsibilities when they submitted non-existent judicial opinions with fake quotes and citations created by the artificial intelligence tool ChatGPT, then continued to stand by the fake opinions after judicial orders called their existence into question”.

Among the “many harms” that flowed from the submission of fake opinions was that it promoted “cynicism about the legal profession and the American judicial system”, while “a future litigant may be tempted to defy a judicial ruling by disingenuously claiming doubt about its authenticity”.

The law firm primarily practises in New York state courts. It uses a legal research service called Fastcase and does not have access to Westlaw or LexisNexis.

But the case involved the Montreal Convention and was in federal court, and the firm’s Fastcase account had limited access to federal cases. Mr Schwartz said this was why he turned to ChatGPT.

Judge Castell said the outcome of the matter would have been “quite different” had the lawyers come clean after the defendant first questioned the existence of the cases, or after the court had required them to produce them.

Instead, they “doubled down and did not begin to dribble out the truth” until the court issued an order to show cause why they ought not be sanctioned.

This was evidence of bad faith on their parts, as was Mr Schwartz’s statement to the court that ChatGPT had “supplemented” his research, when in fact it was the only source of his substantive arguments.

Mr Schwartz testified at the sanctions hearing that he was “operating under the false perception” that ChatGPT “could not possibly be fabricating cases on its own”.

He said: “My reaction was, ChatGPT is finding that case somewhere. Maybe it’s unpublished. Maybe it was appealed. Maybe access is difficult to get. I just never thought it could be made up.”

The law firm told the court that it has arranged for outside counsel to conduct mandatory training on technological competence and artificial intelligence.

Judge Castell credited “the sincerity of the respondents when they described their embarrassment and remorse”. The fines were “sufficient but not more than necessary to advance the goals of specific and general deterrence”.

Earlier this month, the Master of the Rolls, Sir Geoffrey Vos, cited the case as a reason why legal regulators and the courts may need to control “whether and in what circumstances and for what purposes” lawyers can use systems like ChatGPT in litigation.




Leave a Comment

By clicking Submit you consent to Legal Futures storing your personal data and confirm you have read our Privacy Policy and section 5 of our Terms & Conditions which deals with user-generated content. All comments will be moderated before posting.

Required fields are marked *
Email address will not be published.

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Blog


Five key issues to consider when adopting an AI-based legal tech

As generative AI starts to play a bigger role in our working lives, there are some key issues that your law firm needs to consider when adopting an AI-based legal tech.


Bulk litigation – not always working in consumers interests

For consumers to get the benefit, bulk litigation needs to be done well, and we are increasingly concerned that there are significant problems in some areas of this market.


ABSs, cost and audits – fixing regulation after Axiom Ince

A feature of law firm collapses and frauds has sometimes been the over-concentration of power in outdated and overburdened systems of control.


Loading animation