Michael Cohen mistakenly referenced fake court cases created by AI


Are you ready to delve into the scandalous world of AI-generated legal documents? If you thought artificial intelligence was only for sci-fi movies, think again. This blog post will take you on a wild ride through the shocking revelation of Michael Cohen, former lawyer for Donald Trump, using fake, AI-generated court cases in a legal document. Buckle up and get ready for a thrilling journey into the intersection of technology and law.

The AI Deception: Unraveling Michael Cohen’s Legal Scandal
In a jaw-dropping turn of events, it has come to light that Michael Cohen, the once-trusted lawyer of Donald Trump, admitted to citing fake, AI-generated court cases in a legal document that ended up before a federal judge. This revelation has sent shockwaves through the legal community, raising serious questions about the reliability of AI-generated content in the legal field.

The Misunderstood Tool: Google Bard’s Role in Cohen’s Deception
According to a filing unsealed on Friday, Cohen used Google’s Bard to perform research, mistakenly believing it to be a “super-charged search engine” rather than an AI chatbot. This misunderstanding led to the inclusion of non-existent cases in a motion asking a federal judge to shorten Cohen’s probation. The use of AI in legal research has never been more controversial, as the implications of this revelation continue to unfold.

The Judge’s Verdict: Exposing the Fictional Cases
US District Judge Jesse Furman was quick to spot the deception, pointing out that “none of these cases exist” in a filing after reviewing the letter brief submitted by Cohen. This prompted the judge to question Cohen’s lawyer, David Schwartz, about the inclusion of fake citations in the motion, as well as whether Cohen himself played a role in drafting the document. The legal fallout from this scandal is still ongoing, with potential sanctions looming for those involved.

The Unintended Deception: Cohen’s Defense and Explanation
In response to the accusations, Cohen submitted a written statement claiming that he did not intend to mislead the court. He stated that he used Google Bard for legal research and shared his findings with his lawyer, Schwartz. Cohen insists that he was unaware of the potential for fake citations and did not expect them to be included in the motion without verification. This defense raises important questions about the responsibility of legal professionals when using AI-generated content.

The Unveiling of AI’s Dark Side: Cohen’s Ignorance of Legal Technology
Cohen’s written statement also revealed his lack of awareness about the risks associated with emerging legal technology, particularly in the realm of AI-generated content. He admitted, “As a non-lawyer, I have not kept up with emerging trends (and related risks) in legal technology and did not know that Google Bard was a generative text service that could show citations and descriptions that looked real but actually were not.” This ignorance sheds light on the potential pitfalls of using advanced technology without a full understanding of its capabilities.

As we witness the shocking aftermath of Michael Cohen’s use of fake, AI-generated court cases, it’s clear that the intersection of technology and law is more complex and fraught with challenges than ever before. This scandal serves as a cautionary tale for legal professionals and technology users alike, highlighting the need for a deeper understanding of the risks and implications of AI in the legal field. So, buckle up and get ready for a thrilling journey into the dark side of AI in the legal world.

Leave a comment

Your email address will not be published. Required fields are marked *