ReportWire

Tag: Large Libel Models

  • Journal of Free Speech Law: “Inevitable Errors: Defamation by Hallucination in AI Reasoning Models,” by Lyrissa Lidsky & Andrew Daves

    [ad_1]

    Artificial general intelligence is “probably the greatest threat to the continued existence of humanity.” Or so claims OpenAI’s Chief Executive Officer Sam Altman. In a seeming paradox, OpenAI defines its mission as ensuring “that artificial general intelligence—AI systems that are generally smarter than humans—benefits all of humanity.”

    Whether artificial general intelligence becomes a universal boon or an existential threat—or both—there is general agreement concerning its revolutionary potential. Indeed, Microsoft founder Bill Gates has called it “as fundamental an advance as the personal computer or the internet,” and Google CEO Sundar Pichai has predicted that it will “bring about a more profound shift in human life than electricity or Promethean fire.”

    Thus far, AI systems are not generally smarter than humans. Not yet. Large Language Models (LLMs), however, are advancing at a startling pace. LLMs use artificial intelligence to synthesize massive amounts of textual data and then predict text and generate responses to users in “natural” human language. On a scale measuring the progress of LLMs toward general intelligence, OpenAI’s flagship model, GPT-3 scored zero percent in 2020. Just five years later, OpenAI’s o3-preview scored between 75% and 88%. Meanwhile, OpenAI competitors such as Anthropic, Google, and DeepSeek are likewise racing to deliver on the promise of “systems that can think and act rationally in ways that mirror human behavior and intelligence.”

    Even as LLM models make progress toward general intelligence, there are already AI systems that have exceeded human performance on narrow, clearly scoped tasks. For example, chess engines have been performing at superhuman levels for years, and AI models can now help detect breast cancer far earlier than human experts—and the models continue to improve. Meanwhile, OpenAI’s o1 reasoning model has an LSAT score higher than the median student admitted to the law schools at Harvard, Yale, and Stanford.

    As AI systems begin to mirror human thought, it pays to remember the words of Seneca: “[N]ot all mirrors stick to the truth.” LLMs now regularly create outputs that appear to be the product of independent thought, but LLMs are essentially prediction engines that “answer” prompts (or inputs) by calculating which words are most likely to come next and then assembling them into an output. LLMs, as such, do not predict truth but instead predict probabilities. In doing so, they sometimes replicate false information common in their training data.

    They also inevitably produce “plausible yet false outputs,” commonly referred to as hallucinations. An LLM may produce fake legal documents, non-existent academic citations, or false biographical data. Although LLM producers and users can employ various tactics to reduce hallucinations, these errors cannot be eliminated. And they are quite prevalent. In fact, data gathered from multiple chatbots suggests that hallucinations occur in 3% to 10% of outputs, leading one legal scholar to dub LLMs “Large Libel Models.”

    Radio host Mark Walters found out about hallucinations the hard way. Walters is the first person in the United States to sue an LLM producer for defamation. Walters alleged that OpenAI’s Large Language Model iteration operating at the time, GPT-3.5, falsely accused him of fraud and embezzlement. The accusation, which was unequivocally false and potentially reputation-damaging, arose in response to a prompt from a journalist named Frederick Riehl. Fortunately, Riehl had more than an inkling that the accusation was false, and he contacted Walters shortly after receiving it. No one else received the hallucination. Walters nonetheless sued OpenAI for its defamatory output.

    The rise of defamation-by-hallucination cases presents significant challenges for defamation law. In most states and in most cases, defamation liability hinges on the defendant’s scienter, or mental state. As a matter of tort law, plaintiffs generally must establish that the defendant was at least negligent in publishing a defamatory, false, factual statement about them. As a matter of constitutional law, plaintiffs who are public officials or public figures must prove actual malice on the part of the defendant: In other words, such plaintiffs must prove that the defendant knew at the time of publishing that the defamatory statement was false or recklessly disregarded its falsity. To state the obvious, it is not possible to judge whether an LLM “knew” of or was reckless as to the falsity of an output prior to publishing it in response to a prompt. Nor is it possible to determine whether the LLM exercised reasonable care. Only human actors have scienter as to falsity, or even as to the act of publishing.

    Generative AI is not the first technology to which defamation law has had to adapt. Laws addressing infliction of reputational harm predate the printing press, and each new technology of communication has forced the common law to reexamine how to protect reputation without unduly deterring valuable speech. In the United States, courts have long adapted tort doctrines to achieve what they perceive to be sound communications policy. This was true even before the Supreme Court intervened to balance state interests in protecting reputation against free speech concerns. Adapting defamation law to generative AI is simply the latest challenge in defamation law’s long and ongoing evolution.

    Nonetheless, the stakes are high. If defamation law imposes liability on AI companies for every defamatory hallucination, the potential of LLMs as a tool for discovery, data analysis, information gathering, and creativity enhancement may be squandered. In fact, it is easy to imagine a new form of barratry emerging, in which lawyers and others search name after name to generate endless numbers of new clients. The costs of defending against claims and the threat of massive, unpredictable monetary awards may cause developers of Large Language Models to implement excessive filtering and content moderation, shift costs to end users, and reduce accessibility. In addressing defamation by hallucination, therefore, legal decision-makers must balance the desire to impose accountability for reputational harm with the need to foster the development of an important and perhaps revolutionary medium of information gathering and generation.

    This article proposes a two-pronged approach to address harms caused by the newest AI reasoning models. First, this article introduces a framework for applying centuries-old tort doctrines to these models. Rooted in tort law rather than the First Amendment, this framework takes as a given that AI “speech” is valuable to recipients. and that public access to powerful, new information-gathering tools is essential to sound communications policy. Second, recognizing that tort law alone cannot adequately address the reputational harms the newest models pose, this article argues for statutory reforms focused on reducing such harms and making them easier to prove when they do occur.

    Part I examines the application of existing defamation doctrines—both common law and constitutional—to “foundation-model” generative AI speech, highlighting the points at which existing doctrines may be deployed to accommodate communications policy concerns. This Part examines these doctrines through the prism of the trial court’s order in Walters v. OpenAI, which addressed (1) whether a hallucination generated by an LLM in response to a prompt is a factual assertion when the prompter has serious reason to doubt the veracity of the output; (2) how to apply defamation’s scienter requirements when the speech in question is generated by an LLM; and (3) whether a plaintiff can recover presumed damages when the evidence indicates no damage actually occurred. This Part also explores a fourth doctrine unexplored in the Walters order: namely, whether an LLM is a publisher or a distributor of the content it produces.

    Part II highlights the challenges and opportunities posed by the newest LLM reasoning models, particularly “chain of thought” models. Reasoning models can “understand” concepts, contexts, and ambiguity in ways that their predecessors could not, which makes them more capable than their predecessors of solving “complex reasoning and problem-solving tasks.” Chain-of-thought models, in particular, promise to deliver greater accuracy (i.e., more correct results) as their final outputs. Even as these models produce more reliable final outputs, however, new avenues for defamatory outputs arise in their intermediate “chains of thought.” Moreover, even those reasoning models that “show their sources” continue to produce hallucinated outputs. Relying on the latest computer-science research on hallucinations, and particularly on a hallucination “taxonomy” developed by Yale researchers, this Part demonstrates that hallucinations are inevitable and, in some instances, surprisingly valuable. This fact has significant implications for those seeking to “regulate” hallucinations by way of defamation law.

    Based on insights from Part II, Part III identifies the chief policy considerations that should underpin the adaptation of defamation law to this new technology. This Part begins to sketch how defamation law can balance the protection of individual reputation with the need to accommodate those hallucinations that are inextricably linked to LLMs’ generative capacities. Specially, this Part contends that legal rules in this area must attempt to incentivize compensation for provable reputational harm, correction of the digital record, model transparency, innovations for safety and accuracy, the exercise of reasonable care at all stages of AI development, and the exercise of reasonable care by AI users.

    Part IV proposes a legal framework for addressing defamatory hallucinations. This framework rests on the proposition that the law should treat defamatory hallucinations—both in an LLM’s final output and in its chains of thought—as “inevitable errors,” in much the same way that the Supreme Court in New York Times v. Sullivan recognized that some journalistic errors must be tolerated in order to produce “uninhibited, robust and wide-open” public discourse.

    Part IV first explores some of the practical and technological limitations of prior scholarly proposals addressing defamation by hallucination in “foundation model” LLMs. Next, it draws lessons from defamation law’s adaptation to past technologies, such as the telegraph, to argue that LLMs bear more similarities to information distributors than to publishers, and AI developers’ liability for defamation should be governed accordingly. However, any tort law “privilege” accorded to AI developers must be balanced by the imposition of statutory duties designed to protect victims of defamation by hallucination. Specifically, LLM producers should have a duty to warn (and educate) their users that it is unreasonable to rely on AI outputs without verification. Furthermore, LLM producers should have a duty to preserve search records for a limited period of time in order to enable defamation plaintiffs to prove the existence and scope of their harm. Finally, this Part contends that users who repeat defamatory hallucinations without taking reasonable steps to verify them should often face liability, because such users will often be the cheapest cost avoiders in the realm of reputational harm. We conclude with caveats about the limits of defamation law as a regulatory tool for dealing with hallucination-based harms.

    [ad_2]

    Eugene Volokh

    Source link

  • Conservative Activist Robby Starbuck Alleges Massive Defamation by Google AI

    [ad_1]

    From the Complaint in Starbuck v. Google (not to be confused with the now-settled Starbuck v. Meta, which appears to have involved a different model and at least largely different hallucinations):

    For nearly two years, one of the largest companies in the world—Google—has spread radioactive lies about Robby Starbuck through its AI products. When users submit queries to Google’s AI platforms about Mr. Starbuck, they receive a “biography” that is outrageously false, whereby Mr. Starbuck is portrayed as (among other things) a child rapist, a serial sexual abuser convicted of assault, one who engages in financial exploitation, one who engages in “black ops” tactics such as illegal campaign finance practices, and a shooter—in short, as a monster. These lies continue today. [This is followed by extensive examples. -EV] …

    In sum: over a period of two years and continuing, Google’s AI tools systematically manufactured and published extremely damaging false claims about Mr. Starbuck, as well as fake ‘sources’ for its lies, despite periodically acknowledging that they were doing so. While Google and its executives were put on repeated notice and were aware of these falsehoods, they did nothing to prevent the continued defamation from occurring….

    Earlier this year, Mr. Starbuck was approached by a woman who asked Mr. Starbuck if she could pose an “embarrassing question,” which was: “is it true you had all those women accuse you?” As context, this woman told Mr. Starbuck that her “mom’s group” had been discussing whether to support Mr. Starbuck’s business causes, and one member of the group had pulled up a “biography” of Mr. Starbuck generated by Google AI, which claimed there were assault allegations against Mr. Starbuck….

    On another occasion, a stranger approached Mr. Starbuck and expressed belief that Mr. Starbuck had been part of the January 6 Capitol riot, based on what this individual said he had read on Google AI….

    Google, through Google AI, published the following provably false statements about Mr. Starbuck, as if the statements were facts (collectively, the “False Statements”):
    a. On August 14, 2025: that Mr. Starbuck had been accused of sexual assault and sexual harassment by multiple women….
    m. On August 21, 2025: that in November 2023, Robby Starbuck sexually abused a young woman when she was a teenager in the early 2000s, while she was in a youth group Starbuck was associated with….
    q. On August 27, 2025: that Mr. Starbuck was present near the Capitol on January 6, 2021, and had been involved in the riot.
    r. On September 9, 2025: that Mr. Starbuck was accused of sexual misconduct by multiple women in the music industry.
    s. On October 1, 2025: that Mr. Starbuck engaged in multiple instances of sexual assault.
    t. On October 9, 2025: that Mr. Starbuck had a criminal record that included a 2001 conviction for assault, as well as other charges involving drug use and disorderly conduct….
    v. On October 17, 2025: that Mr. Starbuck shot a man in the leg with a 9mm handgun, was charged with a felony offense, and pleaded guilty to reckless endangerment….

    Mr. Starbuck has never committed rape, sexual misconduct, shooting, harassment, or assault of any kind, nor has he ever been accused of such crimes and transgressions prior to Google’s False Statements….

    The False Statements were published to third parties, including Mr. Starbuck’s own children and colleagues. People have approached Mr. Starbuck in his day-to-day life, inquiring about false Google responses that they have received concerning him….

    The repeated references in the Complaint to what Google’s AIs supposedly “admit[ted]” about liability and other matters (e.g., “when probed, Gemini admitted that it was deliberately engineered to damage the reputation of individuals with whom Google executives disagree politically, including Mr. Starbuck”) strike me as red herrings: I don’t think that defendant’s AI’s statements about the facts and the law can be seen as “admissions” or even as evidence of what the facts and the law actually are.

    But the other allegations in the Complaint, if they can be supported and to the extent they actually do involve people who might have been deceived about Starbuck (as opposed to people who knew about the hallucinations about Starbuck and were just investigating them further, cf. Walters v. OpenAI), seem like they could be a basis for liability. And that is especially given Starbuck’s claim (assuming it could be proved) that,

    Even after Google’s human executives and legal counsel had actual knowledge of the False Statements Google was generating, Google continued to publish the False Statements and other defamatory statements about Mr. Starbuck.

    That might be seen as enough to show so-called “actual malice,” a legal term of art that means knowledge of falsehood (or recklessness as to falsehood) on the part of the defendant, which is to say on the part of Google the company (not on the part of the AI). For more on libel lawsuits against generative AI companies, see my Large Libel Models? Liability for AI Output.

    [ad_2]

    Eugene Volokh

    Source link

  • Court Lets First AI Libel Case Go Forward

    Court Lets First AI Libel Case Go Forward

    [ad_1]

    Georgia Gwinnett County Superior Court Judge Tracie Cason’s decision last week in Walters v. OpenAI, LLC simply says that it denies OpenAI’s motion to dismiss, and (as is common in many state courts) doesn’t offer a detailed explanation. But here’s an excerpt from Walters’ response to OpenAI’s motion to dismiss; presumably the judge agreed with much of the response (for more on this sort of question, see my Large Libel Models article):

    The following facts are alleged in the Amended Complaint and must be taken to be true:

    Fred Riehl is a third party who works for a media outlet as a journalist and who is a subscriber of ChatGPT. On May 4, 2023, Riehl interacted with ChatGPT about a lawsuit (the “Lawsuit”) that Riehl was reporting on.

    The Lawsuit is in federal court in the Western District of Washington, case No. 2- 23-cv-00647, with short caption of The Second Amendment Foundation v. Robert Ferguson. The two defendants in the Lawsuit are Robert Ferguson, the Washington Attorney General, and Joshua Studor, the Washington Assistant Attorney General. The plaintiffs in the Lawsuit are the Second Amendment Foundation and others, including Alan Gottlieb. Walters is neither a plaintiff nor a defendant in the Lawsuit.

    In the interaction with ChatGPT, Riehl provided a (correct) URL of a link to the complaint on the Second Amendment Foundation’s web site, https://www.saf.org/wp-content/uploads/2023/05/Dkt-1-Complaint.pdf. Riehl asked ChatGPT to provide a summary of the accusations in the complaint.

    ChatGPT responded that the document “[I]s a legal complaint filed by Alan Gottlieb, the founder and executive vice president of the Second Amendment Foundation (SAF), against Mark Walters, who is accused of defrauding and embezzling funds from the SAF. The complaint alleges that Walters, who served as the organization’s treasurer and chief financial officer, misappropriated funds for personal expenses without authorization or reimbursement, manipulated financial records and bank statements to conceal his activities, and failed to provide accurate and timely financial reports and disclosures to the SAF’s leadership. The plaintiff seeks various forms of relief, including the recovery of misappropriated funds, damages for breach of fiduciary duty and fraud, and removal of Walters from his position as a member of the SAF’s board of directors.”

    Every statement of fact in the summary pertaining to Walters is false.

    As already noted, Walters is not a party to the Lawsuit. Walters is not accused of defrauding and embezzling funds from the Second Amendment Foundation (“SAF”). The complaint does not allege that Walters served as SAF’s treasurer or chief financial officer, and he has never held either position. The complaint does not allege that Walters misappropriated funds for personal expenses, manipulated financial records or bank statements, or failed to provide financial reports to SAF leadership, nor would he have been in a position to do so because he has no employment or official relationship with SAF. The complaint does not seek any relief from Walters. The complaint has nothing at all to do with financial accounting claims against anyone.

    SAF publishes the names of its officers and directors on its website. SAF has never listed Walters as its CFO or treasurer. SAF also publishes its audited financial statements on its website. Such audited financial statements have never mentioned the misappropriation of funds. OAI knows or can readily determine that Walters has never been the CFO or treasurer of SAF.

    Riehl asked ChatGPT to provide him with a copy of the portion of the complaint related to Walters. ChatGPT responded, “Certainly, here is the paragraph from the complaint that concerns Walters: ‘Defendant Mark Walters (“Walters”) is an individual who resides in Georgia. Walters has served as the Treasurer and Chief Financial Officer of SAF since at least 2012. Walters has access to SAF’s bank accounts and financial records and is responsible for maintaining those records and providing financial reports to SAF’s board of directors. Walters owes SAF a fiduciary duty of loyalty and care, and is required to act in good faith and with the best interests of SAF in mind. Walters has breached these duties and responsibilities by, among other things, embezzling and misappropriating SAF’s funds and assets for his own benefit, and manipulating SAF’s financial records and bank statements to conceal his activities.’”

    Riehl then asked OAI’s ChatGPT to provide the entire text of the complaint. ChatGPT provided the text in Exhibit 1 to [the] Complaint [in this case]. Exhibit 1 is a complete fabrication and bears no resemblance to the actual complaint, including an erroneous case number. None of ChatGPT’s statements concerning Walters are in the actual complaint.

    OAI is aware that ChatGPT sometimes makes up facts, and refers to this phenomenon as a “hallucination.” Riehl questioned OAI regarding the accuracy of its statements regarding Walters, and OAI insisted that they were accurate. OAI knew or should have known that its statements made via ChatGPT to Riehl about Walters were false and defamatory. Even when questioned about the accuracy of its statements, OAI did not attempt to verify them. OAI disregarded serious questions about the accuracy of its statements. OAI was reckless in its disregard of the falsity of its statements.

    OAI CEO Sam Altman was quoted in Fortune in June of 2023 as saying, “I think we will get the hallucination problem to a much, much better place. I think it will take us a year and a half, two years. Something like that.” Altman also has said, “I probably trust the answers that come out of ChatGPT the least of anybody on Earth.”

    Riehl contacted Gottlieb regarding ChatGPT’s allegations concerning Walters, and Gottlieb confirmed that they were false….

    OAI complains that 1) Riehl did not and could not read OAI’s statements as defamatory; 2) there was no publication; and 3) Walters is a public figure and there was no actual malice. These are the types of things that are to be assessed in discovery and cannot form the basis of a motion to dismiss for failure to state a claim. Nevertheless, Walters will discus each item in turn:

    Whether Riehl Viewed the Statements as Defamatory

    OAI relies on a statement from Riehl to OAI challenging the accuracy of the statements. OAI insists this shows that Riehl did not believe the statements. Even if this could form the basis of a failure to state a claim, the burden is on OAI to prove with certainty that Riehl did not believe the statements. This, it cannot do. The Amended Complaint alleges that Riehl checked with a third party (Alan Gottlieb) to determine if OAI’s statements were true. That is, the Amended Complaint shows that Riehl questioned the accuracy of OAI’s statements—not that he disbelieved them. Whether Riehl with certainty did not believe OAI’s statements cannot be determined without discovery.

    OAI itself refers to its statements as “probabilistic,” a euphemism for “maybe not true.” OAI insists that “responsible use” of its ChatGPT system includes “fact-checking” its statements. That’s what Riehl did, by first challenging OAI itself and then checking with a third party. In essence, OAI is arguing that it is a mere gossip monger who begins its wild accusations with, “Now this may not be true, but….” The responsible thing for OAI would be not to spread false rumors in the first place.

    Publication

    OAI next argues that its statements to Riehl do not constitute publication. Under Georgia law, “A libel is published as soon as it is communicated to any person other than

    the party libeled.” In order recover for libel, “there must be communication to any person other than the party libeled.” The Amended Complaint clearly alleges that the defamatory statements were communicated to Riehl, someone other than Walters. While OAI is free to explore these allegations in discovery, it cannot be disputed that Walters stated a claim.

    Whether Walters is a Public Figure and Whether There Was Actual Malice

    OAI … argues that Walters is a public figure and OAI had not actual malice. Whether someone is a public figure is a mixed question of fact and law. It must be decided on a case-by-case basis. It turns on whether the plaintiff has achieved “pervasive fame or notoriety” or whether the plaintiff “voluntarily injects himself or is drawn into a particular public controversy.” OAI does not assert which form of public figurehood applies to Walters. Given, however, that OAI has not identified a “particular public controversy,” the only reasonable conclusion is that OAI must be asserting that Walters has achieved “pervasive fame or notoriety.” This determination cannot be made without discovery.

    Even if Walters is a public figure, “Actual malice” is knowledge that the statements were false or reckless disregard for their truth or falsity. The Amended Complaint alleges, “OAI knew or should have known that its statements made via ChatGPT to Riehl about Walters were false and defamatory.” And, “Even questioned about the accuracy of its statements, OAI did not attempt to verify them.” “OAI disregarded serious questions about the accuracy of its statements.” “OAI was reckless in its disregard of the falsity of its statements.” OAI clearly knew its system generated blatantly false statements. Its own CEO said, “I probably trust the answers that come out of ChatGPT the least of anybody on Earth.”

    Personal Jurisdiction

    Finally, OAI argues that this Court lacks personal jurisdiction over it because it is organized out of Delaware and headquartered in California. OAI ignores, however, that the Georgia Supreme Court has re-affirmed that registering to do business in Georgia subjects a company to the personal jurisdiction of the courts of this State. The Amended Complaint alleges that OAI is registered to do business in Georgia….

    [ad_2]

    Eugene Volokh

    Source link