Communications Can Be Defamatory Even If Readers Realize There's a Considerable Risk of Error
Various commenters have suggested that AI programs’ output can’t be defamatory because reasonable readers wouldn’t view the statements as “100% reliable” or “gospel truth” or the like. Others have taken the more modest position that reasonable readers would at least recognize that there’s a significant risk of error (especially given AI programs’ disclaimers that note such a risk). And our own Orin Kerr has suggested that “no one who tries ChatGPT could think its output is factually accurate,” so I take it he’d estimate the risk of error as very high.
But, as I’ve noted before, defamation law routinely imposes liability for communicating assertions even when there is a clear indication that the assertion may well be false.
For instance, “when a person repeats a slanderous charge, even though identifying the source or indicating it is merely a rumor, this constitutes republication and has the same effect as the original publication of the slander.” When speakers identify something as rumor, they are implicitly saying “this may be inaccurate”—but that doesn’t get them off the hook.
Indeed, according to the Restatement (Second) of Torts, “the republisher of either a libel or a slander is subject to liability even though he expressly states that he does not believe the statement that he repeats to be true.” It’s even more clear that a disclaimer that the statement merely may be inaccurate can’t prevent liability.
Likewise, say that you present both an accusation and the response to the accusation. By doing that, you’re making clear that the accusation “may [be] inaccurate.”
Yet that doesn’t stop you from being liable for repeating the accusation. (There are some narrow privileges that defamation law has developed to free people to repeat certain kinds of possibly erroneous content without risk of liability, in particular contexts where such repetition is seen as especially necessary. But those privileges are needed precisely because otherwise presenting both an accusation and a response is actionable.)
And this is especially so because of what OpenAI itself notes in its GPT-4 Technical Report:
This tendency [to, among other things, produce untruthful content] can be particularly harmful as models become increasingly convincing and believable, leading to overreliance on them by users. Counterintuitively, hallucinations can become more dangerous as models become more truthful, as users build trust in the model when it provides truthful information in areas where they have some familiarity.
Couple that with OpenAI’s promotion of GPT-4′s successes in reliably performing on various benchmarks—bar exams, SATs, etc.—and it seems likely that reasonable readers will perceive GPT-4 (and especially future, even more advanced, versions) as generally fairly reliable. They wouldn’t view it as perfectly reliable, but, again, rumors are famously not perfectly reliable, yet people do sometimes act based on them, and repeating rumors can indeed lead to defamation lawsuits. They may certainly view it as more reliable than a Ouija board, a monkey on a typewriter, a fortune-teller, or the various other analogies that I’ve heard proposed (mor on those here). And one can be a reasonable reader even if one doesn’t have much understanding of how these AIs work, or even if one doesn’t have much experience with testing the AIs to see how often they err.
So, yes, when an AI program generates and communicates statements about how someone was found guilty of tax fraud, accused of harassment, and so on—and includes completely bogus quotes, though supposedly from real and prominent media outlets—there is a significant legal basis for treating those statements as defamatory, and the AI company as potentially liable for that defamation.
The post Communications Can Be Defamatory Even If Readers Realize There’s a Considerable Risk of Error appeared first on Reason.com.
Anyone can join.
Anyone can contribute.
Anyone can become informed about their world.
"United We Stand" Click Here To Create Your Personal Citizen Journalist Account Today, Be Sure To Invite Your Friends.
Please Help Support BeforeitsNews by trying our Natural Health Products below!
Order by Phone at 888-809-8385 or online at https://mitocopper.com M - F 9am to 5pm EST
Order by Phone at 866-388-7003 or online at https://www.herbanomic.com M - F 9am to 5pm EST
Order by Phone at 866-388-7003 or online at https://www.herbanomics.com M - F 9am to 5pm EST
Humic & Fulvic Trace Minerals Complex - Nature's most important supplement! Vivid Dreams again!
HNEX HydroNano EXtracellular Water - Improve immune system health and reduce inflammation.
Ultimate Clinical Potency Curcumin - Natural pain relief, reduce inflammation and so much more.
MitoCopper - Bioavailable Copper destroys pathogens and gives you more energy. (See Blood Video)
Oxy Powder - Natural Colon Cleanser! Cleans out toxic buildup with oxygen!
Nascent Iodine - Promotes detoxification, mental focus and thyroid health.
Smart Meter Cover - Reduces Smart Meter radiation by 96%! (See Video).