CNN
 — 

Supreme Court justices appeared broadly concerned Tuesday about the potential unintended consequences of allowing websites to be sued for their automatic recommendations of user content, highlighting the challenges facing attorneys who want to hold Google accountable for suggesting YouTube videos created by terrorist groups.

For nearly three hours on Tuesday, the nine justices peppered attorneys representing Google, the US government and the family of Nohemi Gonzalez, an American student killed in a 2015 ISIS attack, with questions about how the court could design a ruling that exposes harmful content recommendations to liability while still protecting innocuous ones.

How – or if – the court draws that line could have significant implications for the way websites choose to rank, display and promote content to their users as they seek to avoid a litigation minefield.

The attorney for the Gonzalez family argued that narrowing Section 230 of the Communications Decency Act – the federal law protecting websites’ right to moderate their platforms as they see fit – would not lead to sweeping consequences for the internet. But both the Court’s liberals and conservatives worried about the impact of such a decision on everything from “pilaf [recipes] from Uzbekistan” to individual users of YouTube, Twitter and other social media platforms.

A big concern of the justices seems to be the waves of lawsuits that could happen if the court rules against Google.

“Lawsuits will be nonstop,” Justice Brett Kavanaugh said at one point.

But Eric Schnapper, representing the plaintiffs, argued that a ruling for Gonzalez would not have far-reaching effects because even if websites could face new liability as a result of the ruling, most suits would likely be thrown out anyway.

“The implications are limited,” Schnapper said, “because the kinds of circumstance in which a recommendation would be actionable are limited.”

Later, Justice Elena Kagan warned that narrowing Section 230 could lead to a wave of lawsuits, even if many of them would eventually be thrown out, in a line of questioning with US Deputy Solicitor General Malcolm Stewart.

“You are creating a world of lawsuits,” Kagan said. “Really, anytime you have content, you also have these presentational and prioritization choices that can be subject to suit.”

Chief Justice John Roberts mused that under a narrowed version of Section 230, terrorism-related cases might only be a small share of a much wider range of future lawsuits against websites alleging antitrust violations, discrimination, defamation and infliction of emotional distress, just to name a few.

“I wouldn’t necessarily agree with ‘there would be lots of lawsuits’ simply because there are a lot of things to sue about,” Stewart said, “but they would not be suits that have much likelihood of prevailing, especially if the court makes clear that even after there’s a recommendation, the website still can’t be treated as the publisher or speaker of the underlying third party.”

Multiple justices pushed Schnapper to clarify how the court should treat recommendation algorithms if the same algorithm that promotes an ISIS video to someone interested in terrorism might be just as likely to recommend a pilaf recipe to someone interested in cooking.

“I’m trying to get you to explain to us how something that is standard on YouTube for virtually anything you have an interest in, suddenly amounts to aiding and abetting [terrorism] because you’re [viewing] in the ISIS category,” Justice Clarence Thomas said.

Schnapper attempted several explanations, including at one point digressing into a hypothetical about the difference between YouTube videos and video thumbnail images, but many of the justices were lost about what he was calling for.

“I admit I’m completely confused by whatever argument you’re making at the present time,” Justice Samuel Alito said.

Roberts added: “It may be significant if the algorithm is the same across … the different subject matters, because then they don’t have a focused algorithm with respect to terrorist activities… Then it might be harder for you to say that there’s selection involved for which you can be held responsible.”

One of the few justices focusing on how changes to Section 230 could affect individual internet users was Justice Amy Coney Barrett, who repeatedly asked whether narrowing the law in the ways Schnapper has proposed could put average social media users in legal jeopardy.

The text of Section 230 explicitly immunizes “users,” and not just social media platforms, from liability for the content posted by third parties. So a change that exposes tech platforms to new lawsuits could also have implications for users, according to several amicus briefs.

Under Schnapper’s interpretation, could liking, retweeting or saying “check this out” expose individuals to lawsuits that they could not deflect by invoking Section 230?

Yes, Schnapper acknowledged, because “that’s content you’ve created.”

Barrett raised the issue again in a question for Justice Department lawyer Stewart. She asked: “So the logic of your position, I think, is that retweets or likes or ‘check this out’ for users, the logic of your position would be that 230 would not protect in that situation either. Correct?”

Stewart said there was distinction between an individual user making a conscious decision to amplify content and an algorithm that is making choices on a systemic basis. But Stewart did not provide a clear answer about how he believed changes to Section 230 could affect individual users.

Tech law experts say an onslaught of defamation litigation is the real threat if Section 230’s protections are weakened and the justices seemed to agree, posing several questions and hypothetical that turned on defamation claims.

“People have focused on the [Antiterrorism Act], because that’s the one point that’s at issue here. But I suspect there will be many, many times more defamation suits,” Chief Justice John Roberts said, while pointing to other types of claims that also may flood the legal system if tech companies no longer had broad Section 230 immunity.

Justice Samuel Alito posed for Schnapper a scenario where a competitor of a restaurant created a video making false claims about the restaurant violating health code and YouTube refusing to take the video down despite knowing its defamatory.

Kagan seized on Alito’s hypothetical later on in the hearing, asking what happens if a platform recommended the false restaurant competitor’s video and called it the greatest video of all time, but didn’t repeat anything about the content of the video.

“Is the provider on the hook for that defamation?” Kagan asked.

This story and headline have been updated with developments from Tuesday’s hearing.

Source link

You May Also Like

The Who’s Who Behind the Modern Artificial Intelligence Movement

While artificial intelligence has taken the limelight over the past year, technology…

New York AG accuses crypto firms of deceiving investors in $1 billion fraud | CNN Business

CNN  —  The fallout from the colossal implosion of Sam Bankman-Fried’s crypto…

International crackdown on criminal marketplaces selling stolen data

One of the most important tips I have focused on in recent…

To benefit all, diverse voices must take part in leading the growth and regulation of AI | TechCrunch

Jorge Calderon Contributor Jorge Calderon is managing director at San Francisco–based Inicio…