Albuquerque Journal

Supreme Court for first time casts doubt on Section 230

Known as Big Tech’s legal shield, rule protects platforms of user content

- BY DAVID G. SAVAGE

WASHINGTON — Internet giants like Google, Facebook, YouTube and Twitter owe much of their success to a legal shield erected by Congress in 1996.

Known as Section 230, it has been called the rule that launched Big Tech. Though it drew little attention then, the law is now seen as a pillar of the wide-open global internet we know today.

While newspapers and TV stations can be held liable for any false and malicious content they publish or broadcast, internet platforms are treated differentl­y under Section 230.

Congress passed the special free-speech rule to protect the new world of online communicat­ion. It said: “No provider or user of an interactiv­e computer service shall be treated as the publisher or speaker of any informatio­n provided by another informatio­n content provider.”

Law professor and author Jeff Kosseff called Section 230 “the 26 words that created the internet” because it allowed websites to develop freely as platforms for the words, photos and videos of others.

And it went unchalleng­ed in the Supreme Court — until now.

This week, the justices will hear two cases that may finally pierce that legal shield and dramatical­ly alter the rules of the game for the internet.

And they are expected to consider a third case later this year involving the First Amendment rights of internet companies amid state efforts to regulate them.

The case to be heard on Tuesday began with a California family’s suit against Google and YouTube for allegedly aiding and abetting an act of internatio­nal terrorism. Their daughter Nohemi Gonzalez was killed in Paris in November 2015 when Islamic State terrorists fired into a restaurant where the 23-year-old student was dining with two friends. It was part of an ISIS rampage in the city that killed 129 people.

Their lawsuit alleged that Google, which owns YouTube, had “knowingly permitted ISIS to post hundreds of radicalizi­ng videos inciting violence and recruiting potential supporters to join the ISIS forces.” Further, they alleged that YouTube “affirmativ­ely recommende­d ISIS videos to users.”

At issue on Tuesday is only their second claim. Can YouTube be sued over the algorithms it created to direct users to similar content — in this case allegedly directing potential terrorists to other ISIS videos? Or does Section 230 protect them against such claims?

More than four dozen tech firms, internet scholars and free-speech advocates have filed friend-of-the-court briefs arguing that the internet companies should not be held liable for using programs that direct users to content they might find interestin­g.

“Recommenda­tion algorithms are what make it possible to find the needles in humanity’s largest haystack,” said Washington attorney Lisa S. Blatt, representi­ng Google and YouTube. She warned that opening the door to lawsuits over algorithms “risks upending the modern internet.”

A federal judge had dismissed the family’s suit based on Section 230, and a divided 9th Circuit Court of Appeals affirmed that decision in 2021.

Until this term, the Supreme Court had refused to hear appeals involving the law. On several occasions, however, Justice Clarence Thomas called for “paring back the sweeping immunity courts have read into Section 230,” particular­ly in cases where websites knew they were posting dangerous lies or criminal schemes.

Some prominent liberals, including Judges Marsha Berzon and Ronald Gould on the 9th Circuit Court, have also called for paring back the scope of Section 230.

They have been joined by advocates — both liberal and conservati­ve — who portray the internet as a cesspool of disinforma­tion and hate speech, a home for stalkers and fraudsters and a contributo­r to teen suicides and mass shootings. Critics also say social media companies get rich and keep viewers online by amplifying the most extreme claims and the angriest voices.

Google and other tech firms were surprised in October when the high court voted for the first time to hear a direct challenge to Section 230 and decide whether websites like YouTube can be sued for their use of algorithms and targeted recommenda­tions.

Their alarm grew in December when the Biden administra­tion took the side of the plaintiffs in Gonzalez vs. Google and said YouTube could be sued for algorithms that “recommend” more videos to viewers.

Justice Department attorneys said the 9th Circuit Court made a mistake by throwing out the claim, and they argued for a new understand­ing of Section 230. They agreed websites are shielded from liability for displaying content provided by others, including ISIS videos, but said they were not shielded for “their own conduct” in recommendi­ng further videos for viewing.

“When YouTube presents a user with a video she did not ask to see, it implicitly tells the user that she will be interested in that content based on the video and account informatio­n and characteri­stics,” they wrote in their filing.

Many experts in internet law said they were puzzled by the Supreme Court’s decision to take up the case and troubled by what it might mean.

“The internet needs curation. We need to be able to find what we’re looking for,” said Eric Goldman, a law professor at Santa Clara University. If websites cannot sort content based on algorithms, he said, “it would not be a functional internet.”

Blatt, Google’s attorney, said, “YouTube does not ‘recommend’ videos in the sense of endorsing them, any more than Google Search endorses search results. YouTube displays videos that may be most relevant to users.”

On Wednesday, the court will hear a related case but one focused only on whether Facebook, Google and Twitter may be sued for allegedly aiding internatio­nal terrorists.

Newspapers in English

Newspapers from United States