Los Angeles Times (Sunday)

Supreme Court holds the internet’s fate in its hands

- MICHAEL HILTZIK Hiltzik writes a blog on latimes.com. Follow him on Facebook or on Twitter @hiltzikm or email michael .hiltzik@latimes.com.

Almost no one noticed in

1996 when Congress gave online social media platforms sweeping legal immunity from what their users posted on them.

The provision, crafted by then-Rep. Christophe­r Cox (R-Newport Beach) and then-Rep. Ron Wyden (DOre.), was known as Section 230 of the Communicat­ions Decency Act. It has since become labeled as the “Magna Carta of the internet” and “the 26 words that created the internet.”

Without Section 230, according to Jeff Kosseff, the law professor whose book on the section bears the latter title, the social media world as we know it today “simply could not exist.”

That’s why advocates of online speech — indeed, of internet communicat­ions generally — are very, very nervous that the Supreme Court has taken up a case that could determine Section 230’s limits or even, in an extreme eventualit­y, its constituti­onality.

The Supreme Court’s decision to review two lower court rulings, including an appellate case from the U.S. 9th Circuit Court of Appeals in San Francisco, marks the first time the court has chosen to review Section 230, after years in which it consistent­ly turned away cases involving the law.

That may not reflect a change in its view of the legal issues so much as a change in how society views the internet platforms at the center of the cases — Google, Facebook, Twitter and other sites that allow users to post their own content with minimal review.

“We’ve been in the midst of a multiyear tech-lash, representi­ng the widely held view that the internet has gone wrong,” says Eric Goldman, an expert in high-tech and privacy law at Santa Clara University Law School. “The Supreme Court is not immune to that level of popular opinion — they’re people too.”

Disgruntle­ment with the big tech platforms stretches from one side of the political spectrum to the other.

Conservati­ves cherish the notion that the platforms are liberal fronts that have been hiding behind their content-moderation policies to disproport­ionately block conservati­ve users and suppress conservati­ve viewpoints; progressiv­es complain that the platforms’ policies haven’t been successful in eradicatin­g harmful content, including disinforma­tion and racism and other hate speech.

The harvest has been laws and legislativ­e proposals aiming to dictate how the platforms moderate content.

Florida enacted a law prohibitin­g social media firms from shutting down politician­s’ accounts based on proponents’ assertions that “big tech oligarchs in Silicon Valley” aim to silence conservati­ves to favor a “radical leftist agenda,” as a federal appeals court observed in a decision overturnin­g the law.

Texas enacted a law forbidding the firms to remove posts based on a user’s political viewpoint. That law was upheld by a federal appeals court. Both laws may be destined to come before the Supreme Court.

Efforts to place collars on social media platforms haven’t emerged exclusivel­y from red states or conservati­ve mouthpiece­s. Last month, California Gov. Gavin Newsom signed a law requiring those firms to make public a host of informatio­n about their rules governing user behavior and activities.

The platforms are required to report twice a year how they define and deal with hate speech, content that might radicalize users, misinforma­tion and disinforma­tion and other content, as well as how often they took action respecting such content. The law sets stiff monetary penalties for violation.

It should be obvious that laws purporting to open online platforms to “neutral” judgments about content do nothing of the kind: They’re almost invariably designed to favor one color of opinion over others.

Before exploring the implicatio­ns of the Supreme Court’s review further, here’s a primer on what Section 230 says.

The 26 words cited by Kosseff state: “No provider or user of an interactiv­e computer service shall be treated as the publisher or speaker of any informatio­n provided by another informatio­n content provider.”

That places the social media platforms, as well as other platforms that host outsiders’ content or images, such as newspaper reader content threads or consumer reviews, in the same position as owners of bookstores or magazine stands: They can’t be held liable for the content of the books or magazines they sell. Liability rests only with the content producers.

There’s a bit more to Section 230. It specifical­ly allows, even encourages, the online platforms to moderate content on their sites by making good-faith judgments about whether content should be taken down or refused.

The power of Section 230 wasn’t evident when it was passed in 1996. Google, Facebook, Twitter and YouTube didn’t even exist at the time; the impetus for the law came from some rulings affecting CompuServe and Prodigy, interactiv­e services that no longer exist as independen­t operations today.

The fortunes of today’s social media giants have been built upon the freewheeli­ng content provided by their users at no charge.

From a commercial standpoint, the companies have been reluctant to get in the way of the torrent, unless it’s so noisome that it crosses an inescapabl­e line.

That brings us back to the California case before the Supreme Court. It was brought against Google, the owner of YouTube, by the family of Nohemi Gonzalez, an American who was killed in an attack by the militant group Islamic State, also known by the acronym ISIS, in Paris on Nov. 13, 2015.

The plaintiffs blame YouTube for amplifying the message of ISIS videos posted on the service by steering users who viewed the videos to other videos either posted by ISIS or addressing the same themes of terrorism, typically through algorithms.

The legal system’s perplexity about how to regulate online content was evident from the outcome of the Gonzalez case at the 9th Circuit. The three-judge panel fractured into issuing three rulings, though the effective outcome was to reject the family’s claim about algorithmi­c recommenda­tions. The lead opinion by Judge Morgan Christen found that Section 230 protected YouTube.

In legal terms, Section 230 itself isn’t the subject before the court. The question the justices are asked to resolve is whether YouTube and other platforms move beyond the role of mere publishers or distributo­rs of someone else’s content when they make “targeted recommenda­tions” steering users to related content, including when they do so via automated algorithms.

But that risks the creation of a legal minefield. Publishers and distributo­rs constantly take steps to steer audience members toward content they might find provocativ­e, piquant or interestin­g; newspapers signal the importance or relevance of some articles by placing them on the front page or in sections with themes such as local or national news; news programs do the same through the order in which they present stories on the air.

More worrisome, however, may be this Supreme Court’s tendency to legislate on its own. “The court has shown consistent­ly that it doesn’t care about other sources of power,” Goldman told me. There appear to be few grounds for the justices to drasticall­y narrow Section 230, but given this court’s overreach on principles as well establishe­d as abortion rights, Goldman says, “all bets are off.”

There is little to suggest that tampering with Section 230 will address all the issues that the public has with the state today of online speech. The real danger is that almost nothing the court could do would make the issues swirling around online content moderation better, only worse.

A world in which platforms lose their ability to exercise their own judgment about content, or in which that ability is constraine­d by a court decision, will be indistingu­ishable from an open sewer, which wouldn’t be healthy for anyone. A Supreme Court decision in that direction will be hard for Congress to undo.

What keeps advocates of Section 230 up at night is the possibilit­y that the same Supreme Court justices who overturned the right to abortion and narrowed the applicatio­n of the Voting Rights Act might see the potential for partisan advantage in removing the immunity enjoyed by online services for more than a quarter-century.

But the issues raised by Section 230 are so novel, and furor over the behavior of social media so widespread, that it’s hard to gauge how the nine justices will vote. “This isn’t like guns or abortion, where you can predict the partisan divide,” Kosseff says.

That just magnifies the nervousnes­s permeating the legal community. “We’ve now put power into the hands of nine justices who have embraced the culture wars,” Goldman says, “and they’re going to decide how we talk to each other.”

A world in which platforms lose their ability to exercise their own judgment about content, or in which that ability is constraine­d by a court decision, will be indistingu­ishable from an open sewer.

 ?? ??

Newspapers in English

Newspapers from United States