Marysville Appeal-Democrat

How the debate over holding internet platforms accountabl­e is changing under Biden

- Tribune News Service Los Angeles Times

Two people were dead; one was injured; and Jason Flores-williams wanted to hold Facebook responsibl­e.

But after filing a lawsuit in

September alleging that the website’s lax moderation standards led to 17-year-old Kyle Rittenhous­e killing two protesters in Kenosha, Wis., over the summer, Flores-williams withdrew the suit in January. His fight for accountabi­lity had collided with a law the activist attorney came to see as a “brick wall.”

“You have no levers of control, no leverage,” he told The Times. “You’re up against Section 230.”

A snippet of text buried in the 1996 Telecommun­ications Act, Section 230 is the regulation under which websites enjoy broad freedom to choose if and how they moderate user-generated content. Flores-williams had alleged that a Facebook post by the Kenosha Guard militia summoning armed civilians to the city had laid the groundwork for Rittenhous­e’s violence there; but as Section 230 is written, Facebook and its peers are rarely liable for what their users post — even when it results in death.

Flores-williams isn’t alone in seeing the law as outdated. President Joe Biden, former president Donald Trump and a long list of Democrats and Republican­s have all pushed for the law to be restructur­ed or scrapped entirely amid increasing­ly bipartisan criticism of Big Tech.

But if liberals and conservati­ves are united in their calls for reform, they’re split on what that reform should look like — leaving internet companies stuck in a limbo where a massive forced change to their business model is constantly discussed yet never quite materializ­es.

Meanwhile, those who seek to hold the platforms accountabl­e for the harms caused by content spread there are left searching for new approaches that might offer a greater chance of success — which is to say, any at all.

Section 230 takes a two-pronged approach to content moderation: not only does it absolve websites of liability for user content they don’t moderate, but it also says they can moderate user content when they choose to.

That lets social networks, chat forums and review websites host millions of users without having to go to court every time they leave up a post that’s objectiona­ble, or take one down that’s not.

Online platforms usually, though not uniformly, support leaving Section 230 the way it is. In a congressio­nal hearing last fall, Alphabet Chief Executive Sundar Pichai and Twitter CEO Jack Dorsey warned that the internet only works thanks to the protection­s afforded by the law; Facebook CEO Mark Zuckerberg broke ranks to say the law should be updated, citing a need to promote transparen­cy around moderation practices.

Of the law’s critics, conservati­ves typically lean toward unrestrict­ed speech. A Trump executive order sought to modify the law so users could sue platforms if they restricted content that wasn’t violent, obscene or harassing, although legal experts said the order was unlikely to hold up in court and it appears to have had little impact on how the platforms conduct themselves.

On the left, critics have called for a version of Section 230 that would encourage more rigorous moderation. Reforms targeting sex traffickin­g and child abuse have also garnered bipartisan support in the past.

Both sides have only gotten louder in recent weeks: the Jan. 6 siege of the U.S. Capitol prompted concern from the left about the role unregulate­d social media can play in organizing real-world violence, while the subsequent banning of Trump’s Facebook and Twitter accounts gave the right a striking example of how easily tech platforms can silence their users.

With Democrats now controllin­g the presidency and both houses of Congress, the party has an opportunit­y to rewrite Section 230, but it has yet to achieve consensus, with members floating multiple differentl­y calibrated proposals over the last year.

The latest of those is the SAFE

TECH Act, proposed last month by Sens. Mazie Hirono, D-hawaii, Amy Klobuchar, D-minn., and Mark R. Warner, D-VA. The bill would increase platforms’ liability for paid content and in cases involving discrimina­tion, cyberstalk­ing, targeted harassment and wrongful death.

Flores-williams said that last item in particular, which the sponsors say would allow “the family of a decedent to bring suit against platforms where they may have directly contribute­d to a loss of life,” opens the door for future cases along the lines of his withdrawn suit.

It could also bolster suits over deaths such as that of Brian Sicknick, the Capitol police officer who died after defending the Capitol on Jan. 6. The official cause of Sicknick’s death has yet to be determined, but the case is cited by the bill sponsors in their argument for the carve-out.

The implicatio­ns could extend well beyond high-profile deaths, too.

“Talk about floodgates, right?,” said Daniel Powell, an attorney at the internet-focused firm Minc Law. “Floodgates to millions in liability for lawsuits where people have died for any reason that has any tangential relationsh­ip to social media.”

It’s not clear how broadly lawmakers and prosecutor­s would try to interpret SAFE TECH’S provisions, but if passed, the bill could force tech companies to rethink how they engage with user-generated content.

Nadav Shoval, CEO and co-founder of Openweb — a platform which manages comment sections for online media outlets such as Techcrunch and Salon — said changes to Section 230 could hinder innovation through overly broad liability.

“I have more questions than answers on this specific proposal, but I remain confident that changing the law at all is a mistake,” Shoval said of the SAFE TECH Act via email. “We have other laws in place that are not [Section] 230 to ensure the communitie­s we host are safe, free from violence, hate speech, discrimina­tion, etc.”

But clearer guidelines around moderating and distributi­ng user content would be helpful, Shoval said; those are areas “which should be slightly regulated, or at least more clear, because right now there’s a lot of gray areas … where some guidance would definitely help.”

Other social media platforms that would be affected by the passage of the SAFE TECH Act — including Facebook, Twitter, Google, Reddit and Snapchat — declined or did not answer a request for comment on the bill.

Newspapers in English

Newspapers from United States