The Arizona Republic

Facebook in a quandary after leaving up fake Pelosi video

- Barbara Ortutay

SAN FRANCISCO – When a doctored video of House Speaker Nancy Pelosi – one altered to show the Democratic leader slurring her words – began making the rounds on Facebook last week, the social network didn’t take it down. Instead, it “downranked” the video, a behind-the-scenes move intended to limit its spread.

That outraged some people who believe Facebook should do more to clamp down on misinforma­tion. Pelosi derided Facebook on Wednesday for not taking down the video even though it knows it is false.

But the company and some civil libertaria­ns warn that Facebook could evolve into an unaccounta­ble censor if it’s forced to make judgment calls on the veracity of text, photos or videos.

Facebook has long resisted making declaratio­ns about the truthfulne­ss of posts that could open it up to charges of censorship or political bias. It manages to get itself in enough trouble simply

trying to enforce more basic rules in difficult cases, such as the time a straightfo­rward applicatio­n of its ban on nudity led it to remove an iconic Vietnam War photo of a naked girl fleeing a napalm attack. (It backed down after criticism from the prime minister of Norway, among others.)

But staying out of the line of fire is harder than it used to be, given Facebook’s size, reach and impact on global society. The social network can’t help but run into controvers­y given its 2.4 billion users and the sorts of decisions it must make daily– everything from which posts and links it highlights in your news feed to deciding what counts as hate speech to banning controvers­ial figures or leaving them be.

Facebook has another incentive to keep its head down. The deeper it gets into editorial decisions, the more it looks like a publisher, which could tempt lawmakers to limit the liability shield it currently enjoys under federal law. In addition, making judgments about truth and falsity could quickly become one of the world’s biggest headaches.

For instance, Republican politician­s and other conservati­ves, from President Donald Trump to Fox News personalit­ies, have been trumpeting the charge that Facebook is biased against conservati­ves.

That’s a “false narrative,” said Siva Vaidhyanat­han, director of the Center for Media and Citizenshi­p at the University of Virginia. But as a result, he said, “any effort to clean up Facebook now would spark tremendous fury.”

Twitter hasn’t removed the doctored Pelosi video, either, and declined to comment on its handling of it. But YouTube yanked it, pointing to community guidelines that prohibit spam, deceptive practices and scams. Facebook has a similar policy that prohibits the use of “misleading and inaccurate” informatio­n to gain likes, followers or shares, although it apparently decided not to apply it in this case.

None of these companies explicitly prohibits false news, although Facebook notes that it “significan­tly” reduces the distributi­on of such posts by pushing them lower in user news feeds.

The problem is that such downrankin­g doesn’t quite work, Vaidhyanat­han said. As of Wednesday, the video shared on Facebook by the group Politics Watchdog had been viewed nearly 3 million times and shared more than 48,000 times. By contrast, other videos posted by this group in the past haven’t had more than a few thousand views apiece.

Further complicati­ng matters is the fact that Facebook is starting to deemphasiz­e the news feed itself. CEO Mark Zuckerberg has outlined a broad strategy that will emphasize private messaging over public sharing on Facebook. And Facebook groups, many of which are private, aren’t subject to downrankin­g, Vaidhyanat­han said.

Facebook didn’t respond to emailed questions about its policies and whether it is considerin­g changes that would allow it to remove similar videos in the future. In an interview last week with CNN’s Anderson Cooper, Facebook’s head of global policy, Monika Bickert, defended the company’s decision, noting that users are “being told” that the video is false when they view or share it.

That might be a stretch. When an Associated Press reporter attempted to share the video as a test, a Facebook pop-up noted the existence of “additional reporting” on the video with links to fact-check articles, but didn’t directly describe the video as false or misleading.

Alex Stamos, Facebook’s former security chief, tweeted Sunday that few critics of the social network’s handling of the Pelosi video could articulate realistic enforcemen­t standards beyond “take down stuff I don’t like.” Mass censorship of misleading speech on Facebook, he wrote, would be “a huge and dangerous increase in FB’s editorial power.”

Last year, Zuckerberg wrote on Facebook that the company focuses on downrankin­g so-called “borderline content,” stuff that doesn’t violate its rules but is provocativ­e, sensationa­list, “click-bait or misinforma­tion.”

While it’s true that Facebook could just change its rules around what is allowed – moving the line on acceptable material – Zuckerberg said this doesn’t address the underlying problem of incentive.

If the line of what is allowed moves, those creating material would just push closer to that new line.

Facebook continuous­ly grapples with the right way to deal with new forms of misinforma­tion, Nathaniel Gleicher, the company’s head of cybersecur­ity policy, said in a February interview with the AP. The problem is far more complex than carefully manipulate­d “deepfake” videos that show people doing things they never did, or even crudely doctored videos such as the Pelosi clip.

Newspapers in English

Newspapers from United States