Criticism spurs Twitter change
Blocking weblinks to story on Bidens was wrong, CEO says
Twitter was wrong to block weblinks to a New York Post story that included potentially damaging allegations against Democratic presidential nominee Joe Biden, CEO Jack Dorsey said on Friday, as the company responded to criticism over its handling of the story that had prompted cries of censorship.
“Straight blocking of URLS was wrong, and we updated our policy and enforcement to fix,” he tweeted. “Our goal is to attempt to add context, and now we have capabilities to do that.”
Dorsey was weighing in after an executive at the social media company announced changes late Thursday to its policy on hacked content following an onslaught of criticism.
Twitter will no longer remove hacked material unless it’s directly shared by hackers or those working with them, the company’s head of legal, policy, trust and safety, Vijaya Gadde, said in a Twitter thread.
And instead of blocking links from being shared, tweets will be labeled to provide context, Gadde said.
“We want to address the concerns that there could be many unintended consequences to journalists, whistleblowers and others in ways that are contrary to Twitter’s purpose of serving the public conversation,” she said.
Twitter and Facebook had moved quickly this week to limit the spread of the story published by the New York Post.
San Francisco-based Twitter initially responded by banning users from sharing links to the article in tweets and direct messages because it violated the company’s policy prohibiting hacked content. But it didn’t alert users about why they couldn’t share the link until hours later.
By Friday, people were free to post the links again. Twitter said that was because the “once-private” information in the article is now “widely available” in the press and on other platforms.
Dorsey had first tweeted that it was “unacceptable” the company hadn’t provided more context around its action. A little over 24 hours later, Gadde announced the company was making changes after receiving “significant feedback (from critical to supportive)” about how it enforced the policy.
Facebook said it was “reducing” the story’s distribution on its platform while waiting for third-party fact-checkers to verify it.