The Norwalk Hour

Meta board urges changes to VIP moderation system

-

LONDON — Facebook parent Meta’s quasi-independen­t oversight board said Tuesday that an internal system that exempted high-profile users, including former U.S. President Donald Trump, from some or all of its content moderation rules needs a major overhaul.

The report by the Oversight Board, which was more than a year in the making, said the system “is flawed in key areas which the company must address.”

Meta asked the board to look into the system after The Wall Street Journal reported last year that it was being abused by many of its elite users, who were posting material that would result in penalties for ordinary people, including for harassment and incitement of violence.

Facebook’s rules reportedly didn’t seem to apply to some VIP users while others faced reviews of rule-breaking posts that never happened, according to the Journal article, which said the system had at least 5.8 million exempted users as of 2020.

The system — known as “XCheck,” or cross-check — was exposed in Facebook documents leaked by Frances Haugen, a former product manager turned whistleblo­wer who captured worldwide headlines with revelation­s alleging that the social media company prioritize­d profits over online safety and galvanized regulators into cracking down on hate speech and misinforma­tion.

Nick Clegg, Meta’s president for global affairs, tweeted that the company requested the review of the system “so that we can continue our work to improve the program.”

To fully address the board’s recommenda­tions, “we’ve agreed to respond within 90 days,” he added.

The company has said crosscheck, which applies to Facebook and Instagram, was designed to prevent “overpolici­ng,” or mistakenly removing content thought to be breaking the platform’s rules.

The Oversight Board’s report said that the cross-check system resulted in users being treated unequally and that it led to delays in taking down content that violated the rules because there were up to five separate checks. Decisions on average took more than five days, it found.

For content posted by American users, the average decision took 12 days, and for Afghanista­n and Syria, it was 17 days. In some cases, it took a lot longer: one piece of content waited 222 days — more than seven months — for a decision, the report said, without providing further details.

Among its 32 recommenda­tions, the board said Meta “should prioritize expression that is important for human rights, including expression which is of special public importance.”

Human rights defenders, advocates for marginaliz­ed communitie­s, public officials and journalist­s should be given higher priority than others put on the cross-check list because they are business partners, such as big companies, political parties, musicians, celebritie­s and artists, the report said.

“If users included due to their commercial importance frequently post violating content, they should no longer benefit from special protection,” the board said.

Addressing other flaws, the board also urged Meta to remove or hide content while it’s being reviewed and said the company should “radically increase transparen­cy around cross-check and how it operates,” such as outlining “clear, public criteria” on who gets to be on the list.

Newspapers in English

Newspapers from United States