Press "Enter" to skip to content

Internal Facebook documents highlight its moderation and misinformation issues

The Facebook Papers, a vast trove of documents supplied by whistleblower Frances Haugen to a consortium of news organizations has been released. The reporting, by Reuters, Bloomberg, The Washington Post and others, paints a picture of a company that repeatedly sought to prioritize dominance and profit over user safety. This was, however, despite a large number of employees warning that the company’s focus on engagement put users at risk of real-world violence.

The Washington Post, for instance, claims that while Facebook CEO Mark Zuckerberg played down reports that the site amplified hate speech in testimony to Congress, he was aware that the problem was far broader than publicly declared. Internal documents seen by the Post claim that the social network had removed less than five percent of hate speech, and that executives — including Zuckerberg — were well aware that Facebook was polarizing people. The claims have already been rebutted by Facebook, which says that the documents have been misrepresented.

Zuckerberg is also accused of squashing a plan to run a Spanish-language voter-registration drive in the US before the 2020 elections. He said that the plan may have appeared “partisan,” with WhatsApp staffers subsequently offering a watered-down version partnering with outside agencies. The CEO was also reportedly behind the decision not to clamp down on COVID-19 misinformation in the early stages of the pandemic as there may be a “material tradeoff with MSI [Meaningful Social Interaction — an internal Facebook metric] impact.” Facebook has refuted the claim, saying that the documents have been mischaracterized.

Reuters reported that Facebook has serially neglected a number of developing nations, allowing hate speech and extremism to flourish. That includes not hiring enough staffers who can speak the local language, appreciate the cultural context and otherwise effectively moderate. The result is that the company has unjustified faith in its automatic moderation systems which are ineffective in non-English speaking countries. Again, Facebook has refuted the accusation that it is neglecting its users in those territories.

One specific region that is singled out for concern is Myanmar, where Facebook has been held responsible for amplifying local tensions. A 2020 document suggests that the company’s automatic moderation system could not flag problematic terms in (local language) Burmese. (It should be noted that, two years previously, Facebook’s failure to properly act to prevent civil unrest in Myanmar was highlighted in a report from Business for Social Responsibility.)

Similarly, Facebook reportedly did not have the tools in place to detect hate speech in the Ethiopian languages of Oromo or Amharic. Facebook has said that it is working to expand its content moderation team and, in the last two years, has recruited Oromo, Amharic and Burmese speakers (as well as a number of other languages).

The New York Times, reports that Facebook’s internal research was well-aware that the Like and Share functions — core elements of how the platform work — had accelerated the spread of hate speech. A document, titled What Is Collateral Damage, says that Facebook’s failure to remedy these issues will see the company “actively (if not necessarily consciously) promoting these types of activities.” Facebook says that, again, these statements are based on incorrect premises, and that it would be illogical for the company to try and actively harm its users.

Bloomberg, meanwhile, has focused on the supposed collapse in Facebook’s engagement metrics. Young people, a key target market for advertisers, are spending less time on Facebook’s platform, with fewer teens opting to sign up. At the same time, the number of users may be artificially inflated in these age groups, with users choosing to create multiple accounts — “Finstas” — to separate their online personas to cater to different groups. Haugen alleges that Facebook “has misrepresented core metrics to investors and advertisers,” and that duplicate accounts are leading to “extensive fraud” against advertisers. Facebook says that it already notifies advertisers of the risk that purchases will reach duplicate accounts in its Help Center, and lists the issue in its SEC filings.

Over the weekend, Axios reported that Facebook’s Sir Nick Clegg warned that the site should expect “more bad headlines” in the coming weeks. Between the material available in the Facebook Papers, another round of Frances Haugen’s testimony in the UK later today and rumors of more whistleblowers coming forward, it’s likely that Facebook will remain in the headlines for some time.

Editor’s note: This article originally appeared on Engadget.

This post was originally published on this site

Be First to Comment

Leave a Reply

Your email address will not be published. Required fields are marked *