Blowing the Whistle on Facebook

Wondering what all the fuss is around the Facebook Papers? Get the lowdown here.

A large trove of recently leaked documents from Meta/Facebook promises to keep the social platform in the news, and in hot water, for some time to come. While other recent “Paper” investigations (think Panama and Paradise) have revealed fraud, tax evasion, and all manner of wrongdoing, it’s not clear that exposing Facebook’s own organizational musings will lead to the same kind of fallout, but whatever happens, this is not likely to end well for Facebook.

Here’s what’s happening and what’s likely to come as we learn more.

What’s going on and who is behind it?

Beginning in September, the Wall Street Journal began a series of unflattering stories they referred to as the Facebook Files. The primary claim in these stories is that the social media giant has for some time been aware of the harm and potential for harm inherent in the design choices of its platform.

In early October a former Facebook employee, Frances Haugen, appeared on CBS’s “60 Minutes” revealing herself to be the inside source for the Wall Street Journal articles. During her interview, she put a fine point on her message saying that Facebook’s own internal research shows that their tools “amplify hate, misinformation, and political unrest, but the company hides what it knows.” Before leaving Facebook, she copied tens of thousands of pages of internal research and other communications like strategy presentations and employee discussion board conversations.

Through her lawyers from Whistleblower Aid, Haugen has filed complaints with the U.S. Securities and Exchange Commission, which is the government body that enforces business and financial laws. The gist of the complaints is that Facebook has been misleading investors given what it knows from its own research versus what it presents to the public. She has also testified before the U.S. Congress and the U.K. Parliament and shared her collection of redacted (to hide personal information of employees and users) documents with them.

Haugen and her attorneys subsequently released the documents to selected news outlets who are analyzing them. CNN reported on October 26 that 17 organizations now have access to the documents. The news groups we know of are The Associated Press, CNBC, CNN, Gizmodo, The Guardian, The Information, The New York Post, The New York Times, Politico, The Wall Street Journal, The Washington Post, and WIRED. More documents are expected to come out as Whistleblower Aid continues their redaction efforts.

What does this mean for Facebook?

We don’t know yet if the SEC intends to open an investigation into Facebook over these charges. The burden will be on the SEC to show clearly that Facebook willingly or through recklessness misled investors, which will be a hard case to make. However, even without prosecution from the SEC, Facebook could have problems with investors. The New York Times reports that these revelations “worry investors like Julie Goodridge, a portfolio manager for NorthStar Asset Management. She, along with the New York State Comptroller’s Office and other investment funds, filed a motion for the next shareholder meeting calling for the removal of Mr. Zuckerberg’s power as majority voting shareholder.”

There are also signs that Facebook investors in general may not be pleased. While this may not be a lasting effect, it’s worth noting that Facebook’s share price has dropped 15% since The Wall Street Journal started their reporting while the S&P index has trended mostly upward over the same time period.

It’s also possible that Haugen’s testimony to lawmakers in Britain and the U.S. will influence lawmakers to regulate more stringently than they might have otherwise. This could be particularly bad timing from Facebook’s perspective since the U.K. is right now debating the Online Harms Bill. In the U.S. this information comes out in an environment where there have already been calls to break up Facebook. Lawmakers might also use Facebook’s conduct to strengthen their case for making changes to Section 230 of the Communications Decency Act, which protects platforms like Facebook from civil lawsuits.

What is Facebook accused of exactly?

Besides the charges of misleading investors, Facebook is facing a slew of image tarnishing revelations. Reporters continue to analyze the documents, so there may be more on the horizon, but the following is a sample of some of the information that has come out so far.

Facebook’s own researchers and internal studies repeatedly showed that Facebook’s features and design choices function against the public’s best interests:

  • Instagram harms teenagers, especially teenage girls, exacerbating eating disorders and increasing suicidal feelings in teens.
  • The more a piece of content attracts comments of outrage and division, the more likely Facebook’s algorithms are to prioritize it in users’ feeds.
  • As people show interest in a topic, Facebook’s recommender algorithms will suggest more extreme versions of similar content.
  • Facebook’s incentives force even traditional media to be more polarizing, producing “darker, more divisive content.” Facebook’s 2018 change to create more “meaningful social interactions” had the opposite effect and devastated revenue for media companies.

Other revelations include:

  • Facebook knew that extremist groups have been using their platform to polarize American voters and failed to take significant action.
  • Facebook acted on only 3 to 5 percent of hate speech and less than 1 percent of speech advocating violence while at the same time claiming that AI has been uncovering the vast majority of bad content.
  • Mark Zuckerberg’s public comments and congressional testimony have often been at odds with what the documents show.
  • According to the civic integrity team, Facebook has been used to fan ethnic violence in some countries.

What does Facebook say?

It should be said that some of the information you’ll hear about what Facebook knew about potential harms comes from the internal employee discussions and is not necessarily official information. Facebook’s internal social network, which resembles the public one, is a place where many employees engage in conversations on a wide range of topics expressing various points of view and individual opinions. (Note that the items we listed above come from internal research studies or presentations rather than the informal employee discussion.)

Officially, Facebook says that they have already made changes to address many of the concerns revealed in these documents. In response to the original Wall Street Journal articles, Facebook said that the stories were deliberate mischaracterizations that attributed false motives to Facebook’s leadership. They maintain that the company is not responsible for the organic political divisions in the country, and they are not responsible for the current state of the media. They say that they have no commercial or moral incentive to create negative experiences for their users. They point out that the company has to make difficult decisions that must balance varying interests and have long advocated for Congress to pass updated regulations to set appropriate guidelines. Moreover, they say they make extensive disclosures in their SEC filings about their challenges giving investors the information they need to make informed decisions.

Was it legal for Haugen to take and publicize these documents?

Facebook can certainly challenge the legality of taking these documents, but there are several whistleblower protections in place that allow employees to reveal information for the purpose of exposing wrongdoing. The whistleblower protections also supersede any nondisclosure agreement that might have been in effect provided the information released is relevant to the allegations of wrongdoing.

While much of what’s been revealed was already known or suspected by Facebook watchers, the Facebook Papers provide documentary evidence that can’t easily be explained or apologized away. We may see more evidence of real or perceived misdeeds as journalists continue to review the documents. The headlines are bound to fade as news cycles move on, but rest assured that researchers, journalists, and lawmakers will continue to discuss the implications of the information they’re finding. There seems little doubt that Facebook will have to make changes to the way they operate. Still to be seen is whether those changes will be self-imposed or exerted on Facebook by various governments around the world.

Does Haugen Have an Agenda?

There has not been any evidence to suggest Haugen is not sincere in her desire to expose what she considers wrongdoing at Facebook. She has testified to legislatures in the U.S., the U.K., and the EU voicing her opinion that strong regulations will be required to keep companies like Facebook in check. She expressed her hope in Belgium that the EU’s Digital Services Act could serve as a global gold standard that could be adopted by the rest of the world.

Updated Nov. 15, 2021 11:06:02 PST: Added the question “Does Haugen Have an Agenda?”

Updated Nov. 11, 2021 17:37:28 PST: The Information was added to the list of publications having access to the Facebook Papers.

More posts like this

We want content moderation to enhance your users’ experience and so they can find their special one more easily.

The Psychology Behind AI Content Moderation: Understanding User Behavior

Social media platforms are experiencing exponential growth, with billions of users actively engaging in content creation and sharing. As the volume of user-generated content continues to rise, the challenge of content moderation becomes increasingly complex. To address this challenge, artificial intelligence (AI) has emerged as a powerful tool for automating the moderation process. However, user…
5 minutes

Content Moderation for Virtual Reality

What is content moderation in virtual reality? Content moderation in virtual reality (VR) is the process of monitoring and managing user-generated content within VR platforms to make sure it meets certain standards and guidelines. This can include text, images, videos, and any actions within the 3D virtual environment. Given the interactive and immersive nature of…
31 minutes

The Impact of Trust and Safety in Marketplaces

Nowadays, its no surprise that an unregulated marketplace with sketchy profiles, violent interactions, scams, and illegal products is doomed to fail. In the current world of online commerce, trust and safety are essential, and if users don't feel comfortable, they won’t buy. As a marketplace owner, ensuring that your platform is a safe and reliable…
9 minutes

How AI is Revolutionizing Content Moderation in Social Media Platforms

Social media platforms have become an integral part of our lives, connecting us with friends, family, and the world at large. Still, with the exponential growth of user-generated content, ensuring a safe and positive user experience has become a daunting task. This is where Artificial Intelligence (AI) comes into play, revolutionizing the way social media…
3 minutes

Customizing AI Content Moderation for Different Industries and Platforms

With the exponential growth of user-generated content across various industries and platforms, the need for effective and tailored content moderation solutions has never been more apparent. Artificial Intelligence (AI) plays a major role in automating content moderation processes, but customization is key to address the unique challenges faced by different industries and platforms. Understanding Industry-Specific…
3 minutes

Emerging Threats in AI Content Moderation : Deep Learning and Contextual Analysis 

With the rise of user-generated content across various platforms, artificial intelligence (AI) has played a crucial role in automating the moderation process. However, as AI algorithms become more sophisticated, emerging threats in content moderation are also on the horizon. This article explores two significant challenges: the use of deep learning and contextual analysis in AI…
4 minutes

The Impact of AI Content Moderation on User Experience and Engagement

User experience and user engagement are two critical metrics that businesses closely monitor to understand how their products, services, or systems are being received by customers. Now that user-generated content (UGC) is on the rise, content moderation plays a main role in ensuring a safe and positive user experience. Artificial intelligence (AI) has emerged as…
4 minutes

Future Technologies : The Next Generation of AI in Content Moderation 

With the exponential growth of user-generated content on various platforms, the task of ensuring a safe and compliant online environment has become increasingly complex. As we look toward the future, emerging technologies, particularly in the field of artificial intelligence (AI), are poised to revolutionize content moderation and usher in a new era of efficiency and…
3 minutes

Global Perspective : How AI Content Moderation Differs Across Cultures and Religion

The internet serves as a vast platform for the exchange of ideas, information, and opinions. However, this free exchange also brings challenges, including the need for content moderation to ensure that online spaces remain safe and respectful. As artificial intelligence (AI) increasingly plays a role in content moderation, it becomes essential to recognize the cultural…
5 minutes

Ethical Consideration in AI Content Moderation : Avoiding Censorship and Biais

Artificial Intelligence has revolutionized various aspects of our lives, including content moderation on online platforms. As the volume of digital content continues to grow exponentially, AI algorithms play a crucial role in filtering and managing this content. However, with great power comes great responsibility, and the ethical considerations surrounding AI content moderation are becoming increasingly…
3 minutes

‍The Future of Dating: Embracing Video to Connect and Thrive

In a rapidly evolving digital landscape, dating apps are continually seeking innovative ways to enhance the user experience and foster meaningful connections. One such trend that has gained significant traction is the integration of video chat features. Video has emerged as a powerful tool to add authenticity, connectivity, and fun to the dating process. In…
4 minutes

17 Questions Trust and Safety Leaders Should Be Able to Answer 

A Trust and Safety leader plays a crucial role in ensuring the safety and security of a platform or community. Here are 17 important questions that a Trust and Safety leader should be able to answer.  What are the key goals and objectives of the Trust and Safety team? The key goals of the Trust…
6 minutes

What is Content Moderation: a Guide

Content moderation is one of the major aspect of managing online platforms and communities. It englobes the review, filtering, and approval or removal of user-generated content to maintain a safe and engaging environment. In this article, we'll provide you with a comprehensive glossary to understand the key concepts, as well as its definition, challenges and…
15 minutes

A Guide to Detect Fake User Accounts

Online social media platforms have become an major part of our daily lives: with the ability to send messages, share files, and connect with others, these networks provide a way, for us users, to stay connected. Those platforms are dealing with a rise of fake accounts and online fraudster making maintaining the security of their…
4 minutes

How Predators Are Abusing Generative AI

The recent rise of generative AI has revolutionized various industries, including Trust and Safety. However, this technological advancement generates new problems. Predators have found ways to abuse generative AI, using it to carry out horrible acts such as child sex abuse material (CSAM), disinformation, fraud, and extremism. In this article, we will explore how predators…
4 minutes

Prevent unwanted content from reaching your platform

Speak to one of our experts and learn about using AI to protect your platform
Talk to an expert