FACEBOOK IS A CORRUPT EMPIRE OF EVIL SAY INSIDERS

United States Government executives including Dianne Feinstein, and her family; Nancy Pelosi, and her family; Kamala Harris and her family; Jennifer Granholm, and her family, Barack Obama, and his family; and current and past White House officials, and their families, own Facebook, Google, Netflix and their affiliates. They have hired their staffing from those companies taken orders from those companies, commanded those companies to harm their enemies and whistle-blowers, received BILLIONS of dollars of unreported political campaign financing from those companies via indirect cash and internet manipulation services.

These, and other government officials ordered Facebook, Google, Netflix to harm Plaintiff because he was a whistle-blower and a competitor. In fact, in a recent federal hearing at the United States Patent Office, the USPTO determined that Plaintiff had invented the core technology that Facebook relies on, but government officials blockaded Plaintiff’s issuance of the paperwork to monetize that fact because Facebook and Google’s shills work in, and operate control over, the USPTO. In another case, government officials ordered Facebook and it’s “Cartel”; and Google and it’s “Cartel” to blockade all media coverage of Plaintiff’s electric car company while exclusively promoting Tesla’s car company, which government officials own covert stock in. Defendants have made comments to the Court and officials that they plan to use a strategy of further defamation against Plaintiff by trying to label him as a “crazy inventor” and that Defendants will seek to portray their ownerships and control of Facebook and Google, in particular, as “benign” support for a “benign entity” which Plaintiff is making “wild conspiracy assertions” about. A long time favored tactic of Defendant.

Alas, though: An assertion is not a “Conspiracy Theory” when the assertion is based on FACTS!

Recently the District of Columbia (the “District”), by the Office of the Attorney General, brought an action against Defendant Mark Zuckerberg, who Plaintiff was up against in his recent USPTO hearing, for violations of the District’s Consumer Protection Procedures Act, D.C. Code §§ 28-3901, et seq. (“CPPA”). In support of its claims, the District stated as follows:

In addition to the following allegations, this District is also in possession of significant evidence that further demonstrates both the propriety of the Court’s exercise of personal jurisdiction over Mr. Zuckerberg as well as his violations of the CPPA. That evidence was obtained through discovery in the District’s litigation against Facebook (now Meta Platforms, Inc.). District of Columbia v. Facebook, Inc. , 2018 CA 008715 B. In an excess of caution regarding the protective order issued in that action, the District has served discovery requests with this complaint for re-production of that material in this case, among other requests. See D.C. Super. Ct. R. 26 cmt. (a)(1) (“The Superior Court rules allow parties to begin discovery at the filing of the complaint; this process gives parties greater options for early discovery than those available under the Federal Rules.”). Citations to that additional evidence do not presently appear in this complaint, and the District will amend (either by right or with leave of Court) once it has obtained that information from Defendant or Facebook directly. “

When the United States Patent Office, in a federal hearing, said that I invented the core technology operating Facebook and then refused to give me my patent paperwork, it seemed like the game might be rigged. I hired ex FBI and CIA agents to check it out… BOY, was it ever rigged. Facebook’s own staff and investors worked INSIDE the U.S. Patent Office. Then they showed me that most of the Obama and Biden Administration COVERTLY OWN FACEBOOK!!!

What U.S. judge or regulator is complaining? Who would complain aboutSands Capital’s failure to file the S.E.C. Form SC 13G notices of acquisition of Facebook, Baidu and Athenahealth stock?

Not S.E.C Chairman Mary L. Schapiro—she held a boatload of “dark pool” Fidelity, Vanguard, AllianceBern, TIAA-CREF and T. Rowe Price funds.

Not Commerce Secretary #1 Rebecca M. Blank—she held TIAA-CREF, Vanguard and Fidelity funds.

Not Commerce Secretary #2 Penny S. Pritzker—she holds up to $23.4 million Morgan Stanley, JPMorgan and Goldman Sachs Facebook dark pools.

Not Attorney General Eric H. Holder—he held T. Rowe Price and Fidelity funds. In fact, Holder held Fidelity Contrafund, the largest singleFacebook mutual fund stock holder, valued at $413 million.

Who in the judiciarywould complain?

Not Leader v. Facebook  Chief Justice John G. Roberts, Jr.—he held  Microsoft, T. Rowe Price, Fidelity, Janus, Vanguard and Blackrockfunds, including Fidelity Contrafund.

Not Leader v. Facebook  Federal Circuit Judges Alan D. Lourie, KimberlyA. Moore and Evan J. Wallach—they held Fidelity, Vanguard and T. Rowe Price funds, including Fidelity Contrafund.

Not Leader v. Facebook  District Court Judge Leonard P. Stark—he held Vanguard and Fidelity funds.

Not Leader v. Facebook  Patent Office Director David J. Kappos—he held over a million dollars of Vanguard funds.

We uncovered a felony, organized crime-level of QUID PRO QUO media manipulation, RICO violations and anti-trust law violations between government officials and Facebook blockading my patent and business rights. Over 500 high level government officials own Facebook and harm it’s competitors to protect their insider trading stocks.

We noticed nobody every prosecutes these people.

Can the Court or the Defendants explain that?

DOES ANYONE IN THE COURT ASSOCIATED WITH MY CASE, OR THEIR FAMILY, OWN STOCK IN Twitter, Google – Alphabet, Facebook – Meta, Netflix or YouTube or Tesla; as that would be a MASSIVE CONFLICT OF INTEREST IN THIS CASE?

Given the suspicious timing of the White House appointments of massively conflicted officials to oversee America’s digital infrastructure:

1. How much did Defendants you know about SANDS CAPITAL’S collusion with the Chinese?

2. What Defendants you going to do about it?

3. What assurances can Defendants give us that the tech people you have hand picked are worthy of America’s trust?

4. Will the new systems really protect Americans’ privacy, property and security?

5. Did Defendants know that their Securities Chair held stock in Facebook and Baidu before the Facebook IPO?

6. Why didn’t Defendants personal White House counsels from Perkins Coie LLP, namely Robert F. Bauer and Anita B. Dunn, husband and wife respectively, submit ethics pledges and financial disclosures? Did Defendants know that Facebook was one of their clients?

7. Where are Eric Schmidt’s financial disclosures and written ethics pledges to the U.S. Government?

8. Did Defendants know that a Florida judge was ordered to recuse himself from a case where he was Facebook Friends with one of the litigating attorneys?

9. What do Defendants 50 million “likes” say about Defendants appointment of two of the four judges in the Leader v. Facebook case, not even counting all their financial holdings in Facebook, or the Patent Office’s “likes”?

Introduction

1. In under two decades, Facebook, Inc. (now known as Meta Platforms, Inc.) (“Facebook”) has grown from a small online social network to an implacable corporate giant.

Facebook offers a variety of products and services, including the well-known Facebook product.

Today, Facebook is larger than any single country—with more than 2.9 billion monthly active users, nearly half the global population. To put that in perspective, Facebook has more users than the populations of the United States, China and Brazil combined. And Facebook has become wealthier than over 150 countries worldwide, including Switzerland, Sweden, and the UAE. Not surprisingly, Facebook has seized enormous influence over global affairs. Facebook controls how people communicate with friends and family, conduct business online, what news they read, and even how they communicate with governments and elected officials. Atop it all is Mark Zuckerberg, the unelected leader of a massive digital empire with billions of inhabitants.

2. But Zuckerberg’s Facebook is far from a disinterested platform for people to communicate, stay in touch with friends, and reconnect with old acquaintances. Instead, Facebook has become a wildly successful and unique business, deriving enormous wealth from acquiring and monetizing the data of those billions of people leading their lives in Facebook’s digital ecosystem. But even that is not enough. Facebook is in a relentless pursuit to expand its reach on humanity and bring an ever-increasing number of people under its influence.

3. To that end, Mark Zuckerberg has been building his version of the Internet where the “default is social.” To him, that means building an Internet where people live their digital lives on Facebook. The goal is to convince people to reveal the most granular details of who they are to Facebook—their religions, their work histories, their likes—so that it can be monetized, and Zuckerberg and his company can continue to grow even wealthier.

4. Facebook—at Zuckerberg’s direction—shifted its business model in this way because it recognized that it could be even more profitable if it could harness and sell the ability to dependably influence its users’ behavior to third parties. Facebook therefore encouraged (and, at times, teamed up with) developers and researchers to collect and analyze Facebook user data so that it could better learn how to manipulate its own users’ moods and influence what they purchase and even whether and how they vote.

5. Facebook has become among the world’s leading innovators in experimenting on how to keep users engaged—meaning more data and more money for Facebook. But at Facebook’s scale, these experimental decisions reverberate globally.

6. That is in part because Facebook has realized an ugly truth: its social platform becomes “stickier” (meaning people will stay on it longer and share more data) when it is filled with toxicity. What this means is that the more Zuckerberg’s Facebook stokes divisiveness and polarization, destabilizes democracies, amplifies genocides, and impacts users’ mental health, the more money Facebook and its leaders make.

7. Given the trillions of dollars at issue, and having no regard for the people it purports to serve, Facebook—at Zuckerberg’s direction—has decided to hide these problems for as long as possible, including intentionally misleading Facebook users as well as the public, the press, and political leaders.

8. One prime example—and the one that forms the basis for the instant suit—was Facebook’s 2010 decision to open up the Facebook Platform to third parties. Again the brainchild of Zuckerberg, this move helped Facebook by persuading outside developers to build eye-catching applications for Facebook—directing even more users, and user data, into the platform. Developers, though, could access the massive trove of user data that Facebook had collected through the “side door” of applications.

9. Zuckerberg had always been aware that the success of Facebook hinged on convincing users that their data was private enough, while selling as much access to those users as possible without driving them away. And Zuckerberg was fully aware that users would be concerned by this newly vulnerable position. So Zuckerberg engaged in a decade-long campaign designed to convince users that Facebook cared about and tried to protect users and their data. Behind closed doors however, Zuckerberg insisted that Facebook’s policies be “as simple as we can get away with.”2 Given that Facebook’s platform was designed to allow abuse, Zuckerberg’s company largely operated without proper safeguards in place to protect users: policy enforcement was lax, review of app violations was inconsistent or subjective, and the policies themselves were unclear and confusing. But in 2018, the world learned of this sham.

10. In March 2018, whistleblower Christopher Wylie publicly revealed that a company called Cambridge Analytica—a London-based electioneering firm—exfiltrated the personal data of more than 70 million Facebook users in the United States, including more than 340,000 District residents, in order to influence the results of the 2016 United States presidential election. This data trove included Facebook users’ ages, interests, pages they’ve liked, groups they belong to, physical locations, political affiliation, religious affiliation, relationships, and photos, as well as their full names, phone numbers, and email addresses.

11. In other words, Cambridge Analytica used the Facebook Platform—in a way that Facebook and Zuckerberg encouraged—to influence and manipulate the outcome of a United 2

Email from Mark Zuckerberg to Sam Lessin (Nov. 19, 2012, 10:39 a.m.), available at https://dataviz.nbcnews.com/projects/20191104-facebook-leaked-documents/assets/facebook-sealed-exhibits.pdf.

States presidential election. The personal data of the more than 70 million U.S. Facebook users that Cambridge Analytica used to manipulate the election accounted for more than half the total votes during the 2016 presidential elections, in an election that was effectively decided by just a few hundred thousand people.

12. Though the data Cambridge Analytica (and many other companies like Media Matters, Fusion GPS, Black Cube and other attack services) used was supposedly private and protected from disclosure by Facebook’s privacy and data policies, Cambridge Analytica knew that it could access this trove of data using Facebook’s existing developer tools, an open secret that was well known to Facebook’s business partners using the platform. Cambridge Analytica also knew that it could leverage Facebook’s lax policy enforcement to continue manipulating the Facebook data it had amassed without fear Facebook would do anything about its operations. All the while, Facebook and Zuckerberg were trying to convince users in their user-facing statements that their data was safe.

13. The Cambridge Analytica revelations shocked the world, but it was no surprise to Facebook or Zuckerberg. Facebook had both a longstanding relationship with Cambridge Analytica and also actively encouraged companies like Cambridge Analytica to use the Facebook Platform to influence and manipulate consumer behavior.

14. What is most troubling is that Facebook looked into Cambridge Analytica and determined that it posed a risk to consumer data but chose to bury those concerns rather than stop them, as that could have hurt Facebook’s (and Zuckerberg’s) bottom line. Instead of coming clean, Facebook continued to help Cambridge Analytica win a United States presidential election.

15. While Facebook and Zuckerberg have, a full three years later, publicly condemned Cambridge Analytica’s data collection, its condemnation, in reality, only demonstrates that what Zuckerberg and Facebook say publicly is part of an intentional plan to mask the devastating consequences of their actions (or inactions).

16. Zuckerberg has said time and again that he and Facebook have a responsibility to protect users, and if they can’t, then they “don’t deserve to serve [them].”

17. Accordingly, the District brings this case to ensure that Mark Zuckerberg is held accountable for his role in Facebook violating the District’s consumer protection laws by misrepresenting the protection of user data and their blatant disregard and misuse of sensitive, personal data belonging to District residents.

The District of Columbia is a municipal corporation empowered to sue and be sued and is the local government for the territory constituting the permanent seat of the federal government. The District brings this case through the Attorney General for the District of Columbia, who is the chief legal officer for the District. The Attorney General is responsible for upholding the public interest and is also specifically authorized to enforce the District’s consumer protection laws, including the CPPA.

Mark Zuckerberg, FACEBOOK (Mar. 21, 2018), https://www.facebook.com/zuck/posts/i-want-to-share-an-update-on-the-cambridge-analytica-situation-including-the-ste/10104712037900071/ (last visited May 20, 2022)

Defendant Mark Zuckerberg is an individual residing in Palo Alto, California who regularly engages with Barack Obama, Dianne Feinstein, Nancy Pelosi, et al and their staff and political financiers and receives billions of dollars of profits from government ‘favored nations contracts.

Zuckerberg was, at all times material to this Complaint, Facebook’s co-founder, Chief Executive Officer, and a member of Facebook’s Board of Directors. Zuckerberg maintains his principal place of business at Facebook’s headquarters at 1 Hacker Way, Menlo Park, California, 94025.

Since 2012, Zuckerberg has served as Chairman of Facebook’s Board and controls approximately 60% of the voting shares. At all times material to this Complaint, acting alone or in concert with others, Zuckerberg was responsible for Facebook’s day-to-day operations, including the product strategy of Facebook, and formulated, directed, controlled, had the authority to control, participated in, or with knowledge, approved of the acts or practices of Facebook, including the acts and practices set forth in this Complaint. Zuckerberg engages in the business of supplying social networking services through the operation of his business, Facebook, which includes the product Facebook and maintains a website, www.facebook.com, and accompanying mobile applications, to consumers in D.C. Zuckerberg has traveled to Washington, D.C. on numerous occasions for matters related to the District’s allegations, including to testify on Facebook’s activities to various agencies and Congress.

Facebook’s Collection of Consumer Data for political Manipulaton

The Facebook website allows consumers to build a social network with other Facebook consumers and share information within that network. It is among the world’s most heavily trafficked websites and has over two billion active consumers around the globe.

Hundreds of thousands of D.C. residents are among Facebook’s consumers. Millions of Californians are Facebook Consumers.

To begin using the Facebook website, a consumer first creates a Facebook account. The consumer can then add other Facebook consumers as “friends” and by accumulating Facebook friends, the consumer builds a social network on the Facebook website.

As Facebook consumers grow their social networks and interact with friends on the Facebook website, their information and activity are digitally collected, recorded, and maintained by Facebook. Relevant here, this data can be divided into two broad categories: (i) data directly supplied by consumers, and (ii) data pertaining to consumers’ activity on and off the Facebook website.

First, consumers directly provide Facebook with personal information. To create a Facebook account, a consumer is required to supply Facebook with basic information such as their name, phone number, email address, birthday, and gender. A consumer then has the option to customize their “Facebook Profile” by supplying additional information to Facebook, such as their hometown, educational history, work experience, relationship status, political and religious views, and personal photographs. Facebook’s website is designed to encourage consumers to continue supplying information in the form of “Posts,” which are shared with that consumer’s 4

In this Complaint, the “Facebook website” refers to both (i) www.facebook.com, which is accessed through an Internet browser, and (ii) the Facebook mobile application, which is accessed through a mobile device like a smartphone or tablet. Many of Facebook’s features and services available on www.facebook.com are also available through the Facebook mobile application.

Posts include, but are not limited to, written statements, photographs and videos, links to websites, and “Check Ins” to geographic locations such as restaurants and bars.

Second, Facebook tracks and maintains data pertaining to consumer activity on its website, on other websites on the Internet, and even what they do offline. For example, Facebook records what advertisements are displayed to each consumer, as well as whether the consumer clicked on the advertisement. Facebook also tracks the date and time each consumer logs into their account, as well as the IP address, device, and browser they used to log in. Facebook also operates a companion mobile application called “Facebook Messenger,” which allows Facebook consumers to send and receive messages and make phone and video calls. For users of Facebook Messenger, Facebook maintains records of messages sent and received and the date and time of phone and video calls made.

Another example of consumer activity data that Facebook collects is a consumer’s “Likes,” one of Facebook’s signature innovations. It allows consumers to click on a

“thumbs-up” icon to Like a vast array of online content. Among other things, Facebook consumers can Like “Posts” made by other Facebook consumers, “Pages” maintained by non-individual entities, and content on external websites.

Facebook’s Like feature incentivizes increased activity on the Facebook website by allowing consumers to reward one another for sharing information—the more Posts a consumer makes, the more Likes they will receive. The Like also serves a broader function because over time, a consumer’s allocation of Likes reveals information about them—the friends they interact with most, the brands that catch their eye, and the issues with which they identify.

Facebook records and maintains each and every one of its consumers’ Likes.

Facebook generates much of its revenue by selling advertising space. Facebook relies on its collection of consumer data—and the personal information and preferences derived from each individual’s data—to sell targeted advertising space to marketers. Facebook’s business model primarily relies on using consumer data to provide advertisers the ability to run targeted ads to particular individuals and demographics. In other words, although Facebook supplies its social networking services free of direct monetary charge to consumers, in exchange, consumers provide Facebook with their personal data, which Facebook monetizes through the sale of targeted advertising.

The Facebook Platform and Third-Party Facebook Applications

In 2007, Facebook launched the Facebook “Platform,” an extensive software environment where third-party developers can build applications that interact with the Facebook website. The Facebook Platform includes various services and tools designed to assist third-party developers to create such applications.

Millions of third-party applications have been developed using the Facebook Platform and made available to Facebook consumers. Some applications are social, such as those that allow consumers to play games against other consumers within their social networks. Others are functional, allowing consumers to integrate information from their calendar and email accounts with their Facebook account.

The Facebook Platform facilitates integration between the Facebook website and third-party applications. For example, a third-party developer can allow Facebook consumers to access their application with a service available on the Facebook Platform called “Facebook Login.” Facebook Login allows a Facebook consumer to access an application directly by using 10 their Facebook account and login credentials (username and password). The Facebook Platform also harmonizes third-party applications’ look and feel with the Facebook website.

The Facebook Platform also includes an application program interface (“API”).

An API specifies how software components interact. In practical terms, Facebook’s website is built upon proprietary source code. The API refers to the code that Facebook makes available to third-party developers, which enables those developers to build applications for the Facebook website. Facebook’s API allows for a third-party application to interact with the Facebook website and governs the extent to which it can access Facebook’s vast collection of consumer data.

34.

In sum, the Facebook Platform was designed to allow for the development of third-party applications that would seamlessly engage with Facebook consumers while at the same time allowing those applications access to Facebook’s vast collection of consumer data.

The Cambridge Analytica Data Harvest

The Harvest of 87 Million Facebook Consumers’ Data

In November 2013, Aleksandr Kogan, a researcher affiliated with Cambridge University, and his company, Global Science Research (GSR), launched a third-party application on the Facebook Platform that identified itself as a personality study for research purposes. The application was called “thisisyourdigitallife” (the “App”) and ran on the Facebook Platform for over two years. The App appealed to Facebook consumers as a personality quiz and offered to generate a personality profile for consumers in exchange for downloading the App and granting access to some of the consumer’s Facebook data.

The App was presented to Facebook as a research tool to study psychological traits. At the time of the App’s launch, third-party applications could be launched on the 11 Facebook Platform without affirmative review or approval by Facebook. Accordingly, Facebook did not review the App before it was allowed on the Facebook Platform, nor did it verify its claim that the information it collected was for academic purposes.

At the time of the App’s launch, Facebook permitted applications to request permission to access a Facebook consumer’s personal data. Prior to installation, a Facebook consumer installing the App (an “App User”) was shown a screen that stated that the App would download some of the App User’s own Facebook data, including their name, gender, birthdate, Likes, and a list of Facebook friends.

To complete the installation, an App User clicked a Facebook Login icon on the information screen. The App was then installed through the Facebook Login service, using the App User’s Facebook login credentials.

Upon installation, the App harvested the personal information of the App User from Facebook’s collection of user data, including at least the App User’s name, gender, birthday, Likes, and list of Facebook friends.

In addition, the App also accessed data of the App User’s Facebook friends that the friend had shared with the App User. This data included at least the Facebook friend’s name, gender, birthdate, current city, and Likes. The vast majority of these Facebook friends never installed the App, never affirmatively consented to supplying the App with their data, and never knew the App had collected their data.

In early 2014, Facebook introduced changes to the Facebook Platform that (i) limited the data that applications could access, including data regarding the installing user’s friends, and (ii) instituted a review and approval process (called an “App Review”) for applications that sought to access data beyond what the updated Facebook Platform would allow.

In May 2014, Kogan applied to App Review to request access to consumer data beyond what the updated Facebook Platform would allow. In only a matter of days, Facebook rejected Kogan’s application on the basis that he was seeking information beyond the App’s stated research purposes. Nevertheless, the App was not audited nor further investigated.

During the time that the App ran on the Facebook Platform, approximately 290,000 Facebook consumers in the United States installed the App, including 852 consumers in D.C. Because the App was improperly allowed to harvest the personal data of App Users as well as App Users’ Facebook friends, approximately 87 million Facebook consumers had their information collected by the App, with more than 70 million in the United States—including over 340,000 District residents.

The Sale and Misuse of Consumer Data by Political Parties and Government Officials For Their Political Campaigns

In 2014, at a time when the App was fully operational on the Facebook Platform and harvesting consumer data, Kogan entered into an agreement with Cambridge Analytica for the sale of data collected by the App. Cambridge Analytica was a political consulting firm based in London, England that provided consulting services to candidates running for political office in the United States and abroad.

Kogan provided Cambridge Analytica with the personal data of the approximately 87 million Facebook consumers whose data was harvested, which included almost half of all D.C. residents. In exchange, Cambridge Analytica paid Kogan over $800,000.

Cambridge Analytica used the data it acquired from Kogan to, among other things, target digital political advertising during the 2016 United States Presidential Election (the

“2016 Election”). Cambridge Analytica received millions of dollars from the Ted Cruz presidential nomination campaign and later the Donald Trump presidential campaigns to provide digital advertising services during the 2016 Election.

At relevant times, Facebook had employees embedded within multiple presidential candidate campaigns who worked alongside employees from Cambridge Analytica.

Facebook knew, or should have known, that these presidential candidate campaigns and Cambridge Analytica were using the Facebook consumer data harvested by Kogan throughout the 2016 Election.

Facebook’s Lack of Oversight and Enforcement of Its Own Policies In Order To Help Corrupt Government Officials

By no later than December 11, 2015, Facebook knew that Kogan had sold Facebook consumer data to Cambridge Analytica. At that time, Facebook also knew that the collection and sale of consumer data violated its Platform Policy.

Facebook’s Platform Policy, which governed its relationship with third-party application developers throughout the App’s operation on the Facebook Platform, expressly prohibited the transfer and sale of consumer data accessed from Facebook. However, Facebook failed to exert meaningful review or compliance mechanisms to enforce its Platform Policy.

Indeed, the App itself contained terms that directly contradicted the Platform Policy, expressly stating that collected data could be used for commercial purposes. Nevertheless, Facebook did not take any action against the App and instead permitted it to harvest and sell Facebook consumers’ data without oversight for several years.

The Platform Policy also permitted Facebook to audit any applications on the Facebook Platform and to take other enforcement measures if it suspected that an application was violating the Platform Policy. In addition, the Platform Policy expressly provided several methods by which Facebook could enforce non-compliance with the Platform Policy. These audit provisions were largely unenforced.

In late December 2015, Facebook terminated the App’s access to the Facebook Platform. Nevertheless, Facebook did not ban, suspend, or limit the privileges of Kogan, Cambridge Analytica, or any of their affiliates, with respect to their access to the Facebook website or the Facebook Platform. Nor did Facebook conduct an audit of Kogan, Cambridge Analytica, or any of their affiliates, or take any other enforcement or remedial action to determine whether the Facebook consumer data that was harvested by the App had been accounted for, deleted, and protected from further use and sharing.

Instead, Facebook simply requested that Kogan and Cambridge Analytica delete all data that they received through the Facebook Platform and accepted their word that they had done so. Facebook did not take any additional steps to determine whether the harvested data was, in fact, accounted for and destroyed. And in fact, the data was not destroyed. It continued to be held and used by Cambridge Analytica through the 2016 Election and beyond. Facebook knew, or should have known, this fact from, among other sources, its employees embedded in presidential candidate campaigns during the 2016 Election who worked alongside Cambridge Analytica employees.

Facebook eventually required written certifications promising that the harvested data was accounted for and destroyed, but Facebook did not receive a certification from Kogan until June 2016 and did not receive a certification from Cambridge Analytica until April 2017.

Facebook further kept this data harvesting a secret from its users for years, waiting until April 2018 to finally disclose to its consumers that their personal information may have been harvested and sold to Cambridge Analytica.

Had Facebook and Zuckerberg disclosed in 2015 or 2016 the sale of Facebook consumer data to Cambridge Analytica, they would have provided consumers with timely material information about their use of the Facebook website. A disclosure that Facebook consumers’ data had been sold to a political consulting firm and was being used to target political advertising for the 2016 Election could have influenced Facebook consumers, including those in D.C. to, among other things, share less information on the Facebook website or deactivate their Facebook accounts. Rather than make such meaningful disclosures, Facebook and Zuckerberg instead profited from Kogan’s and Cambridge Analytica’s misuse of this stolen consumer data by selling millions of dollars of advertising space to Cambridge Analytica and presidential candidate campaigns during the 2016 Election.

Facebook knew of other third-party applications that similarly violated its Platform Policy through selling or improperly using consumer data. Facebook also failed to take reasonable measures to enforce its Platform Policy in connection with other third-party applications and failed to disclose to users when their data was sold or otherwise used in a manner inconsistent with Facebook’s policies.

Facebook’s Misleading Statements and Practices Regarding Third-Party Application Access to Consumer Data

Facebook’s failure to disclose third-party applications’ access to consumer data on its platform was compounded by the fact that the limited disclosures Facebook did make were ambiguous, misleading, and deceptive. These disclosures primarily are contained in two lengthy documents, a Terms of Service and Data Policy, that consumers must agree to in order to create a Facebook account. These documents together set out the general terms of use for the Facebook website and contain some statements regarding how third-party applications could access a consumer’s data. However, as shown by Facebook’s actions (and inactions) in connection with third parties, including the App and Cambridge Analytica, the representations made in these documents were misleading and deceptive.

For the duration of the App’s launch and operation on the Facebook Platform, Facebook’s Terms of Service represented that Facebook required applications to respect a Facebook consumer’s privacy. This representation, taken with Facebook’s public statements that it would protect consumers’ private information and its representations in the Platform Policy that it had the ability to audit applications and take enforcement measures against applications, gave consumers the impression that Facebook had implemented and maintained reasonable oversight and safeguards to protect consumers’ privacy.

These representations were misleading and deceptive, as demonstrated by Facebook’s lack of oversight and enforcement relating to third parties, such as the App. For example, Facebook failed to conduct meaningful oversight or enforcement of the App at several relevant times when it knew, or should have known, that the App was operating in violation of Facebook’s policies: (i) when the App was first launched on the Facebook Platform; (ii) after Facebook became aware, through its receipt and rejection of Kogan’s application through App Review, that the App was seeking consumer data to be used beyond the App’s stated research purpose; and (iii) after it learned that data collected by the App had been sold to Cambridge Analytica.

In addition, Facebook’s Data Policy also contained misrepresentations about third-party applications’ access to Facebook consumer data. From at least November 15, 2013 to at least January 30, 2015, the Data Policy provided that if an application asks permission from someone else to access your information, the application will be allowed to use that information only in connection with the person that gave the permission, and no one else. This representation was deceptive and misleading as demonstrated by Kogan’s use of the App to harvest consumer data, and then sell it to Cambridge Analytica. Facebook failed to implement and maintain 17

Reasonable oversight of applications operating on the Facebook Platform to safeguard consumers’ private data, and it knew or should have known that it did not have measures in place to control how applications used and/or shared data.

Facebook also misled its consumers generally about third-party applications’

access to their data. Facebook publicly represented that consumers controlled how their data is shared on the Facebook website. But as shown by the App, third-party applications that a Facebook consumer had never downloaded could still access their information through a Facebook friend who downloaded the App. The Facebook Platform thus afforded third parties an end-run to access consumer data, which third-party applications exploited. This was a material fact that Facebook failed to disclose, or failed to adequately disclose, to its consumers.

Adding to the potential customer confusion is the fact that consumers could not restrict third-party application access to their data through Facebook’s Privacy Settings, even though that is where a consumer would expect to have the ability to control how their data is shared. Instead, Facebook allocated privacy settings related to applications to a separate location under a separate Application Settings tab.

Through Privacy Settings, a consumer controls how their Facebook information is shared with other Facebook consumers. For example, a consumer can control what kinds of other Facebook users can view their account information. This can be manipulated to allow for sharing to all Facebook consumers (most expansive), only Facebook friends (the less expansive default), and a customized list of Facebook friends (the least expansive).

By contrast, through Application Settings, a consumer controls how their Facebook information is shared with third-party applications. There is a high potential for consumer confusion here. For example, from at least November 2013 through at least April 18 2014, even if a consumer restricted access to their information to only Facebook friends through their Privacy Settings, the information could still be accessible by any application that the consumer’s friends downloaded.

In sum, Facebook’s representations regarding consumer privacy in connection with applications were misleading and deceptive. Moreover, Facebook’s lack of adequate disclosures and multi-tiered privacy options added to consumer confusion regarding how consumer information was shared with applications.

Facebook’s representations to consumers that it will protect the privacy of consumers’ personal information, when, in fact, it did not implement or maintain reasonable privacy safeguards and failed to take reasonable measures in response to the harvesting and use of data by Cambridge Analytica, are misrepresentations of material facts that tend to mislead consumers.

Facebook’s representations to consumers that it requires applications and third-party developers to respect the privacy of consumers’ personal information, when, in fact, it did not implement or maintain reasonable oversight of third-party applications (such as conduct appropriate audits of applications), are misrepresentations of material facts that tend to mislead consumers.

Facebook’s representations to consumers that consumers’ agreements with third-party applications will control how those applications use consumer data, when, in fact, applications were able to collect and use consumer data without regard to those agreements, are misrepresentations of material facts that tend to mislead consumers.

Facebook’s failure to inform consumers, or to adequately inform consumers, that their personal information may be shared with third-party applications without their knowledge or affirmative consent, is a material fact, the omission of which tended to mislead consumers.

Facebook’s failure to tell consumers for over two years that their personal information was improperly harvested and sold by Kogan to Cambridge Analytica in violation of Facebook’s policies is a material fact, the omission of which tended to mislead consumers.

Facebook’s failure to explain to consumers whether and how they could control how information is shared with third-party applications and how to change privacy settings with respect to applications, and its representations that consumers can control how their information is shared, constitute ambiguities as to material facts that have the tendency to mislead consumers.

Mark Zuckerberg’s Control Over the Facebook Platform

Zuckerberg’s role and responsibility at Facebook

Zuckerberg is Facebook’s co-founder, CEO, and Chairman, and is inseparable from the company he founded. He continues to manage the company’s day-to-day operations––

at an exacting level of detail––including directing, participating in, and guiding the Facebook Platform.

He maintains an unparalleled level of control over the operations of Facebook as it has grown into the largest social media company in the world. In fact, throughout Facebook’s inception and explosive growth, Zuckerberg has retained control of the company. Zuckerberg currently controls almost 60% of Facebook’s voting shares and has the power to appoint a majority of the company’s board members.

Zuckerberg is not just a figurehead at Facebook; he is personally involved in nearly every major decision the company makes, and his level of influence is no secret. Dave 20

Arnold, a Facebook spokesman, recently touted Zuckerberg’s “active” leadership in the New York Times stating, “Mark has taken an active role in the leadership of Facebook from its founding through to today . . . We’re fortunate to have such engaged leaders, including Mark, Sheryl and the entire leadership team.”

Zuckerberg not only manages the company, but he also serves as the driving force behind its products. Early versions of the Facebook Platform––which promoted the “concept of openness and transparency as the high level ideal”––arose from Zuckerberg’s vision.6 According to Zuckerberg, Facebook “actually shifted a bit more of a focus not just on directly making it so people can use Facebook and share and be open on Facebook, but instead on making it so that the systems themselves have open properties.”

Zuckerberg has even gone as far as claiming that privacy is no longer a “social norm” because “[p]eople have really gotten comfortable not only sharing more information and different kinds, but more openly and with more people[.]”8

Zuckerberg’s vision of “open properties” directly led to developing the API tool called “Open Graph” that Facebook launched in April 2010 as Graph API version 1.0. Open Graph allowed consumers to personalize their user experience and “have instantly social connection”

IE:

Mike Isaac, Sheera Frenkel, and Cecilia Kang, Now More Than Ever, Facebook Is a

Mark Zuckerberg Production’, NEW YORK TIMES (May 16, 2020), https://www.nytimes.com/2020/05/16/technology/zuckerberg-facebook-coronavirus.html.

Fred Vogelstein, The Wired Interview: Facebook’s Mark Zuckerberg, WIRED (June 29, 2009), https://www.wired.com/2009/06/mark-zuckerberg-speaks/.

Id. (emphasis added).

Marshall Kirkpatrick, Facebook’s Zuckerberg Says The Age of Privacy Is Over, NEW

YORK TIMES (Jan. 10, 2020),

https://archive.nytimes.com/www.nytimes.com/external/readwriteweb/2010/01/10/10readwritew eb-facebooks-zuckerberg-says-the-age-of-privac-82963.html.

But to create these experiences, Facebook made information such as a person’s likes, dislikes, and interests instantly viewable to the person’s Facebook friends––and gave developers access to that personal data––including access to the data of a user’s Facebook friends, with little to no oversight.

On April 21, 2010, Zuckerberg publicly unveiled Graph API version 1.0 at Facebook’s annual F8 App Developer conference, saying: “We are building a web where the default is social.”10 For Zuckerberg, “[t]he thing [he] really care[d] about is the mission, making the world open.” 11

Shortly thereafter however, Facebook faced criticism over the growing concern from Facebook users and lawmakers that Graph API version 1.0 greatly expanded the scope of publicly available data to applications and developers by automatically allowing access to a user’s current city, hometown, education, work, likes, interests, and friends list without user consent.12

Zuckerberg responded by posting a public apology in the Washington Post acknowledging that “[s]ometimes we move too fast––and after listening to recent concerns,

IE:

Erick Schonfeld, Zuckerberg: “We are Building a Web Where the Default is Social” , TECHCRUNCH (Apr. 21, 2010), https://techcrunch.com/2010/04/21/zuckerbergs-buildin-web-default-social/.

Ryan Singel, Mark Zuckerberg: I Donated to Open Source, Facebook Competitor, WIRED (May 20, 2010), https://www.wired.com/2010/05/zuckerberg-interview/.

Jason Kincaid, Senators Call Out Facebook On ‘Instant Personalization’, Other Privacy Issues, TECHCRUNCH (Apr. 27, 2010), https://techcrunch.com/2010/04/27/senators-call-out-facebook-on-instant-personalization-other-privacy-

issues/?_ga=2.192578537.216944569.1650296804-940477689.1650296804.

The biggest message we have heard recently is that people want easier control over their information. Simply put, many of you thought our controls were too complex.

Our intention was to give you lots of granular controls; but that may not be what many of you wanted. We just missed the mark.

We have heard the feedback. There needs to be a simpler way to control your information. In the coming weeks, we will add privacy controls that are much simpler to use. We will also give you an easy way to turn off all third-party services.

Within Facebook, Zuckerberg directly oversaw the product development and engineering work that was exposing consumer data to abuse: Zuckerberg is in charge of the Facebook Platform and is able to overrule any decisions made by the company. But Zuckerberg’s involvement includes not only deciding just how open the Facebook Platform should be and what categories of user data should be shared with developers––he is also making day-to-day decisions about minute details of the Platform’s operations, including specific policy changes and enforcement decisions.

And contrary to his public statements, Zuckerberg was intent on finding a way of leveraging Platform changes to amass more data—and accordingly more money—for Facebook.

In November 2012, Zuckerberg unveiled his plan for Facebook’s data-sharing business model––which included giving apps and developers who spent money on Facebook or shared data back to Facebook––access to even more consumer data.

IE:

Mark Zuckerberg, From Facebook, answering privacy concerns with new settings, WASHINGTON POST (May 24, 2010), https://www.washingtonpost.com/wp-dyn/content/article/2010/05/23/AR2010052303828.html.

n a November 19, 2012, email to senior-level Facebook executives entitled “Platform Model Thoughts,” Zuckerberg discussed the value of “pulling non-app friends out of friends.get”––in other words making friends’ data available only through private APIs, which would further obfuscate a developers’ ability to access data from Facebook consumers:

After thinking about platform business for a long time, I wanted to send out a note explaining where I’m leaning on this. This isn’t final and we’ll have a chance to discuss this in person before we decide this for sure, but since this is complex, I wanted to write out my thoughts. This is long, but hopefully helpful. The quick summary is that I think we should go with full reciprocity and access to app friends for no charge. Full reciprocity means that apps are required to give any user who connects to FB a prominent option to share all of their social content within that service […] back to Facebook. […]

[W]e’re trying to enable people to share everything they want, and to do it on Facebook.

Sometimes the best way to enable people to share something is to have a developer build a special purpose app or network for that type of content and to make that app social by having Facebook plug into it. However, that may be good for the world but it’s not good for us unless people also share back to Facebook and that content increases the value of our network. So ultimately, I think the purpose of platform—even the read side—is to increase sharing back into Facebook.

[…]

It seems like we need some way to fast app switch to the FB app to show a dialog on our side that lets you select which of your friends you want to invite to an app. We need to make sure this experience actually is possible to build and make as good as we want, especially on iOS where we’re more constrained. We also need to figure out how we’re going to charge for it. I want to make sure this is explicitly tied to pulling non-app friends out of friends.get.

[…]

What I’m assuming we’ll do here is have a few basic thresholds of API usage and once you pass a threshold you either need to pay us some fixed amount to get to the next threshold or you get rate limited at the lower threshold.

[…]

Overall, I feel good about this direction. The purpose of platform is to tie the universe of all the social apps together so we can enable a lot more sharing and still remain the central social hub. I think this finds the right balance between ubiquity, reciprocity and profit.

Zuckerberg’s data-sharing business model underscores just how Facebook was able to get developers to use the Platform and why apps were able to amass so much Facebook consumer data: by leveraging access to friend data, Facebook’s “most valuable data.”

To carry out his business vision for Facebook, Zuckerberg oversaw changes to the Platform policies, including at the granular level of providing specific direction and assigning employees to the task. For example, in a November 19, 2012 email, Zuckerberg assigned specific employees, including former Harvard classmate Sam Lessin and longtime Facebook employee Javier Olivan, Head of International Growth, to the task of putting together a “real framework and process” for enforcement actions, giving them guidelines to create a process that is

“minimally intrusive/annoying” and “as simple as we can get away with.”

Zuckerberg also provided direct input on Facebook’s internal policies relating to data sharing on the Platform, and was so involved that he personally reviewed certain applications’ use of data. This was especially true for developers who Facebook deemed “competitors” and were therefore restricted from accessing certain categories of data, such as friend data, on the Platform. While making Facebook data generally available to developers was in Facebook’s interests, that decision cut the other way when it came to competitors.

IE:

Email from Mark Zuckerberg to Sam Lessin (Nov. 19, 2012, 10:39 a.m.), available at https://dataviz.nbcnews.com/projects/20191104-facebook-leaked-documents/assets/facebook-sealed-exhibits.pdf.

Email from Douglas Purdy to Constantin Koumouzelis, Marie Hagman, Zhen Fang, Eddie O’Neil, and George Lee (Apr. 10, 2013, 8:25 p.m.), available at https://dataviz.nbcnews.com/projects/20191104-facebook-leaked-documents/assets/facebook-sealed-exhibits.pdf.

Any changes on the Platform relating to these competitors would be escalated to Zuckerberg for approval and only permitted after his sign-off.

For example, in an April 2013 email, Doug Purdy, Facebook’s Director of Product Management, discussed an internal plan to implement certain API restrictions stating,

“We maintain a small list of strategic competitors that Mark personally reviewed. Apps produced by the companies on this list are subject to a number of restrictions outlined below. Any usage beyond that specified is not permitted without Mark level sign-off.”

Zuckerberg also instructed top-level executives to enforce Facebook’s “policies against competitors much more strongly.”20 In January 2013, Zuckerberg confirmed a new Facebook policy blocking competitors WeChat, Kakao, Line, and Google+ products from spending on ads or accessing friend data on the Platform, noting “[t]hose companies are trying to build social networks and replace us. The revenue is immaterial to us compared to any risk.”

Similarly, in a January 24, 2013 email, Zuckerberg gave the green light to Justin Osofsky, Facebook’s former Vice President of Global Operations, to “shut down” Vine, a popular video-sharing app’s access to the “friends API”––that app was owned by Facebook’s competitor Twitter, Inc.

Email from Mark Zuckerberg to Sam Lessin (Nov. 19, 2012, 10:39 a.m.), available at https://dataviz.nbcnews.com/projects/20191104-facebook-leaked-documents/assets/facebook-sealed-exhibits.pdf.

Email from Mark Zuckerberg to Javier Olivan, Sheryl Sandberg, Dan Rose, Mike Vernal, Justin Osofsky, Elliot Schrage, Sam Lessin, et. al. (Jan. 10, 2013, 1:28 a.m.), available at https://dataviz.nbcnews.com/projects/20191104-facebook-leaked-documents/assets/facebook-sealed-exhibits.pdf.

Email from Dan Rose to Mike Vernal, Justin Osofsky, Mark Zuckerberg, Kevin Systrom, Douglas Purdy and Dan Rose (Jan. 24, 2013, 12:21 p.m.), available at https://dataviz.nbcnews.com/projects/20191104-facebook-leaked-documents/assets/facebook-sealed-exhibits.pdf.

Zuckerberg was aware of the risks posed by an “Open” Platform but failed to act because he was told to control elections and he could not do so as well as he could if he followed the rules.

Zuckerberg was personally aware of the risks that sharing consumer data with apps posed, but actively disregarded those risks because sharing data was otherwise beneficial and lucrative to Facebook’s business model and Platform growth.

Even as early as 2012, Zuckerberg was aware of potential harms that might result from sharing consumer data. In an October 27, 2012, email concerning apps leaking personal data, Zuckerberg stated,

I’m getting more on board with locking down some parts of platform, including friends’

data and potentially email addresses for mobile apps. I’m generally skeptical that there is as much data leak strategic risk as you think. I agree there is clear risk on the advertiser side, but I haven’t figured out how that connects to the rest of the platform. I think we leak info to developers, but I just can’t think of any instances where that data has leaked from developer to developer and caused a real issue for us.

Despite being aware of risks on the Platform, Zuckerberg failed to take the necessary steps to protect consumer data––including the personal data of hundreds of thousands of District consumers.

Just months before the Cambridge Analytica data harvest became public, Zuckerberg admitted that his “personal challenge for 2018” was to fix abuse on the platform:

“we currently make too many errors enforcing our policies and preventing misuse of our tools.”

When details about Cambridge Analytica became public in March 2018, Zuckerberg publicly

IE:

Email from Mark Zuckerberg (Oct. 27, 2012, 6:06 a.m.), available at https://dataviz.nbcnews.com/projects/20191104-facebook-leaked-documents/assets/facebook-sealed-exhibits.pdf.

Mark Zuckerberg, FACEBOOK (Jan. 4, 2018),

https://www.facebook.com/zuck/posts/10104380170714571 (last visited May 20, 2022).

acknowledged the incident as a “breach of trust between Facebook and the people who share their data with us and expect us to protect it.”

Zuckerberg’s shift in tone and acknowledgment of Cambridge Analytica shows Facebook knew it could no longer hide behind its lax enforcement actions in the past. At Zuckerberg’s direction, Facebook shifted to promising future improvements to the security of the Platform, thereby hoping to obscure the company’s past harms and failures to adequately protect users’ privacy.

For example, despite Facebook’s ability to audit apps well before Cambridge Analytica became public in 2018, Zuckerberg vowed only after the breach became public to

“conduct a full audit of any app with suspicious activity” and promised to “ban any developer from our platform that does not agree to a thorough audit.”26 Zuckerberg also promised additional restrictions on developers’ access to user data and additional tools to help users more easily “revoke those apps’ permissions to [user] data.”

Zuckerberg has publicly taken personal responsibility for Facebook’s failures leading up the Cambridge Analytica incident as well. In April 2018 at a Congressional hearing relating to data privacy, Zuckerberg himself testified, “I started Facebook, I run it, and I’m responsible for what happens here.”28 In another statement in April 2018 concerning data privacy

IE:

Mark Zuckerberg, FACEBOOK (Mar. 21, 2018),

https://www.facebook.com/zuck/posts/10104712037900071 (last visited May 20, 2022).

Facebook CEO Mark Zuckerberg Hearing on Data Privacy and Protection, C-SPAN

(Apr. 10, 2018), https://www.c-span.org/video/?443543-1/facebook-ceo-mark-zuckerberg-testifies-data-protection (complete opening statement in Senate Hearing).

on the Platform, Zuckerberg stated, “We didn’t take a broad enough view of what our responsibility is, and that was a huge mistake. It was my mistake.”

In sum, Zuckerberg is broadly responsible for his Platform “vision” which required implementing an open Platform and exposing consumers’ personal data, was aware of the trade-off between the security of users’ data and Facebook’s own profits that these choices posed, and was directly responsible for Facebook’s lax enforcement standards and actions.

As CEO, Zuckerberg had the authority to control and direct––and had knowledge of––Facebook’s deceptive trade practices and misrepresentations to District consumers.

As such, Zuckerberg participated in, directed, managed, or otherwise knew of, and had the authority to control, Facebook’s inconsistent actions regarding privacy, including deceptive trade practices, misrepresentations, and ambiguities that violate the CPPA––namely: a.

Facebook’s representations to consumers that Facebook would protect the privacy of consumers’ personal information, when, in fact, it did not implement or maintain reasonable privacy safeguards and failed to take reasonable measures in response to the harvesting and use of data by Cambridge Analytica;

Facebook’s representations to consumers that it requires applications and third-party developers to respect the privacy of consumers’ personal information, when, in fact, it did not implement or maintain reasonable oversight of third-party applications (such as conduct appropriate audits of applications);

Facebook’s representations to consumers that consumers’ agreements with third-party applications will control how those applications use consumer data, when, in fact, applications were able to collect and use consumer data without regard to those agreements;

Facebook’s failure to inform consumers, or to adequately inform consumers, that their personal information may be shared with third-party

Hard Questions: Q&A With Mark Zuckerberg on Protecting People’s Information, FACEBOOK NEWSROOM (Apr. 4, 2018), https://about.fb.com/news/2018/04/hard-questions-protecting-peoples-information/ (emphasis added).

… applications without their knowledge or affirmative consent; e.

Facebook’s failure to tell consumers for over two years that their personal information was improperly harvested and sold by Kogan to Cambridge Analytica in violation of Facebook’s policies; and

Facebook’s failure to explain to consumers how to control how information is shared with third-party applications and how to change privacy settings with respect to applications, and its representations that consumers can control how their information is shared.

Count I: Violations of the Consumer Protection Procedures Act

 

The District incorporates by reference the foregoing allegations.

100. The CPPA is a remedial statute that is to be broadly construed. It establishes an enforceable right to truthful information from merchants about consumer goods and services that are or would be purchased, leased, or received in the District of Columbia.

101. The services that Facebook provides consumers are for personal, household, or family purposes and, therefore, are consumer goods and services.

102. Facebook, in the ordinary course of business, supplies consumer goods and services and, therefore, is a merchant under the CPPA.

103. Facebook users receive consumer goods and services from Facebook in the form of social networking services and, therefore, are consumers under the CPPA.

104. The CPPA prohibits unfair and deceptive trade practices in connection with the offer, sale, and supply of consumer goods and services.

105. Facebook’s representations to consumers, both express and implied, that it will protect the privacy of consumers’ personal information, that it requires applications and third-party developers to respect the privacy of consumers’ personal information, and that consumers’ agreement with third-party applications will control how those applications use consumer data, 30 are misrepresentations concerning material facts that have a tendency to mislead consumers and are unfair and deceptive trade practices that violate the CPPA, D.C. Code § 28-3904(e).

106. Facebook’s failure to disclose, or failure to adequately disclose, to consumers that their personal information may be shared with third-party applications without their knowledge or affirmative consent, is a material fact, the omission of which tended to mislead consumers and are unfair and deceptive trade practices that violate the CPPA, D.C. Code § 28-3904(f).

107. Facebook’s failure to disclose, or failure to adequately disclose, to consumers that their personal information was improperly harvested and used by third-party applications and others in violation of Facebook’s policies, such as in the Kogan and Cambridge Analytica example, is a material fact, the omission of which tended to mislead consumers and are unfair and deceptive trade practices that violate the CPPA, D.C. Code § 28-3904(f).

108. Facebook’s failure to explain to consumers how to control how information is shared with third-party applications and how to change privacy settings with respect to applications, as well as its representations to consumers, both express and implied, that it will protect the privacy of consumers’ personal information, that it requires applications and third-party developers to respect the privacy of consumers’ personal information, and that consumers’

agreement with third-party applications will control how those applications use consumer data, constitute ambiguities as to material facts that have the tendency to mislead consumers and are unfair and deceptive trade practices that violate the CPPA, D.C. Code § 28-3904(f-1).

109. At all times relevant to this Complaint, Defendant Mark Zuckerberg (i) was aware or should have been aware of Facebook’s data-sharing policy terms; (ii) possessed and/or exercised the authority to control the policies and practices of Facebook, Inc.; (iii) was responsible for creating and implementing the deceptive policies and trade practices of 31

Facebook, Inc. that are described in this Complaint; (iv) participated in the deceptive trade practices that are described in this Complaint; (v); directed, managed, or supervised, those employees at Facebook, Inc. who participated in the deceptive trade practices described in this Complaint; and (vi) knew or should have known of the deceptive trade practices that are described in this Complaint and had the power to stop them, but did not. “

While much that is revealed in the above is disturbing regarding the rights of Plaintiff, it is VERY disturbing for every citizen in the world. How could a company grow to become so vast, so evil, so manipulative of industry, elections and information unless corrupt government officials, THAT OWN FACEBOOK, GOOGLE, NETFLIX, TESLA, et al, were protecting and directing that company? It is IMPOSSIBLE for Facebook to have grown so large and so corrupt without Mark Zuckerberg exchanging compensation with government officials including those mentioned herein, who he knows, personally. The recordings between Barack Obama, David Plouffe, Nancy Pelosi, the Zuckerbergs and others reveal hard core conspiracy and is not a THEORY! The DC court evidence reveals hard core conspiracy and is not a THEORY! The leaked public officials emails and text messages reveal hard core conspiracy and is not a THEORY! Government officials used FACEBOOK, GOOGLE, NETFLIX, TESLA, et al, WHICH: 1.) THEY OWN; 2.) GET MONEY FROM; 3.) GET ELECTION MANIPULATION SERVICES FROM; 4.) THEY PARTY, DINE, HAVE SEX AND ORGANIZE WITH: AND ENGAGE IN HUNDREDS OF OTHER CONFLICTS OF INTEREST WITH!

In this, or another court trial, some Plaintiff is going to subpoena every bank and account record of FACEBOOK, GOOGLE, NETFLIX, TESLA, et al,. (FINCEN already has such records and could introduce them in lieu of a subpoena). Every wire transfer from ANY government account to FACEBOOK, GOOGLE, NETFLIX, TESLA, et al, will be scrutinized in detail, since 2000, and compared with what their competitors received during the same time frame. Wouldn’t it be interesting if certain public officials, who are Mark’s buddies, arranged to get contracts for Facebook, that nobody else got, while those officials were in office?

The bottom line is that there is now hard core forensic evidence on the books that proves that the government allowed its executives and officials to own, protect from regulation and operate illicit Silicon Valley operations that used their resources to harm Plaintiff, harm industry, harm society, harm elections and cause the deaths of children who were not immune to their Bullying-As-A-Service business models!