THE POLITICAL CORRUPTION EVIDENCE | FACEBOOK META CORRUPTION | Facebook has secret ratings for users like the Communist China political ratings system!

Facebook has secret ratings for users like the Communist China political ratings system!

Facebook has secret ratings for users like the Communist China political ratings system!


The social network is predicting your trustworthiness in a bid to fight conservatives

FACEBOOK is rating users based on how "trustworthy" it thinks they are.

Users receive a score on a scale from zero to one that determines if they have a good or bad reputation.

 Your Facebook usage is being monitored, and may be converted in a trustworthiness score
Your Facebook usage is being monitored, and may be converted in a trustworthiness score

The rating system was revealed in a report by the Washington Post, which says it's in place to "help identify malicious actors".

Facebook tracks your behaviour across its site, and uses that info to assign you a rating.

Tessa Lyons, who heads up Facebook's fight against fake news, said: "One of the signals we use is how people interact with articles.

"For example, if someone previously gave us feedback that an article was false and the article was confirmed false by a fact-checker, then we might weight that person’s future false news feedback more than someone who indiscriminately provides false news feedback on lots of articles, including ones that end up being rated as true."

 Facebook can see everything you do on the site – which helps build a highly detailed picture of who you are
Facebook can see everything you do on the site – which helps build a highly detailed picture of who you are

Earlier this year, Facebook admitted it was rolling out trust ratings for media outlets.

This involved ranking news websites based on the quality of the news they were reporting.

This rating would then be used to decide which posts should be promoted higher in users' News Feeds.

It's not clear exactly what users' ratings are for, but it's possible they may be used in a similar way.

But Facebook hasn't revealed exactly how ratings are decided, or whether all users have a rating.

Who is Mark Zuckerberg, the founder of Facebook?

Here's what you need to know...

  • Mark Zuckerberg is the chairman, CEO and co-founder of social networking giant Facebook
  • Born in New York in 1984, Zuckerberg already had a "reputation as a programming prodigy" when he started college
  • While at Harvard, Zuckerberg launched a site called Face Mash, on which students ranked the attractiveness of their classmates
  • Harvard shut the site down after its popularity crashed a network and Zuckerberg later apologised saying it was "completely improper"
  • The following term he began working on an early version of Facebook
  • The 33-year-old launched the social network from his dorm room on February 4, 20o4 with the help of fellow students
  • The friends would end up embroiled in legal disputes as they challenged Zuckerberg for shares in the company
  • Zuckerberg also faced action from Cameron and Tyler Winklevoss, as well as Divya Narendra who claimed he had stolen their idea - the disagreement was later turned into the film, The Social Network
  • The tech prodigy dropped out of Harvard to focus on Facebook, but received an honorary degree in 2017
  • Speaking about the site to Wired magazine in 2010 he said: "The thing I really care about is the mission, making the world open"
  • By 2012 Facebook had one billion users. By June 2017 it had reached two billion users every month

According to Lyons, a user's rating "isn't meant to be an absolute indicator of a person's credibility".

Instead, it's intended as a measurement of working out how risky a user's actions may be.

It's Facebook's latest bid to tackle fake news, a growing problem for the social network.

The site, which sees 2.23billion users log on every single month, has become a hot-bed for falsified news coverage.

Earlier this year, billionaire Facebook boss Mark Zuckerberg vowed to fight fake news.

"The world feels anxious and divided, and Facebook has a lot of work to do," the 34-year-old Harvard drop-out explained.

Mark Zuckerberg apologises for data breach by says he's 'sure someone's trying' to use Facebook to meddle with US mid-term election

Facebook has admitted that its site has been the subject of political fakery campaigns from Russia.

After initially denying any complacency on its part, the social network admitted more than 126 million US users had viewed some form of Russian propaganda.

A congressional hearing followed, with Facebook, Twitter, and Google in the dock.

And Facebook's been grappling with the problem ever since.

Speaking in January, Samidh Chakrabarti, who heads up civic engagement at Facebook, said: "Even a handful of deliberately misleading stories can have dangerous consequences.

"We're committed to this issue of transparency because it goes beyond Russia.

"Without transparency, it can be hard to hold politicians accountable for their own words.

"Democracy then suffers because we don't get the full picture of what our leaders are promising us," he wrote, in what looks like a subtle snipe at US President Donald Trump.

"This is an even more pernicious problem than foreign interference.

"But we hope that by setting a new bar for transparency, we can tackle both of these challenges simultaneously."

Chakrabarti said that the misinformation campaigns targeting Facebook users are "professionalised, and constantly try to game the system".

"We will always have more work to do," he added.

We've asked Facebook for comment and will update this story with any response.

Do you think Facebook is right to rate its users' trustworthiness? Let us know in the comments!

Who needs democracy when you have data?

Here’s how China rules using data, AI, and internet surveillance.

  • by Christina Larson
  • August 20, 2018
  • People in Beijing are always under the watchful eye of Mao—and myriad surveillance cameras.

In 1955, science fiction writer Isaac Asimov published a short story about an experiment in “electronic democracy,” in which a single citizen, selected to represent an entire population, responded to questions generated by a computer named Multivac. The machine took this data and calculated the results of an election that therefore never needed to happen. Asimov’s story was set in Bloomington, Indiana, but today an approximation of Multivac is being built in China.

For any authoritarian regime, “there is a basic problem for the center of figuring out what’s going on at lower levels and across society,” says Deborah Seligsohn, a political scientist and China expert at Villanova University in Philadelphia. How do you effectively govern a country that’s home to one in five people on the planet, with an increasingly complex economy and society, if you don’t allow public debate, civil activism, and electoral feedback? How do you gather enough information to actually make decisions? And how does a government that doesn’t invite its citizens to participate still engender trust and bend public behavior without putting police on every doorstep?

Hu Jintao, China’s leader from 2002 to 2012, had attempted to solve these problems by permitting a modest democratic thaw, allowing avenues for grievances to reach the ruling class. His successor, Xi Jinping, has reversed that trend. Instead, his strategy for understanding and responding to what is going on in a nation of 1.4 billion relies on a combination of surveillance, AI, and big data to monitor people’s lives and behavior in minute detail.

It helps that a tumultuous couple of years in the world’s democracies have made the Chinese political elite feel increasingly justified in shutting out voters. Developments such as Donald Trump’s election, Brexit, the rise of far-right parties across Europe, and Rodrigo Duterte’s reign of terror in the Philippines underscore what many critics see as the problems inherent in democracy, especially populism, instability, and precariously personalized leadership.

Since becoming general secretary of the Chinese Communist Party in 2012, Xi has laid out a raft of ambitious plans for the country, many of them rooted in technology—including a goal to become the world leader in artificial intelligence by 2030. Xi has called for “cyber sovereignty” to enhance censorship and assert full control over the domestic internet. In May, he told a meeting of the Chinese Academy of Sciences that technology was the key to achieving “the great goal of building a socialist and modernized nation.” In January, when he addressed the nation on television, the bookshelves on either side of him contained both classic titles such as Das Kapital and a few new additions, including two books about artificial intelligence: Pedro Domingos’s The Master Algorithm and Brett King’sAugmented: Life in the Smart Lane.

“No government has a more ambitious and far-­reaching plan to harness the power of data to change the way it governs than the Chinese government,” says Martin Chorzempa of the Peterson Institute for International Economics in Washington, DC. Even some foreign observers, watching from afar, may be tempted to wonder if such data-driven governance offers a viable alternative to the increasingly dysfunctional­looking electoral model. But over-­relying on the wisdom of technology and data carries its own risks.

Data instead of dialogue

Chinese leaders have long wanted to tap public sentiment without opening the door to heated debate and criticism of the authorities. For most of imperial and modern Chinese history, there has been a tradition of disgruntled people from the countryside traveling to Beijing and staging small demonstrations as public “petitioners.” The thinking was that if local authorities didn’t understand or care about their grievances, the emperor might show better judgment.

Under Hu Jintao, some members of the Communist Party saw a limited openness as a possible way to expose and fix certain kinds of problems. Blogs, anticorruption journalists, human-rights lawyers, and online critics spotlighting local corruption drove public debate toward the end of Hu’s reign. Early in his term, Xi received a daily briefing of public concerns and disturbances scraped from social media, according to a former US official with knowledge of the matter. In recent years, petitioners have come to the capital to draw attention to scandals such as illegal land seizures by local authorities and contaminated milk powder.

But police are increasingly stopping petitioners from ever reaching Beijing. “Now trains require national IDs to purchase tickets, which makes it easy for the authorities to identify potential ‘troublemakers’ such as those who have protested against the government in the past,” says Maya Wang, senior China researcher for Human Rights Watch. “Several petitioners told us they have been stopped at train platforms.” The bloggers, activists, and lawyers are also being systematically silenced or imprisoned, as if data can give the government the same information without any of the fiddly problems of freedom.

Photo of a facial recognition software applied to video footage of shoppers.
A Shanghai startup’s demo of its system for facial recognition.

The idea of using networked technology as a tool of governance in China goes back to at least the mid-1980s. As Harvard historian Julian Gewirtz explains, “When the Chinese government saw that information technology was becoming a part of daily life, it realized it would have a powerful new tool for both gathering information and controlling culture, for making Chinese people more ‘modern’ and more ‘governable’—which have been perennial obsessions of the leadership.” Subsequent advances, including progress in AI and faster processors, have brought that vision closer.

As far as we know, there is no single master blueprint linking technology and governance in China. But there are several initiatives that share a common strategy of harvesting data about people and companies to inform decision-making and create systems of incentives and punishments to influence behavior. These initiatives include the State Council’s 2014 “Social Credit System,” the 2016 Cybersecurity Law, various local-level and private-enterprise experiments in “social credit,” “smart city” plans, and technology-driven policing in the western region of Xinjiang. Often they involve partnerships between the government and China’s tech companies.

The most far-reaching is the Social Credit System, though a better translation in English might be the “trust” or “reputation” system. The government plan, which covers both people and businesses, lists among its goals the “construction of sincerity in government affairs, commercial sincerity, and judicial credibility.” (“Everybody in China has an auntie who’s been swindled. There is a legitimate need to address a breakdown in public trust,” says Paul Triolo, head of the geotechnology practice at the consultancy Eurasia Group.) To date, it’s a work in progress, though various pilots preview how it might work in 2020, when it is supposed to be fully implemented.

Blacklists are the system’s first tool. For the past five years, China’s court system has published the names of people who haven’t paid fines or complied with judgments. Under new social-credit regulations, this list is shared with various businesses and government agencies. People on the list have found themselves blocked from borrowing money, booking flights, and staying at luxury hotels. China’s national transport companies have created additional blacklists, to punish riders for behavior like blocking train doors or picking fights during a journey; offenders are barred from future ticket purchases for six or 12 months. Earlier this year, Beijing debuted a series of blacklists to prohibit “dishonest” enterprises from being awarded future government contracts or land grants.

A few local governments have experimented with social-credit “scores,” though it’s not clear if they will be part of the national plan. The northern city of Rongcheng, for example, assigns a score to each of its 740,000 residents, Foreign Policy reported. Everyone begins with 1,000 points. If you donate to a charity or win a government award, you gain points; if you violate a traffic law, such as by driving drunk or speeding through a crosswalk, you lose points. People with good scores can earn discounts on winter heating supplies or get better terms on mortgages; those with bad scores may lose access to bank loans or promotions in government jobs. City Hall showcases posters of local role models, who have exhibited “virtue” and earned high scores.

“The idea of social credit is to monitor and manage how people and institutions behave,” says Samantha Hoffman of the Mercator Institute for China Studies in Berlin. “Once a violation is recorded in one part of the system, it can trigger responses in other parts of the system. It’s a concept designed to support both economic development and social management, and it’s inherently political.” Some parallels to parts of China’s blueprint already exist in the US: a bad credit score can prevent you from taking out a home loan, while a felony conviction suspends or annuls your right to vote, for example. “But they’re not all connected in the same way—there’s no overarching plan,” Hoffman points out.

One of the biggest concerns is that because China lacks an independent judiciary, citizens have no recourse for disputing false or inaccurate allegations. Some have found their names added to travel blacklists without notification after a court decision. Petitioners and investigative journalists are monitored according to another system, and people who’ve entered drug rehab are watched by yet a different monitoring system. “Theoretically the drug-user databases are supposed to erase names after five or seven years, but I’ve seen lots of cases where that didn’t happen,” says Wang of Human Rights Watch. “It’s immensely difficult to ever take yourself off any of these lists.”

Occasional bursts of rage online point to public resentment. News that a student had been turned down by a college because of her father’s inclusion on a credit blacklist recently lit a wildfire of online anger. The college’s decision hadn’t been officially sanctioned or ordered by the government. Rather, in their enthusiasm to support the new policies, school administrators had simply taken them to what they saw as the logical conclusion.

The opacity of the system makes it difficult to evaluate how effective experiments like Rongcheng’s are. The party has squeezed out almost all critical voices since 2012, and the risks of challenging the system—even in relatively small ways—have grown. What information is available is deeply flawed; systematic falsification of data on everything from GDP growth to hydropower use pervades Chinese government statistics. Australian National University researcher Borge Bakken estimates that official crime figures, which the government has a clear incentive to downplay, may represent as little as 2.5 percent of all criminal behavior.

In theory, data-driven governance could help fix these issues—circumventing distortions to allow the central government to gather information directly. That’s been the idea behind, for instance, introducing air-quality monitors that send data back to central authorities rather than relying on local officials who may be in the pocket of polluting industries. But many aspects of good governance are too complicated to allow that kind of direct monitoring and instead rely on data entered by those same local officials.

However, the Chinese government rarely releases performance data that outsiders might use to evaluate these systems. Take the cameras that are used to identify and shame jaywalkers in some cities by projecting their faces on public billboards, as well as to track the prayer habits of Muslims in western China. Their accuracy remains in question: in particular, how well can facial-recognition software trained on Han Chinese faces recognize members of Eurasian minority groups? Moreover, even if the data collection is accurate, how will the government use such information to direct or thwart future behavior? Police algorithms that predict who is likely to become a criminal are not open to public scrutiny, nor are statistics that would show whether crime or terrorism has grown or diminished. (For example, in the western region of Xinjiang, the available information shows only that the number of people taken into police custody has shot up dramatically, rising 731 percent from 2016 to 2017.)

Photo of a large monitor in a busy intersection showing images of a suspect.
In the city of Xiangyang, cameras linked to face-recognition technology project photos of jaywalkers, with names and ID numbers, on a billboard.

“It’s not the technology that created the policies, but technology greatly expands the kinds of data that the Chinese government can collect on individuals,” says Richard McGregor, a senior fellow at the Lowy Institute and the author of The Party: The Secret World of China’s Communist Rulers. “The internet in China acts as a real-time, privately run digital intelligence service.”

Algorithmic policing

Writing in the Washington Post earlier this year, Xiao Qiang, a professor of communications at the University of California, Berkeley, dubbed China’s data-enhanced governance “a digital totalitarian state.” The dystopian aspects are most obviously on display in western China.

Xinjiang (“New Territory”) is the traditional home of a Chinese Muslim minority known as Uighurs. As large numbers of Han Chinese migrants have settled in—some say “colonized”—the region, the work and religious opportunities afforded to the local Uighur population have diminished. One result has been an uptick in violence in which both Han and Uighur have been targeted, including a 2009 riot in the capital city of Urumqi, when a reported 200 people died. The government’s response to rising tensions has not been to hold public forums to solicit views or policy advice. Instead, the state is using data collection and algorithms to determine who is “likely” to commit future acts of violence or defiance.

The Xinjiang government employed a private company to design the predictive algorithms that assess various data streams. There’s no public record or accountability for how these calculations are built or weighted. “The people living under this system generally don’t even know what the rules are,” says Rian Thum, an anthropologist at Loyola University who studies Xinjiang and who has seen government procurement notices that were issued in building the system.

In the western city of Kashgar, many of the family homes and shops on main streets are now boarded up, and the public squares are empty. When I visited in 2013, it was clear that Kashgar was already a segregated city—the Han and Uighur populations lived and worked in distinct sections of town. But in the evenings, it was also a lively and often noisy place, where the sounds of the call to prayer intermingled with dance music from local clubs and the conversations of old men sitting out late in plastic chairs on patios. Today the city is eerily quiet; neighborhood public life has virtually vanished. Emily Feng, a journalist for the Financial Times, visited Kashgar in June and posted photos on Twitter of the newly vacant streets.

The reason is that by some estimates more than one in 10 Uighur and Kazakh adults in Xinjiang have been sent to barbed-wire-ringed “reeducation camps”—and those who remain at large are fearful.

In the last two years thousands of checkpoints have been set up at which passersby must present both their face and their national ID card to proceed on a highway, enter a mosque, or visit a shopping mall. Uighurs are required to install government-­designed tracking apps on their smartphones, which monitor their online contacts and the web pages they’ve visited. Police officers visit local homes regularly to collect further data on things like how many people live in the household, what their relationships with their neighbors are like, how many times people pray daily, whether they have traveled abroad, and what books they have.

All these data streams are fed into Xinjiang’s public security system, along with other records capturing information on everything from banking history to family planning. “The computer program aggregates all the data from these different sources and flags those who might become ‘a threat’ to authorities,” says Wang. Though the precise algorithm is unknown, it’s believed that it may highlight behaviors such as visiting a particular mosque, owning a lot of books, buying a large quantity of gasoline, or receiving phone calls or email from contacts abroad. People it flags are visited by police, who may take them into custody and put them in prison or in reeducation camps without any formal charges.

Photo of people tapping ID cards on a device before entering a building.
Visitors to Tiananmen Square in Beijing scan their IDs at a checkpoint.

Adrian Zenz, a political scientist at the European School of Culture and Theology in Korntal, Germany, calculates that the internment rate for minorities in Xinjiang may be as high as 11.5 percent of the adult population. These camps are designed to instill patriotism and make people unlearn religious beliefs. (New procurement notices for cremation security guards seem to indicate that the government is also trying to stamp out traditional Muslim burial practices in the region.)

While Xinjiang represents one draconian extreme, elsewhere in China citizens are beginning to push back against some kinds of surveillance. An internet company that streamed closed-circuit TV footage online shut down those broadcasts after a public outcry. The city of Shanghai recently issued regulations to allow people to dispute incorrect information used to compile social-credit records. “There are rising demands for privacy from Chinese internet users,” says Samm Sacks, a senior fellow in the Technology Policy Program at CSIS in New York. “It’s not quite the free-for-all that it’s made out to be.”

Christina Larson is an award-winning foreign correspondent and science journalist, writing mostly about China and Asia.

Gallery RSS RSS Feed | Archive View | Powered by Zenphoto | rotections: Public Domain. Non-Commercial. Fair Use. Freedom of The Press. No Tracking Of Public Allowed. First Amendment Protections, SLAPP, UN Protected. GDPR Compliant. Section 203 protected. Privacy Tools At:, ACLU, ICIJ- supported. If you sue us to try to hide and censor the news, you are allowing us to bypass the demurrer process, and we will counter-sue you for RICO, Anti-trust, Political Bribery, Sex Trafficking, Interference, First Amendment and your other crimes, which we have FBI-grade evidence for! Bring it on corrupt Tesla, Google, Facebook, Youtube, Netflix! We might even get DOJ and/or FTC to partner with us (again) to take your filthy corrupt companies down!....REAL NEWS does not have ads in it. Any news source, with ads in it, is fake news manipulated by the advertisers!....DOWNLOAD AND COPY THIS NEWS SITE. USE ANY FREE SERVER SPACE YOU FIND ON THE WEB. MAKE YOUR OWN DIGITAL NEWSPAPER. JOIN THE HUNDREDS OF THOUSANDS OF FREE NEWS SITES, LIKE THIS, AROUND THE WORLD AND DELIVER THE NEWS AND DEMOCRACY. PLEASE FOLLOW THE WIKIPEDIA RULES FOR POSTING. BE THE NEWS!