October 25, 2021

The Facebook Papers and their fallout.


By The New York Times

For weeks, Facebook has been shaken by revelations that have ignited a firestorm of criticism from lawmakers, regulators and the public.

Reports by The Wall Street Journal from research documents provided by a whistle-blower put Facebook under a microscope. Those reports showed how Facebook knew Instagram was worsening body image issues among teenagers, among other issues.

Putting Out Fires


The whistle-blower goes public.

The whistle-blower, Frances Haugen, went public during an interview on “60 Minutes” in early October. On Oct. 5, Ms. Haugen testified before a Senate subcommittee for more than three hours. She said Facebook had purposely hidden disturbing research about how teenagers felt worse about themselves after using its products and how it was willing to use hateful content on its site to keep users coming back. In her testimony, she encouraged lawmakers to demand more documents and internal research, suggesting the documents she had provided were just the tip of the iceberg.

The Whistle-Blower Testifies


Facebook Whistle-Blower Urges Lawmakers to Regulate the Company
Oct. 5, 2021



After Ms. Haugen testified, executives publicly questioned her credibility and called her accusations untrue. But internally, they tried to position their stances to hang on to the good will of more than 63,000 employees and assuage their concerns.

A Rift Inside Facebook



After Whistle-Blower Goes Public, Facebook Tries Calming Employees
Oct. 10, 2021


Leaked documents reveal internal struggles.

Reporters have since covered more internal documents from the company, which owns Instagram and WhatsApp in addition to the core Facebook social network. Documents about Instagram, for instance, reveal a company that is struggling with retaining, engaging and attracting young users.

Fighting for Teenage Users


Instagram Struggles With Fears of Losing Its ‘Pipeline’: Young Users
Oct. 16, 2021


Other documents raise questions about Facebook’s role in election misinformation and the pro-Trump attack on the Capitol on Jan. 6. Company documents show the degree to which Facebook knew of extremist movements and groups on its site that were trying to polarize American voters before the election. Employees believed Facebook could have done more, according to the documents.

Misinformation and a Riot


Internal Alarm, Public Shrugs: Facebook’s Employees Dissect Its Election Role
Oct. 22, 2021


The challenges are global.


In India, Facebook’s biggest market, the problems are bigger, too. Internal documents show a struggle with misinformation, hate speech and celebrations of violence. Dozens of studies and memos written by Facebook employees provide stark evidence of one of the most serious criticisms levied by human rights activists and politicians against the world-spanning company: It moves into a country without fully understanding its potential impact on local culture and politics, and fails to deploy the resources to act on issues once they occur.

Hate Speech and Violence in India


In India, Facebook Grapples With an Amplified Version of Its Problems
Oct. 23, 2021



The latest revelations, published on Monday morning, show internal research that undercuts the heart of social networking — “likes” and sharing — that Facebook revolutionized. According to the documents, researchers determined over and over that people misused key features or that those features amplified toxic content, among other effects. In an August 2019 internal memo, several researchers said it was Facebook’s “core product mechanics” — meaning the basics of how the product functioned — that had let misinformation and hate speech flourish on the site.

The Trouble With Likes


Facebook Wrestles With the Features It Used to Define Social Networking
Oct. 25, 2021



Without government-mandated transparency, Facebook can present a false picture of its efforts to address hate speech and other extreme content,Ms. Haugen told Britain’s Parliament. The company says artificial intelligence software catches more than 90 percent of hate speech, but Ms. Haugen said the number was less than 5 percent.

“They are very good at dancing with data,” she said.

The Facebook Papers: What you need to know about the trove of insider documents

BILL CHAPPELL

Facebook's rank-and-file employees warned their leaders about the company's effects on society and politics in the U.S. — and its inability to effectively moderate content in other countries magnified those dangers. Those are two of the main takeaways from thousands of internal Facebook documents that NPR and other news outlets have reviewed.

The documents, known collectively as the Facebook Papers, were shared in redacted form with Congress after whistleblower Frances Haugen, a former Facebook product manager, disclosed them to the Securities and Exchange Commission.

Haugen alleges that the trove of statements and data prove that Facebook's leaders have repeatedly and knowingly put the company's image and profitability ahead of the public good — even at the risk of violence and other harm.


TECHNOLOGY

Here are 4 key points from the Facebook whistleblower's testimony on Capitol Hill


Some of the internal documents initially emerged last month in The Wall Street Journal. They include internal research findings and internal audits that the company performed on its own practices.

Here are four main takeaways from news outlets' review of the documents:
Facebook employees hotly debated its policies, especially after Jan. 6

When then-President Donald Trump's supporters mounted an insurrection at the U.S. Capitol on Jan. 6, Facebook rushed to take technical measures that aimed to clamp down on misinformation and content that might incite further violence. The next day, it banned Trump from the platform — at first temporarily, but then permanently.


In the weeks leading up to the violence, Facebook worked to defuse vitriol and conspiracy theories from Trump voters who refused to accept his defeat. As NPR's Shannon Bond and Bobby Allyn have reported, the company repeatedly shut down groups affiliated with the Stop the Steal movement. But those groups were attracting hundreds of thousands of users, and Facebook was unable to keep pace as the conspiracy theorists regrouped.

The post-election turmoil put a quick end to the relief that many at Facebook felt on Nov. 3, when the U.S. election played out mostly peacefully and without inklings of foreign meddling.

But then came Jan. 6 — and as the assault on the Capitol riveted and horrified audiences in the U.S. and elsewhere, Facebook employees aired their frustration and anger.


UNTANGLING DISINFORMATION

How the 'Stop the Steal' movement outwitted Facebook ahead of the Jan. 6 insurrection

"We've been fueling this fire for a long time and we shouldn't be surprised it's now out of control," one employee wrote on an internal message board, the documents show.

"Hang in there everyone," Mike Schroepfer, Facebook's chief technology officer, wrote on a message board, calling for calm as he explained the company's approach to the riot, according to the documents.

In response to Schroepfer's message, Facebook employees said it was too little too late.

"I came here hoping to effect change and improve society, but all I've seen is atrophy and abdication of responsibility," one commenter said, according to the documents.

In a statement to NPR, Facebook spokesman Andy Stone said Facebook did not bear responsibility for the Capitol siege.

"The responsibility for the violence that occurred on January 6 lies with those who attacked our Capitol and those who encouraged them," Stone said.
Content standards were contorted, often out of fear of riling high-profile accounts

One of the earliest revelations from the internal documents is the detail they provide about Facebook's separate set of content standards for high-profile accounts, such as those for Trump, or for celebrities.

During Trump's presidency, he regularly made false and inflammatory statements about a wide range of matters. But only a small handful were removed by Facebook, as when the then-president made dangerous claims like saying COVID-19 was less dangerous than the flu or stating that children were "almost immune from this disease."

Facebook has previously defended its approach to such controversial and misleading statements, saying politicians like Trump should be allowed to say what they believe so the public knows what they think. Facebook CEO Mark Zuckerberg has also repeatedly insisted that Facebook is merely a platform, not the "arbiter of truth."


TECHNOLOGY

Oversight Board slams Facebook for giving special treatment to high-profile users


But the documents suggest Facebook's policy of treating influential people differently — codified in a VIP system called XCheck — was created in large part to prevent a public relations backlash from celebrities and other high-profile users.

The entire premise of the XCheck system, the Journal's Jeff Horwitz told NPR in September, "is to never publicly tangle with anyone who is influential enough to do you harm."

Facebook's own Oversight Board sharply criticized the program last week, saying the company has not been forthcoming enough about its varying standards for content moderation.

A Facebook spokesperson told NPR in a statement that the company asked the board to review the program because it aims "to be clearer in our explanations to them going forward."
Young people see Facebook content as "boring, misleading, and negative"

For much of the past decade, senior citizens have been the fastest-growing U.S. demographic on Facebook — a dramatic turnabout for a company whose founding mystique rests on the image of a hoodie-wearing coder creating a space for college kids to connect on.

During the same timespan, Facebook has seen younger people become less likely to join the site. It's a worrying trend for the company — Facebook insiders got an update on that trend this year, in an internal presentation that is reflected in the documents.

"Most young adults perceive Facebook as a place for people in their 40s and 50s," the company's researchers said, according to The Verge. "Young adults perceive content as boring, misleading, and negative. They often have to get past irrelevant content to get to what matters."


BUSINESS

Ex-Facebook employee says company has known about disinformation problem for years


Along with that stumbling block, young users were found to have negative views of Facebook due to privacy concerns and its potential "impact to their wellbeing," The Verge reports.

Haugen previously leaked a Facebook study that found that 13.5% of British teen girls in a survey said their suicidal thoughts became more frequent after they joined Instagram.

In addition to its namesake platform, Facebook owns Instagram and WhatsApp.

"It is clear that Facebook prioritizes profit over the well-being of children and all users," Sen. Marsha Blackburn, R-Tenn., said during a Senate hearing this month in which Haugen testified.
Facebook's global reach exceeds its grasp

While much of the focus on Facebook in the U.S. has to do with its role in enabling and intensifying political divisions, the documents also fault the company for its activities in numerous other countries.

The documents portray Facebook as failing to deal with a number of social and language complexities stemming from its more than 2.8 billion users worldwide. The results have been especially dangerous and harmful in countries where unrest or rights abuses are common, the documents state.

"Two years ago, Apple threatened to pull Facebook and Instagram from its app store over concerns about the platform being used as a tool to trade and sell maids in the Mideast," The Associated Press reports.

The company routinely struggles with posts and comments in Arabic, both on its main platform and on Instagram, according to the documents. Arabic is one of the world's most widely spoken languages, but its many dialects are highly distinct from each another.

Facebook "doesn't have anyone who can speak most of them or can understand most of them in terms of sort of the vernacular," Horwitz told NPR. "And it also doesn't have a system to route content in those dialects to the right people."

The problem extends beyond Arabic and has a wide range of effects.

"In countries like Afghanistan and Myanmar, these loopholes have allowed inflammatory language to flourish on the platform," the AP reports, "while in Syria and the Palestinian territories, Facebook suppresses ordinary speech, imposing blanket bans on common words."

As similar stories emerged over the weekend about India and Ethiopia, Facebook said that it has more than 40,000 people "working on safety and security, including global content review teams in over 20 sites around the world reviewing content in over 70 languages."