Facebook executives feel
the heat of content controversies
Send a link to a friend
[October 28, 2016]
By Kristina Cooke, Dan Levine and Dustin Volz
SAN
FRANCISCO/WASHINGTON (Reuters) - After Facebook's removal of an iconic
Vietnam war photo stirred an international uproar last month, the social
network's executives quickly backtracked and cleared its publication.
But the image - showing a naked Vietnamese girl burned by napalm - had
previously been used in training sessions as an example of a post that
should be removed, two former Facebook employees told Reuters.
Trainers told content-monitoring staffers that the photo violated
Facebook policy, despite its historical significance, because it
depicted a naked child, in distress, photographed without her consent,
the employees told Reuters.
The social network has taken great pains to craft rules that can be
applied uniformly with minimal discretion. The reversal on the war
photo, however, shows how Facebook's top executives sometimes overrule
company policy and its legions of low- and mid-level content monitors.
Facebook has often insisted that it is a technology company - not a
media company - but an elite group of at least five senior executives
regularly directs content policy and makes editorial judgment calls,
particularly in high-profile controversies, eight current and former
Facebook executives told Reuters.
One of those key decision-makers - Justin Osofsky, who runs the
community operations division - wrote a Facebook post acknowledging that
the removal of the war photo was a "mistake.”
“Sometimes,” he wrote, “the global and historical significance of a
photo like ‘Terror of War’ outweighs the importance of keeping nudity
off Facebook.”
Facebook spokeswoman Christine Chen declined to comment on the company’s
use of the photo in training sessions.
Facebook has long resisted calls to publicly detail its policies and
practices on censoring postings. That approach has drawn criticism from
users who have had content removed and free-speech advocates, who cite a
lack of transparency and a lack of an appeals process for many content
decisions.
At the same time, some governments and anti-terror groups are pressuring
the company to remove more posts they consider offensive or dangerous.
HIGH-LEVEL REVIEW
The current and former Facebook executives, most of them speaking on
condition of anonymity, told Reuters in detail how complaints move
through the company’s content-policing apparatus. The toughest calls,
they said, rise to an elite group of executives.
Another of the key decision-makers is Global Policy Chief Monika Bickert,
who helped rule on the fracas over the war photo.
"That was one we took a hard look at, and we decided it definitely
belonged on the site," said Bickert, a former federal prosecutor.
She declined to elaborate on the decision-making process.
Facebook chief operating officer Sheryl Sandberg followed up with an
apology to Norwegian Prime Minister Erna Solberg, who had posted the
photo on her own account after Facebook removed it from others in her
country.
In addition to Sandberg, Osofsky and Bickert, executives involved in
sensitive content issues include Joel Kaplan, Facebook’s
Washington-based government relations chief; and Elliot Schrage, the
vice president for public policy and communications.
All five studied at Harvard, and four of them have both undergraduate
and graduate degrees from the elite institution. All but Sandberg hold
law degrees. Three of the executives have longstanding personal ties to
Sandberg.
Chief Executive Mark Zuckerberg, a Harvard drop-out, occasionally gets
involved with content controversies, Bickert said.
These executives also weigh in on content policy changes meant to
reflect shifting social context and political sensitivities around the
world, current and former executives said.
Facebook officials said the five people identified by Reuters were not
the only ones involved in high-level content decisions.
“Facebook has a broad, diverse and global network involved in content
policy and enforcement, with different managers and senior executives
being pulled in depending on the region and the issue at hand,” Chen
said.
Chen declined to name any other executives who were involved in content
policy.
A WAR OVER FREE EXPRESSION
The company's reticence to explain censorship decisions has drawn
criticism in many countries around the globe.
Last month, Facebook disabled the accounts of editors at two of the most
widely read Palestinian online publications, Shehab News Agency and Quds.
In keeping with standard company practice, Facebook didn't publicly
offer a reason for the action or pinpoint any content it considered
inappropriate.
The company told Reuters that the removal was simply an error.
Some Palestinian advocacy groups and media outlets condemned the
shutdowns as censorship stemming from what they described as Facebook’s
improper alliance with the Israeli government.
[to top of second column] |
Monika Bickert, Facebook's head of global policy management, is
interviewed by Reuters in Washington DC February 2, 2016.
REUTERS/Gary Cameron/File Photo
Israel’s government has pushed Facebook to block hundreds of pages it
believes incite violence against Jews, said Noam Sela, spokesman for
Israeli cabinet Minister Gilad Erdan.
Sela said the Israeli government “had a connection” at Facebook to
handle complaints but declined to elaborate on the relationship.
“It’s not working as well as we would like,” Sela said. “We have more
work to do to get Facebook to remove these pages.”
Ezz al-Din al-Akhras, a Quds supervisor, said that Facebook's head of
policy in the Middle East had gotten in touch after the uproar over the
shutdowns and that three of four suspended accounts were restored.
“We hope the Facebook campaign of suspending and removing Palestinian
accounts will stop,” he said. “We do not practice incitement; we are
only conveying news from Palestine to the world.”
Facebook said the restoration of the accounts was not a response to
complaints. It declined to comment on whether top executives were
involved.
The company has cited technological glitches in other recent cases where
content was removed, then restored, including the takedown of a video
that showed the aftermath of a Minneapolis police shooting.
Chen declined to explain the glitch.
She said the company was reviewing its appeals process in response to
public feedback. Facebook currently allows appeals of company actions
involving entire profiles set up by people or institutions, or full
pages on those profiles, but not for individual posts.
THICK RULEBOOK
To manage the huge volume of content complaints - more than a million a
day - the company employs a multi-layered system. It starts with
automated routing of complaints to content-policing teams in Dublin,
Hyderabad, Austin and Menlo Park, who make initial rulings, current and
former executives said.
These low-level staffers and contractors consult a thick rulebook that
interprets the comparatively spare "community standards" that Facebook
customers are asked to follow. The company trains front-line monitors to
follow rules and use as little discretion as possible.
When a removal sparks more complaints, regional managers function as a
mid-level appeals court. Continuing controversy could then push the
issue to top U.S. executives.
Senior executives also weigh in on policy updates. Osofsky and Kaplan,
for instance, wrote a blog post last week, in response to “continued
feedback” on content removals, explaining that the company would start
weighing news value more heavily in deciding whether to block content.
In an earlier post, responding to the Napalm-girl controversy, Osofsky
said Facebook's policies usually work well, but not always.
"In many cases, there's no clear line between an image of nudity or
violence that carries global and historic significance and one that
doesn't," Osofsky wrote.
The Vietnam war photo - depicting horrors suffered by a girl named Phan
Thi Kim Phuc - was first removed from an account in Norway by a
front-line monitor.
In protest, the Norwegian newspaper Aftenposten printed the image on its
front page and posted it on Facebook, which removed it. That prompted
the prime minister to post the photo - only to have Facebook remove it
again.
Facebook then issued a statement defending the action, saying it was
“difficult to create a distinction between allowing a photograph of a
nude child in one instance and not others.”
The next day, executives reversed the call, with Sandberg telling the
prime minister: “Even with clear standards, screening millions of posts
on a case-by-case basis every week is challenging."
(Additional reporting by Yasmeen Abutaleb and Joseph Menn in San
Francisco, Nidal al-Mughrabi in Gaza and Terje Solsvik in Oslo; Editing
by Jonathan Weber and Brian Thevenot)
[© 2016 Thomson Reuters. All rights
reserved.] Copyright 2016 Reuters. All rights reserved. This material may not be published,
broadcast, rewritten or redistributed.
|