Facebook pretending to care about democracy now is the height of hypocrisy
25/1/18
Facebook has admitted that
sometimes, it might actually be bad for democracy. Facebook is right about
that. However, I’m not sure that the social media platform really understands why
this is the case.
The admission comes in a series
of official blog posts by Facebook insiders about what effect social media can
have on democracy. “I wish I could guarantee that the positives are
destined to outweigh the negatives, but I can‘t,” wrote Samidh
Chakrabarti, a Facebook product manager. He continued: “....we have a
moral duty to understand how these technologies are being used and what can be
done to make communities like Facebook as representative, civil and trustworthy
as possible.”
First off, it’s important to
understand the political and media context in which Facebook has felt forced to
make these comments. That context is alleged ‘Russian interference’ in the 2016
election through the promotion of political ads designed to take advantage of
social division. Facebook is responding to a not small cohort of Americans who
genuinely believe that Russian Facebook ads are destroying democracy. The
second thing to understand is that while Facebook’s admission may sound like
noble self-reflection, the truth is that what Facebook says and what it means
are two very different things.
There is a temptation among some
to believe that the social media giant is a neutral actor that cares about
fairness and democracy and that it is doing its very best to ensure it has a
positive effect on democracy. This could not be further from the truth.
If Facebook’s recent history is
anything to go by, the California-based company is not actually a big fan of
democracy at all. Even before Facebook decided to become selectively outraged
about the ubiquitousness of propaganda and ‘fake news’ on its platform, it was
already engaging in political censorship. Take this 2016 story in
which Facebook employees admit to suppressing conservative news on the platform,
for example. Not only that, but employees were told to artificially “inject”Facebook-approved
stories into the trending news module when they weren’t popular enough to make
it there organically. The employees were also told not to include news about
Facebook itself into the trending category.
Facebook’s news section
operates like a traditional newsroom, reflecting the biases of its workers and
the institutional imperatives of the corporation,” Michael Nunez wrote
for Gizmodo. With that kind of ability and willingness to manipulate, Facebook
itself possesses huge potential to affect political outcomes, far more than
some Russian ads.
Facebook has said it believes
that simply adding the ability to click an“I voted” sticker
can increase actual voter turnout significantly through a combination
of simply seeing the sticker and feeling the peer pressure to vote if your
friends have done so. This is supposed to be one of the good things Facebook
has done for democracy, but there are so many ways that Facebook could use this
kind of thing to surreptitiously promote its own political agenda.
What if Facebook were to
artificially push certain news stories in specific locations – say, where an
election was taking place – and then add the “I voted” button
for users in that area. Or alternatively chose not to add that button for
certain races where a lower turnout might be deemed a good thing.
What Facebook means when it says
it is worried about how its platform is being used is that it’s not entirely
comfortable with the fact that it can’t fully control the political narrative.
Even Facebook believes it has created a monster. It would like to control what
our impressionable minds might see and read – lest we fall victim to unapproved
opinions or ideologies. But Facebook also knows that such control is not
entirely possible – and therein lies their true crisis.
Even the steps Facebook has taken
to address alleged Russian interference in the 2016 election are questionable.
In his blog post, Chakrabarti writes that the platform has “made it
easier to report false news” and has “taken steps in
partnership with third-party fact checkers to rank these stories lower” in
the news feed. Once the fact-checkers identify a story as fake, Facebook can
reduce impressions of that story by 80%, he says. But who are these third-party
fact checkers? Facebook doesn’t tell us.
“We’re also working to make it
harder for bad actors to profit from false news,” he writes. But
again, we don’t get a definition of bad actor, either. One assumes Russia is the
bad actor referred to – but if Facebook was truly concerned about government
propaganda and its effect on election outcomes, the crackdown would surely not
be limited to one government. Are some governments bad actors and other
governments good actors? Is some propaganda good and some bad? Are some
sock-puppet accounts acceptable and others not? Can we get a breakdown?
Facebook has also been kind
enough to help users figure out whether they were unfortunate enough
to have come into contact with any Russian-linked posts. It’s part of
their “action plan against foreign interference”. Again, we might
benefit from a definition here of “foreign interference.” Facebook
is an international platform, thus the potential exists for elections to be
‘interfered’ with through Facebook all over the world, not just in the United States.
Does Facebook’s fight against foreign interference incorporate all those
efforts equally? This kind of information would be really helpful, if Facebook
would be kind enough to provide it.
Facebook is not alone in its
mission to rid the world of nasty Russian propaganda. Twitter is at it, too.
Last week, the company sent out emails to users warning them that they may have
come into contact with Russian propaganda on the microblogging platform. Curiously,
no similar warnings have been sent to users who came into contact with American
propaganda online – despite the fact that we’ve known for years that the US government
has been using sock-puppet accounts to spread its own propaganda and
misinformation online.
Google has also dipped its toes
in the water. Eric Schmidt, the executive chairman of Google’s parent company
Alphabet Inc., said recently that Google was trying to create special
algorithms and“engineer the systems” to make RT’s content less visible
on the search engine.
Media coverage of Facebook’s
comments was fairly uniform. Most outlets have been treating the blog posts as
a ‘see, we told you!’ moment, focusing entirely on the Russia angle but
ignoring the many other ways in which Facebook has itself attempted to corrupt
the free flow of information and manipulated its users. The reporting is almost
sympathetic: Poor innocent Facebook is coming to terms with the fact that
sometimes bad things happen online.
The Washington Post called
Facebook blog posts the “most critical self-assessment yet.” Another
piece in the Post opines on Facebook’s “year of reckoning.” Reuters reported that
the sharing of “misleading headlines” became a “global
issue” after accusations that Russia had used Facebook to interfere
in the 2016 election. The implication is almost that misleading headlines are
some kind of new phenomenon and Facebook is out there on the frontlines of the
battle.
Facebook wants you to stay mad
about Russian ads. It wants you to believe that its democracy-loving executives
are truly sorry and doing all they can to make the platform as good for
democracy as possible. What they don’t want is for us to examine their own
practices too closely. But that’s exactly what we should be doing – instead of
congratulating them on their disingenuous foray into self-reflection.
All comments [ 0 ]
Your comments