Skip to main content

The context you need, when you need it

When news breaks, you need to understand what actually matters — and what to do about it. At Vox, our mission to help you make sense of the world has never been more vital. But we can’t do it on our own.

We rely on readers like you to fund our journalism. Will you support our work and become a Vox Member today?

Join now

This is Facebook’s plan to fight fake news

In the wake of the 2016 election, Facebook has been receiving an ongoing wave of criticism for its handling of fake news in its newsfeed used by more than a billion daily users. Initially, Facebook CEO Mark Zuckerberg dismissed concerns about fake news, arguing that it was a “pretty crazy idea” to blame Facebook for Donald Trump’s upset win.

Zuckerberg’s blasé response triggered a fierce debate inside the social media giant. And now Facebook is changing its tune in a dramatic way. In a Thursday blog post, Adam Mosseri, the Facebook executive responsible for the newsfeed, announced that Facebook will begin working with prominent third-party fact-checking organizations like Snopes, Factcheck.org, and Politifact to verify news articles.

Stories flagged by these fact-checking organizations will be marked as “disputed” in the newsfeed, and these stories will be penalized in the newsfeed algorithm, meaning that they’ll tend to show up further down the list but won’t be removed from Facebook altogether.

It’s exactly the kind of ambitious steps that Facebook’s critics have been demanding for weeks. It’s also likely to cause a backlash among conservatives, many of whom believe that a company from liberal Silicon Valley will simply use the “fake news” label as a pretext for suppressing conservative viewpoints generally.

But that’s a shortsighted concern. In the long run, it’s not good for conservatives for the right-of-center political conversation to be strongly influenced by totally bogus news stories. Facebook’s new approach will raise the profile of more responsible conservative news organizations like the Washington Times, the Washington Examiner, and National Review. And in the long run, that’s going to be good for both the conservative movement and American democracy more generally.

Facebook has more and more influence over the news we see

A handful of big tech companies — Twitter, Google, and especially Facebook — have gained a huge and growing influence over what news people see. Forty-four percent of US adults tell pollsters they got news from Facebook in 2016. That’s vastly larger than other news-focused social media sites like Twitter (9 percent) and Reddit (2 percent). And while many people still get their news from television programs or newspapers, those media are divided among many competing news organizations. This means that Facebook has a larger influence over ordinary Americans’ media diets than almost any other news organization.

Normally we think that organizations with a lot of power have an obligation to use that power responsibly. But Facebook’s leaders have traditionally resisted thinking of themselves in those terms. Zuckerberg has repeatedly insisted that “we are a tech company, not a media company” — implying that Facebook should not be in the business of making editorial decisions or held responsible for the fake material cluttering its newsfeed.

But the 2016 election campaign showed the folly of this see-no-evil approach. The basic problem is that Facebook’s massive growth actually created an incentive for people to create websites specifically designed to manufacture, and profit from, fake news.

For example, investigations by BuzzFeed and the Guardian found that a group of cynical Macedonian hucksters had created dozens of right-wing news sites that publish low-quality pro-Trump news stories. Some are plagiarized from other conservative news sites. Others appear to be totally made up, with headlines like “Proof surfaces that Obama was born in Kenya,” “Bill Clinton’s sex tape just leaked,” and “Pope Francis forbids Catholics from voting for Hillary!”

“Yes, the info in the blogs is bad, false, and misleading but the rationale is that ‘if it gets the people to click on it and engage, then use it,’” a Macedonian student told BuzzFeed.

There has always been shoddy journalism and accidental mistakes in news articles, of course. But until Facebook’s newsfeed came along, there was no real incentive for someone to create sites full of false but clickable stories, since these sites would get very little traffic and hence few, if any, advertising dollars.

Facebook’s growing popularity created a market opportunity for fake news sites, while the growth of internet advertising networks made it easy to turn traffic into cash. Both Google and Facebook have already taken steps to ban fake news sites from using their lucrative ad networks — but there are lots of other ad networks these sites can use.

Facebook is relying on third-party organizations to battle fake news

The fake news situation puts Facebook in an awkward position, because it likes to think of itself as simply providing a neutral platform for their users to share information with one another. Making direct editorial judgments about which news stories are fake would seemingly run contrary to that platform-company ethos.

So Facebook is essentially outsourcing these editorial judgments to independent organizations that are already in the fact-checking business. Facebook says it will rely on Poynter’s International Fact Checking Network to identify suitable fact-checking organizations. A Facebook spokesperson tells me that at launch, Facebook will rely on ABC News, the Associated Press, Snopes, Factcheck.org, and PolitiFact to verify news articles.

“We’ll use the reports from our community, along with other signals, to send stories to these organizations,” Mosseri writes. “If the fact checking organizations identify a story as fake, it will get flagged as disputed and there will be a link to the corresponding article explaining why.”

If users flag an article as inaccurate, Facebook will refer the article to the third-party fact-checkers, who will examine the article and render a verdict. If a story is ruled false by fact-checkers, Facebook will add a “disputed” tag to the article anytime a user shares it.

Disputed stories will also be penalized by Facebook’s newsfeed algorithm, which decides which stories to show you first. In the past, Facebook’s newsfeed has relied heavily on “engagement” — that is, how often someone clicks, likes, and shares a link — to decide which stories to show users first.

As one Facebook critic put it: “News Feed optimizes for engagement. As we’ve learned in this election, bullshit is highly engaging.”

Facebook will still take engagement into account, but now it will also take accuracy into account. Disputed stories can still be shared on the newsfeed, but they’ll tend to appear further down in the newsfeed than stories that have not been marked as inaccurate by fact-checkers.

And if users try to share a disputed story, Facebook will warn users that the story is disputed and ask if they really want to share it. That might help discourage users from unknowingly sharing bogus information.

More in Technology

Podcasts
Are humanoid robots all hype?Are humanoid robots all hype?
Podcast
Podcasts

AI is making them better — but they’re not going to be doing your chores anytime soon.

By Avishay Artsy and Sean Rameswaram
Future Perfect
The old tech that could help stop the next airborne pandemicThe old tech that could help stop the next airborne pandemic
Future Perfect

Glycol vapors, explained.

By Shayna Korol
Future Perfect
Elon Musk could lose his case against OpenAI — and still get what he wantsElon Musk could lose his case against OpenAI — and still get what he wants
Future Perfect

It’s not about who wins. It’s about the dirty laundry you air along the way.

By Sara Herschander
Life
Why banning kids from AI isn’t the answerWhy banning kids from AI isn’t the answer
Life

What kids really need in the age of artificial intelligence.

By Anna North
Culture
Anthropic owes authors $1.5B for pirating work — but the claims process is a Kafkaesque messAnthropic owes authors $1.5B for pirating work — but the claims process is a Kafkaesque mess
Culture

“Your AI monster ate all our work. Now you’re trying to pay us off with this piece of garbage that doesn’t work.”

By Constance Grady
Future Perfect
Some deaf children are hearing again because of a new gene therapySome deaf children are hearing again because of a new gene therapy
Future Perfect

A medical field that almost died is quietly fixing one disease at a time.

By Bryan Walsh