Skip to main content

The context you need, when you need it

When news breaks, you need to understand what actually matters — and what to do about it. At Vox, our mission to help you make sense of the world has never been more vital. But we can’t do it on our own.

We rely on readers like you to fund our journalism. Will you support our work and become a Vox Member today?

Join now

How TikTok’s hate speech detection tool set off a debate about racial bias on the app

TikTok says it’s fixing an error that caused its algorithms to flag BLM content, but some Black creators are still frustrated.

An image of the TikTok logo on a black phone screen.
An image of the TikTok logo on a black phone screen.
Some Black creators are frustrated with how the app seemingly flagged phrases about Black people as inappropriate, which the company says was an error in its hate speech detection system.
Photo Illustration by Nikolas Kokovlis/NurPhoto via Getty Images
Shirin Ghaffary
Shirin Ghaffary was a senior Vox correspondent covering the social media industry. Previously, Ghaffary worked at BuzzFeed News, the San Francisco Chronicle, and TechCrunch.

“This is why I’m pissed the fuck off. We’re tired,” said popular Black influencer Ziggi Tyler in a recent viral video on TikTok. “Anything Black-related is inappropriate content,” he continued later in the video.

Tyler was expressing his frustration with TikTok about a discovery he made while editing his bio in the app’s Creator Marketplace, which connects popular account holders with brands who pay them to promote products or services. Tyler noticed that when he typed phrases about Black content in his Marketplace creator bio, such as “Black Lives Matter” or “Black success,” the app flagged his content as “inappropriate.” But when he typed in phrases like “white supremacy” or “white success,” he received no such warning.

For Tyler and many of his followers, the incident seemed to fit within a larger pattern of how Black content is moderated on social media. They said it was evidence of what they believe is the app’s racial bias against Black people; some urged their followers to leave the app while others tagged TikTok’s corporate account and demanded answers. Tyler’s original video about the incident has received over 1.2 million views and over 25,000 comments; his follow-up video has received another nearly 1 million views.

“I’m not going to sit here and let that happen,” Tyler, a 23-year-old recent college graduate from Chicago, told Recode. “Especially on a platform that makes all these pages saying things like, ‘We support you, it’s Black history month in February.’”

A spokesperson for TikTok told Recode that the issue was an error with its hate speech detection systems that it is actively working to resolve and that it is not indicative of racial bias. TikTok’s policies do not restrict posting about Black Lives Matter, according to a spokesperson.

In this instance, TikTok told Recode that the app is mistakenly flagging phrases like “Black Lives Matter” because its hate speech detector is triggered by a combination of words involving the words “Black” and “audience” — because “audience” contains the word “die” in it.

“Our TikTok Creator Marketplace protections, which flag phrases typically associated with hate speech, were erroneously set to flag phrases without respect to word order,” a company spokesperson said in a statement. “We recognize and apologize for how frustrating this was to experience, and our team is working quickly to fix this significant error. To be clear, Black Lives Matter does not violate our policies and currently has over 27B views on our platform.” TikTok says it has reached out to Tyler directly and that he hasn’t responded.

But Tyler said he didn’t find TikTok’s explanation to Recode to be adequate and that he felt the company should have identified an issue in its hate speech detection system sooner.

“Regardless of what the algorithm is and how it picked up, somebody had to program that algorithm,” Tyler told Recode. “And if [the problem] is the algorithm, and the marketplace has been available since [2020], why wasn’t this a conversation you had with your team, knowing there have been racial controversies?” he asked.

Tyler isn’t alone in his frustration. He’s one of many Black creators who have been protesting TikTok recently because they say they are unrecognized and underserved. Many of these Black TikTokers are participating in what they’re calling the “#BlackTikTok Strike,” in which they are refusing to make up original dances to a hit song because they are angry that Black artists on the app are not being properly credited for the viral dances that they first choreograph and that other creators imitate.

These issues also connect to another criticism that’s been leveled at TikTok, Instagram, YouTube, and other social media platforms over the years: That their algorithms, which recommend and filter the posts everyone sees, often have inherent racial and gender biases.

In 2019, for example, a study showed that leading AI models for detecting hate speech are 1.5 times more likely to flag tweets written by African Americans as “offensive” compared to other tweets.

Findings like that have fostered an ongoing debate about the merits and potential harms that come with relying on algorithms — particularly developing AI models — to automatically detect and moderate social media posts.

Major social media companies like TikTok, Google, Facebook, and Twitter, though they acknowledge that these algorithmic models can be flawed, are still making them a key part of their rapidly expanding hate speech detection systems. They say they need a less labor-intensive way to keep up with the ever-expanding volume of content on the internet.

Tyler’s TikTok video also shows the tensions surrounding these apps’ lack of transparency about how they police content. In June 2020, during Black Lives Matter protests across the US, some activists accused TikTok of censoring certain popular #BlackLivesMatter posts — which for a time the app showed as having zero views even when they had billions of views. TikTok denied this and said it was a technical glitch also affecting other hashtags. And in late 2019, TikTok executives were reportedly discussing tamping down political discussion on the app, according to Forbes, to avoid political controversy.

A spokesperson for TikTok acknowledged larger frustrations about Black representation on the platform and said that, earlier this month, the company launched an official @BlackTikTok account to help foster the Black TikTok community and that overall, its teams are committed to developing recommendation systems that reflect inclusivity and diversity.

But for Tyler, the company has a lot more work to do. “This instance is just the tip of the iceberg and underneath the water level you have all of these issues,” said Tyler.

More in Technology

Podcasts
Are humanoid robots all hype?Are humanoid robots all hype?
Podcast
Podcasts

AI is making them better — but they’re not going to be doing your chores anytime soon.

By Avishay Artsy and Sean Rameswaram
Future Perfect
The old tech that could help stop the next airborne pandemicThe old tech that could help stop the next airborne pandemic
Future Perfect

Glycol vapors, explained.

By Shayna Korol
Future Perfect
Elon Musk could lose his case against OpenAI — and still get what he wantsElon Musk could lose his case against OpenAI — and still get what he wants
Future Perfect

It’s not about who wins. It’s about the dirty laundry you air along the way.

By Sara Herschander
Life
Why banning kids from AI isn’t the answerWhy banning kids from AI isn’t the answer
Life

What kids really need in the age of artificial intelligence.

By Anna North
Culture
Anthropic owes authors $1.5B for pirating work — but the claims process is a Kafkaesque messAnthropic owes authors $1.5B for pirating work — but the claims process is a Kafkaesque mess
Culture

“Your AI monster ate all our work. Now you’re trying to pay us off with this piece of garbage that doesn’t work.”

By Constance Grady
Future Perfect
Some deaf children are hearing again because of a new gene therapySome deaf children are hearing again because of a new gene therapy
Future Perfect

A medical field that almost died is quietly fixing one disease at a time.

By Bryan Walsh