When I think of TikTok, the first things that come to mind are dance choreographies, short-lived challenges and cute animal videos. Ask someone else and they’ll probably mention similar concepts such as the Tim Burton challenge, singing bowls’ comeback or the unboxing trend. But no one in their right mind would mention live streamed suicides or teen deaths, right? Well, as it turns out, they wouldn’t be wrong.
TikTok has already made headlines for its strange way of moderating certain types of content. While comments are not getting deleted quickly enough, TikTok moderators are (rightly so) being accused of discriminating and racist content moderation. In other words, although we’re quick to glamourise the sensation that the video-sharing app has become, we tend to forget or ignore its dark side. This time, let’s not do that—let’s look at TikTok’s worst aspects so that we can work towards fixing those (data privacy problems put aside just this once).
In February 2019, a 19-year-old vlogger living in Curitiba, Brazil, took his own life on a TikTok livestream after warning his fans a day earlier that he was planning a “special performance.” Around 280 people watched the man kill himself on the stream, which continued to show his body until TikTok moderators finally took it down. During that time, users posted nearly 500 comments and 15 complaints. It took TikTok three hours to warn police and over an hour and a half to take the video down.
Reportedly, TikTok took steps to prevent the post from going viral first before notifying the authorities and Business Insider reports that the video-sharing app’s first move was to notify its PR team immediately. This story only came out into the open a year after the incident took place, so TikTok’s PR team obviously did a good job at stifling it.
Then, in September 2020, a video of another man committing suicide by shooting himself in the head with a shotgun began circulating on the app. Despite not seeing it myself, I witnessed the mass outcry and shock firsthand in reply videos and their comments section. Against odds, the video remained on the platform for a few days, which resulted in TikTok being heavily criticised for its poor moderation efforts.
In response to the moderators’ inaction, several users ended up posting engagement they had with TikTok moderators, who reportedly told them the video “doesn’t violate our Community Guidelines.” In the meantime, users took the matter into their own hands by sharing videos that warned others about the presence of the suicide clip on TikTok. “Please try not to go on TikTok today or tomorrow,” one video said. “There is a very graphic and gorey suicide video going around right now!”
“Please stop treating this like a meme, please stop treating this like a joke, this is a real person who passed and his family is grieving,” said another TikTok user. In July, the app’s moderation guidelines were questioned once again, after its algorithm promoted a collection of anti-semitic memes soundtracked by the lyrics, “We’re going on a trip to a place called Auschwitz, it’s shower time.” Nearly 100 users featured the song in their videos, which remained on the app for three days.
TikTok’s Transparency Report published in July 2020, says that the app removed over 49 million videos globally in the second half of last year, with 98.2 per cent of those being taken down before they were reported. 89.4 per cent of these were removed before they received any views. Yet, TikTok is known for censoring users and content that doesn’t violate any guidelines, including a teenager who criticised China, those deemed ugly or disabled and Black creators.
Fast forward to October 2020, and another death can be somehow ‘assigned’ to TikTok. 21-year-old Areline Martinez was shot in the head by one of her friends in what has been referred to as an accident, as Mexico News Daily first reported. Martinez was killed while attempting to stage a kidnapping for a TikTok video.
Previous videos posted on Martinez’s TikTok page featured scenes in which she was blindfolded with her hands and feet bound, while men surrounded her and pointed guns at her head. TikTok has since removed these videos. Many of the friends who were involved in the fake kidnapping fled the scene after the killing, though a “behind the scenes” video posted to TikTok before Martinez was killed was used by authorities to identify the individuals.
Undoubtedly, TikTok moderators cannot catch every instance of inappropriate content, but the timeline above clearly highlights the amount of content that goes unnoticed on the app for too long—or sometimes simply ignored by moderators until users start getting involved. TikTok’s content-moderation is a time bomb waiting to explode in our face.
Because teens are using the app not just as a channel for light-hearted fun but also as a space to discuss personal problems, traumas and politics, the more serious the TikTok conversation gets, the more potential mischief and “coordinated inauthentic behaviour,” as the app calls it, its users will face from bad actors. Even Bill Gates called TikTok “a poison chalice.” The question that remains now is how; how can this be stopped?
If you’re struggling with mental health issues and feel like you need help, you can contact the suicide prevention specialists Samaritans in the UK here or the National Suicide Prevention Lifeline in the US here.