EU warns TikTok over illegal content and misinformation

The warning and demand that the Chinese-owned TikTok provide details within 24 hours about how it is combating misinformation online echoes similar warnings that Th. Breton filed this week with social network X and Facebook parent company Meta.

“Given that your platform is widely used by children and teenagers, you have a special responsibility to protect them from violent content depicting hostage-taking and other gruesome videos that are reported to be widely distributed on your platform without adequate safeguards,” – in a letter to the head of TikTok, Shou Zi Chew, Th. Breton.

Th. Breton stressed that TikTok and other major internet platforms need to comply with the Digital Services Act (DSA), which gives the EU the power to fine platforms up to 6 percent. their global circulation if they do not fight against illegal content.

Like Elon Musk, the owner of the X platform, and Mark Zuckerberg, the head of Meta, Th. Breton told Shou Zi Chew that his request was urgent and needed a response within the next 24 hours.

The full text of the letter was published in Th. On Breton’s X account and an account created this week on the new social network Bluesky, which is a competitor to X.

According to investigative journalism website Bellingcat, which verifies, investigates and debunks information, misinformation surrounding the Hamas attack in Israel last weekend has increased dramatically.

It documented examples of false or misleading videos being posted on X, TikTok and other platforms.

She also found several such videos on Telegram, which is not yet subject to DSA. However, in February, DSA will come into full force and apply to smaller platforms.

#warns #TikTok #illegal #content #misinformation

How⁤ does​ the spread of misinformation on TikTok compare to other social media platforms in terms of impact and user vulnerability?

⁢**Interview with‌ Dr. Sarah Chen, Media Literacy Expert**

**Editor:**‌ Today, ​we delve into the urgent concerns raised by European ⁣Commissioner Thierry Breton regarding the spread⁤ of misinformation‌ and violent ⁤content on platforms like‍ TikTok. Joining us is⁢ Dr. Sarah Chen, a media literacy expert⁤ who has worked ⁤extensively on issues of​ social media governance ⁣and child safety online. Welcome, Dr. Chen.

**Dr. Chen:** Thank you ⁣for having me.

**Editor:** Commissioner Breton recently issued a ⁢warning to TikTok, demanding details ⁢on their efforts to combat misinformation‌ within ⁤24 hours.⁢ Why is this⁤ issue particularly pressing for TikTok in comparison to other platforms?

**Dr.‍ Chen:** TikTok has a unique user base that skews younger, ⁣with many users being children and teenagers. This demographic is ⁢particularly ⁤vulnerable to misinformation and‌ harmful content, making the platform’s responsibility​ much ‌greater. The ⁣immediacy ‍of these concerns stems from not only the⁤ potential‌ for⁤ misinformation to spread ‍rapidly but also the lack of adequate safeguards against graphic content that can adversely affect ‍young audiences.

**Editor:** In his ⁢letter to TikTok CEO ​Shou Zi Chew, Breton emphasized the need for protective measures against violent content. How effective do ⁣you think TikTok has been​ in addressing these issues so far?

**Dr. Chen:** TikTok ‍has made strides in​ combating⁤ misinformation, implementing features to flag false content and engaging with fact-checkers‍ to enhance its information integrity. However, the challenge remains significant. The ⁢platform has acknowledged the‌ impacts of misinformation⁢ on public trust and safety,​ as outlined in their recent updates ⁢ [[1](https://newsroom.tiktok.com/en-us/an-update-on-our-work-to-counter-misinformation)]. Yet, many ⁢users‍ report that violent or inappropriate ⁤content continues to slip ‍through the cracks, raising‌ questions ‌about the effectiveness of their current ⁣measures.

**Editor:** What ​steps do you⁤ think⁣ TikTok should take immediately to address these concerns raised by Breton?

**Dr. Chen:** Firstly,‌ TikTok should enhance⁢ its content moderation systems to more effectively identify and remove violent and misleading content. This includes employing advanced AI⁢ tools ⁢as well as human moderators ⁢trained to‍ recognize harmful ‍material. Additionally, they need to be more transparent about their processes and provide straightforward reporting mechanisms for users to ⁣flag inappropriate ⁤content. Regular updates and communication‌ with ‌the public about⁣ these efforts can help ​rebuild trust.

**Editor:**⁣ Do you think there is a broader implication‌ for social media platforms ⁤beyond TikTok?

**Dr. Chen:** Absolutely. The issues raised by Breton are reflective of a larger, systemic problem across all major social media platforms. Regulators are increasingly holding ‌companies accountable for what happens on their sites, ‍and this trend is likely to continue. Effective regulation‍ will require collaboration between governments and tech companies to create‌ standards ‍that prioritize user safety, especially for younger users.

**Editor:** Thank⁣ you, Dr. Chen, for your insights on this critical issue. It’s clear that platforms like TikTok have a long ⁣road​ ahead in‌ terms of ensuring safety and accuracy for ⁢their users.

**Dr. ⁢Chen:** Thank you for having me. It’s crucial that we‍ keep these⁣ discussions ongoing as technology and its⁢ impact ⁤on society evolves.

Leave a Replay