The favorite app among the youth, TikTok, continues to stir up controversy. With over 21 million users in France, the platform recently boasted about the impressive performance of TikTok Shop. However, a new investigation by Amnesty International has exposed the shortcomings of its recommendation algorithm.

Contrary to protecting minors as required by European law, the platform appears to promote harmful content related to depression and suicide. Simultaneously, another report by Science Feedback identifies TikTok as the network most affected by misinformation.

Algorithms Intensifying Spirals of Distress

By creating fake profiles of 13-year-old teens, Amnesty International’s team monitored how TikTok selects videos that appear in the “For You” feed. Within just twenty minutes, these test accounts were flooded with content that invoked sadness, loss of self-confidence, and suicide.

Even worse, after three hours of watching, the feeds were overwhelmed with dark videos, some explicitly detailing self-harm methods…

These findings confirm the long-standing accusations from several families: an algorithm that traps young users in a “vicious cycle” where suffering drives engagement.

The group Algos Victima, formed by parents who have lost their children, accuses TikTok of exploiting the emotional vulnerability of teenagers as an economic resource. This sentiment is echoed by Justine Payoux of Amnesty International France: “The longer a user stays on the platform, the more the platform profits.”

Widespread Misinformation and Lack of Regulation

In parallel, a study conducted by Science Feedback, funded by the European Fund for Media and Information, shows that approximately 20% of content on TikTok contains false or misleading information.

Already criticized for spreading fake news about the coronavirus in 2020, this rate is the highest among major social media platforms, far surpassing Facebook and X.

The researchers condemn a structural bias: content with low credibility generates more engagement, and thus more visibility. This “misinformation premium” stems directly from TikTok’s recommendation mechanisms, which focus on maximizing screen time. By amplifying emotionally charged content – whether false or morbid – the algorithm perpetuates the attention economy’s dynamics.

TikTok Facing Accountability in Europe

In light of these revelations, Amnesty International announces it will file a lawsuit against TikTok for violating the Digital Services Act. The European legislation requires platforms to mitigate systemic risks and protect minors. However, according to the NGO, TikTok’s technical adjustments, such as the “Not Interested” button, are insufficient to curb the spread of dangerous content.

This case could represent a turning point in the regulation of algorithms in Europe, as TikTok might be forced to significantly revise its business model if it cannot demonstrate control over the impacts of its recommendation system.