Your daily adult tube feed all in one place!
In a bid to be more 'transparent', TikTok has introduced a new tool that explains why specific videos are being recommended to users.
The tool, available on the 'For You' feed on the TikTok app, brings up a window entitled 'Why you're seeing this video'.
It then gives a list of reasons, which can include 'this video is popular in your region' or 'this video was recently posted by an account you follow'.
TikTok, which is owned by Chinese company Bytedance, says the new feature brings 'more context' and 'meaningful transparency' for users.
The tool, available on the 'For You' page on the TikTok app, brings up a window entitled 'Why you're seeing this video'
The platform uses an algorithm to promote videos to users on the For You page from people they don't even follow – but this has proved controversial for promoting dangerous content.
For example, a recent Daily Mail investigation found teenagers are being bombarded with self-harm and suicide content on TikTok within minutes of joining the platform.
'At TikTok, we want people to feel empowered creating, connecting, and engaging on our platform,' TikTok says in a blog post.
'That's why we equip creators and viewers with a range of features, tools, and resources so they can stay in control of their experience.
'Today we're adding to that toolbox with a feature that helps bring more context to content recommended in For You feeds.'
There's not a set launch date for the new tool but a TikTok spokesperson said it will roll out 'over the coming weeks'.
Back in June 2020, TikTok published a detailed blog post explaining how the For You page works, as part of its transparency push.
When TikTok users open the app, they are presented with the For You feed – a stream of videos recommended to users.
A recent Daily Mail investigation found teenagers are being bombarded with self-harm and suicide content on TikTok within minutes of joining the platform. An account set up by the Daily Mail as 14-year-old Emily was shown posts about suicidal thoughts within five minutes of expressing an interest in depression content (pictured)
For You is powered by an algorithm that makes personalised recommendations based on several metrics, including viewing behaviour and interactions with videos.
Similar to the recommendations offered by search engines, streaming services and other social media platforms, this is designed to give users a more personalised experience.
Unfortunately, an inherent issue with social media algorithms is that users are directed to a stream of posts related to a video that they've viewed.
Just as a user who sees an explosion of football-related videos after watching a person kick a ball into a goal, self-harm or suicide-related posts can end up bombarding a user's feed.
This came to a tragic head in 2017 when London school girl Molly Russell ended her life after viewing Instagram content linked to anxiety, depression, self-harm and suicide.
Molly's father, Ian, who now campaigns for online safety, has previously said the 'pushy algorithms' of social media 'helped kill my daughter'.
A recent report from the Center for Countering Digital Hate (CCDH) found TikTok's For You page is 'bombarding vulnerable teenagers with dangerous content that might encourage self-harm, suicide, disordered eating and eating disorders'.
Researchers at the non-profit organisation set up accounts posing as 13-year-old girls and 'liked' harmful content whenever it was recommended.
Videos referencing suicide were served to one account within 2.6 minutes, while eating disorder content was served to one account within eight minutes, they found.
CCDH said TikTok's For You page lacks 'meaningful transparency' and its algorithm operates 'in an opaque manner'.
The platform allows users as young as 13 to create an account by asking for a date of birth, although it doesn't stop kids below this age from using it.
The platform uses an algorithm to promote videos to users on the For You page – but this has proved controversial for promoting dangerous content (file photo)
'All platforms need to ensure content being targeted at children is consistent with what parents rightly expect to be safe,' said Richard Collard, NSPCC policy and regulatory manager for child safety online.
TikTok says it does not allow self-harm and suicide content if it 'promotes, glorifies or normalises'.
It's also working on 'interrupting repetitive patterns and diversifying recommendations' that are pushed by its algorithm.
'This activity and resulting experience does not reflect genuine behaviour or viewing experiences of real people,' a TikTok spokesperson said.
'We regularly consult with health experts, remove violations of our policies, and provide access to supportive resources for anyone in need.
'We're mindful that triggering content is unique to each individual and remain focused on fostering a safe and comfortable space for everyone, including people who choose to share their recovery journeys or educate others on these important topics.'
In 2021, TikTok began testing ways to avoid recommending a series of similar content on topics that 'may be fine as a single video but potentially problematic if viewed repeatedly', such dieting, extreme fitness and sadness.
Read similar stories here...
TikTok tests landscape videos in bid to 'move into YouTube territory'
TikTok to add 3,000 engineers globally while rivals laid off thousands
Twitter Blue is relaunched - but iPhone users have to pay more