Bar children from suggested friend lists to prevent groomer contact, Ofcom tells social media platforms
New online safety regulator has published its first draft codes of practice around stopping illegal content spreading online
Your support helps us to tell the story
From reproductive rights to climate change to Big Tech, The Independent is on the ground when the story is developing. Whether it's investigating the financials of Elon Musk's pro-Trump PAC or producing our latest documentary, 'The A Word', which shines a light on the American women fighting for reproductive rights, we know how important it is to parse out the facts from the messaging.
At such a critical moment in US history, we need reporters on the ground. Your donation allows us to keep sending journalists to speak to both sides of the story.
The Independent is trusted by Americans across the entire political spectrum. And unlike many other quality news outlets, we choose not to lock Americans out of our reporting and analysis with paywalls. We believe quality journalism should be available to everyone, paid for by those who can afford it.
Your support makes all the difference.The biggest social media platforms will be required to protect children online by keeping them off suggested friend lists to prevent contact by groomers, Ofcom has said.
The new online safety regulator has published its first draft codes of practice under the Online Safety Act, which was signed into law last week.
The first codes focus on illegal material online – such as child sexual abuse material, grooming content and fraud.
Under the code, the largest platforms will be required, by default, to ensure that children on their sites are not presented with lists of suggested friends, do not appear in other users’ lists, that their location information is not visible to other users and that people outside their agreed connections cannot direct message them.
Ofcom is set to publish further codes in the coming months on other areas of online safety, such as guidance for adult sites on keeping children away and protecting children from harmful content promoting things such as suicide or self-harm.
Each of the draft codes will have a consultation period before requiring final approval from Parliament.
The regulator’s own timetable says it hopes to begin enforcing its first codes of practice by the end of 2024.
The illegal content code also encourages larger sites to use hash-matching technology to identify illegal images of sexual abuse and use automated tools to detect websites that have been identified as hosting abuse material.
In fighting fraud and terrorism, Ofcom says services should use automatic detection systems to find and remove posts linked to the sale of stolen financial information and block all accounts run by proscribed terrorist organisations.
The codes of practice also propose that tech firms nominate an accountable person who reports to senior management on compliance around illegal content, reporting and complaints duties, ensure their content moderation teams are well-resourced and trained, offer easy reporting and blocking tools to use, and carry out safety tests on recommendation algorithms.
Dame Melanie Dawes, Ofcom’s chief executive, said: “Regulation is here, and we’re wasting no time in setting out how we expect tech firms to protect people from illegal harm online while upholding freedom of expression.
“Children have told us about the dangers they face, and we’re determined to create a safer life online for young people in particular.”
Technology Secretary Michelle Donelan said the publication of the first codes marked a “crucial” step in making the Online Safety Act a reality by “cleaning up the wild west of social media and making the UK the safest place in the world to be online”.
“Before the Bill became law, we worked with Ofcom to make sure they could act swiftly to tackle the most harmful illegal content first,” she said.
“By working with companies to set out how they can comply with these duties, the first of their kind anywhere in the world, the process of implementation starts today.”
Ofcom said it had been and would continue working with social media and other in-scope platforms over the coming months to help ensure they were in compliance with the proposed codes when they come into full force.
Campaign groups have backed the first proposals from the regulator.
Susie Hargreaves, chief executive of the Internet Watch Foundation, said: “We stand ready to work with Ofcom and with companies looking to do the right thing to comply with the new laws.
“It’s right that protecting children and ensuring the spread of child sexual abuse imagery is stopped is top of the agenda.
“It’s vital companies are proactive in assessing and understanding the potential risks on their platforms and taking steps to make sure safety is designed in.
“Making the internet safer does not end with this Bill becoming an Act. The scale of child sexual abuse and the harms children are exposed to online, have escalated in the years this legislation has been going through Parliament.
“Companies in the scope of the regulations now have a huge opportunity to be part of a real step forward in terms of child safety.”