Children contacting Childline about AI-generated online harms, NSPCC says

The children’s charity has launched a new partnership with a US-based safety group to lobby for child safety rules in the development of AI tools.

Martyn Landi
Thursday 25 January 2024 00:01 GMT
Debate around how to safely develop AI tools has intensified over the last 12 months (Dominic Lipinski/PA)
Debate around how to safely develop AI tools has intensified over the last 12 months (Dominic Lipinski/PA) (PA Wire)

Your support helps us to tell the story

From reproductive rights to climate change to Big Tech, The Independent is on the ground when the story is developing. Whether it's investigating the financials of Elon Musk's pro-Trump PAC or producing our latest documentary, 'The A Word', which shines a light on the American women fighting for reproductive rights, we know how important it is to parse out the facts from the messaging.

At such a critical moment in US history, we need reporters on the ground. Your donation allows us to keep sending journalists to speak to both sides of the story.

The Independent is trusted by Americans across the entire political spectrum. And unlike many other quality news outlets, we choose not to lock Americans out of our reporting and analysis with paywalls. We believe quality journalism should be available to everyone, paid for by those who can afford it.

Your support makes all the difference.

Young people are contacting Childline about AI-generated child sexual abuse material, the NSPCC has said as it announced a new partnership aimed at championing child safety during the development of new AI technology.

The children’s charity said young people had also contacted Childline about harm linked to generative AI, including the spread of misinformation and bullying.

It comes as the charity unveiled a new advocacy partnership with US children’s safety group Common Sense Media which will see the groups look to pressure tech firms and politicians over ensuring child safety and wellbeing are central to decision-making around artificial intelligence.

Debate around how to safely develop AI tools has intensified over the last 12 months as the technology has become increasingly mainstream, thanks to the rise of generative AI chatbots such as ChatGPT.

The risk children face from unregulated and unsafe AI is already far too high, and their safety and experiences must be at the centre of conversations about its development and regulations

Sir Peter Wanless, NSPCC

As part of the new partnership, the NSPCC and Common Sense Media will also roll out education programmes designed to increase digital literacy skills in schools and carry out more joint research on the impact of generative AI on children.

NSPCC chief executive Sir Peter Wanless said: “As one of the leading voices helping to achieve the Online Safety Act in the UK, we have long acknowledged the need for global collaboration by governments, civil society and tech firms to drive children’s safety online.

“This cannot be clearer than in AI where a rush to glean the significant benefits of technology has led to worldwide concerns about the danger it can also pose.

“The risk children face from unregulated and unsafe AI is already far too high, and their safety and experiences must be at the centre of conversations about its development and regulations.

“This partnership will seek to do that while also empowering young people with digital literacy skills to help them thrive.”

James P Steyer, founder and chief executive of Common Sense Media, said: “To safely and responsibly harness the potential of AI for children, global organisations must work together to ensure that the government and private sector have children’s best interests at heart for all technology design, development, and deployment.

“The NSPCC, and the UK more broadly, is seeing great progress by activating its educators, parents, and policymakers, and we look forward to amplifying their efforts in the US and globally.”

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in