‘Godfather of AI’ quits Google and gives warning about the future of technology

‘I don’t think they should scale this up more until they have understood whether they can control it,’ Geoffrey Hinton says

Andrew Griffin,Graeme Massie
Wednesday 03 May 2023 14:04 BST
Comments
AI ‘godfather’ warns of ‘existential risk’ of robotic intelligence after quitting Google

Your support helps us to tell the story

This election is still a dead heat, according to most polls. In a fight with such wafer-thin margins, we need reporters on the ground talking to the people Trump and Harris are courting. Your support allows us to keep sending journalists to the story.

The Independent is trusted by 27 million Americans from across the entire political spectrum every month. Unlike many other quality news outlets, we choose not to lock you out of our reporting and analysis with paywalls. But quality journalism must still be paid for.

Help us keep bring these critical stories to light. Your support makes all the difference.

Geoffrey Hinton, known as the “godfather of AI”, has quit Google with a warning about the dangers the technology poses.

Geoffrey Hinton worked at Google for more than a decade and developed technology that paved the way for current AI systems such as ChatGPT. However, the tech expert has since quit the company so that he could sound the alarm about the dangers of artificial intelligence.

He told The New York Times that he regrets the work he contributed to the field. “I console myself with the normal excuse: If I hadn’t done it, somebody else would have,” he said.

In the short term, he fears the technology will mean that people will “not be able to know what is true anymore” because of the proliferation of fake images, videos and text, he said.

But in the future, AI systems could eventually learn unexpected, dangerous behaviour, and that such systems will eventually power killer robots. He also warned that the technology could cause harmful disruption to the labour market.

“The idea that this stuff could actually get smarter than people — a few people believed that,” he says. “But most people thought it was way off. And I thought it was way off. I thought it was 30 to 50 years or even longer away. Obviously, I no longer think that.”

He warned that regulation is required to ensure that companies such as Google and Microsoft do not get locked into a dangerous race. Those companies might already be working on dangerous systems in secret, he suggested.

“I don’t think they should scale this up more until they have understood whether they can control it,” he said.

Dr Hinton is not the first artificial intelligence expert to sound the alarm on the dangers of the technologies that they have built. In recent months, two major open letters warned about the “profound risks to society and humanity” that it poses, and was signed by many of the people who had helped create it.

Like many others, Dr Hinton said he had become more concerned about the risks of the technology over the last year, with the development of technologies such as OpenAI’s ChatGPT and Google’s Bard. His view changed when he came to believe that the systems were beginning to behave in ways that were not possible in the human brain.

“Maybe what is going on in these systems is actually a lot better than what is going on in the brain,” he told the New York Times.

That is likely to become even more true in the years to come, as companies further refine and train their AI systems. As they do, the systems become more dangerous, he fears.

“Look at how it was five years ago and how it is now,” he said of the technology. “Take the difference and propagate it forwards. That’s scary.”

Join our commenting forum

Join thought-provoking conversations, follow other Independent readers and see their replies

Comments

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in