UK AI Safety Institute to open San Francisco office

The institute is looking to establish closer ties with the tech sector in Silicon Valley.

Martyn Landi
Monday 20 May 2024 10:02 BST
Rishi Sunak attended the AI Safety Summit at Bletchley Park (PA)
Rishi Sunak attended the AI Safety Summit at Bletchley Park (PA) (PA Archive)

Your support helps us to tell the story

From reproductive rights to climate change to Big Tech, The Independent is on the ground when the story is developing. Whether it's investigating the financials of Elon Musk's pro-Trump PAC or producing our latest documentary, 'The A Word', which shines a light on the American women fighting for reproductive rights, we know how important it is to parse out the facts from the messaging.

At such a critical moment in US history, we need reporters on the ground. Your donation allows us to keep sending journalists to speak to both sides of the story.

The Independent is trusted by Americans across the entire political spectrum. And unlike many other quality news outlets, we choose not to lock Americans out of our reporting and analysis with paywalls. We believe quality journalism should be available to everyone, paid for by those who can afford it.

Your support makes all the difference.

The Government’s AI Safety Institute is to open an overseas office in San Francisco to get closer to the tech giants based in Silicon Valley.

The Department for Science, Innovation and Technology (DSIT) said the new office is expected to open in the summer, and would look to engage with the world’s biggest AI labs, which are based in the region.

The UK’s AI Safety Institute was first announced at the AI Safety Summit at Bletchley Park last year and, alongside several similar international institutes, will carry out research and conduct evaluations of AI models to help inform AI safety policy.

The institute has released its first selection of recent results from safety testing five publicly available advanced AI models, although the institute said it had anonymised the results.

This expansion represents British leadership in AI in action

Michelle Donelan, Technology Secretary

In its findings, the institute said several models had completed cyber security challenges but struggled to complete more advanced challenges, and several models had also shown similar to PhD-level knowledge of chemistry and biology.

The institute said all the models it tested remain highly vulnerable to basic “jailbreaks” to sidestep safety controls, and some will produce harmful outputs even without dedicated attempts to get around safety mechanisms.

The testing also noted that it found models were unable to complete more complex, time-consuming tasks without human oversight.

The announcements come ahead of the AI Seoul Summit, which the UK is co-hosting with South Korea, with Prime Minister Rishi Sunak and Technology Secretary Michelle Donelan taking part along with international counterparts and tech firms.

On the expansion of the AI Safety Institute, Ms Donelan said: “This expansion represents British leadership in AI in action.

“It is a pivotal moment in the UK’s ability to study both the risks and potential of AI from a global lens, strengthening our partnership with the US and paving the way for other countries to tap into our expertise as we continue to lead the world on AI safety.

Our ambition is to continue pushing the frontier of this field by developing state-of-the-art evaluations, with an emphasis on national security-related risks

Ian Hogarth, AI Safety Institute chairman

“Since the Prime Minister and I founded the AI Safety Institute, it has grown from strength to strength and in just over a year, here in London, we have built the world’s leading Government AI research team, attracting top talent from the UK and beyond.

“Opening our doors overseas and building on our alliance with the US is central to my plan to set new, international standards on AI safety which we will discuss at the Seoul Summit this week.”

AI Safety Institute (AISI) chairman Ian Hogarth said its ongoing testing of models would help with the development of better safety controls in the sector.

“The results of these tests mark the first time we’ve been able to share some details of our model evaluation work with the public,” he said.

“Our evaluations will help to contribute to an empirical assessment of model capabilities and the lack of robustness when it comes to existing safeguards.

“AI safety is still a very young and emerging field. These results represent only a small portion of the evaluation approach AISI is developing.

“Our ambition is to continue pushing the frontier of this field by developing state-of-the-art evaluations, with an emphasis on national security-related risks.”

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in