AI Safety: UK and US sign landmark agreement
- Published
- comments
The UK and US have signed a landmark deal to work together on testing advanced artificial intelligence (AI).
The agreement signed on Monday says both countries will work together on developing "robust" methods for evaluating the safety of AI tools and the systems that underpin them.
It is the first bilateral agreement of its kind.
UK tech minister Michelle Donelan said it is "the defining technology challenge of our generation".
"We have always been clear that ensuring the safe development of AI is a shared global issue," she said.
"Only by working together can we address the technology's risks head on and harness its enormous potential to help us all live easier and healthier lives."
The secretary of state for science, innovation and technology added that the agreement builds upon commitments made at the AI Safety Summit held in Bletchley Park in November 2023.
The event, attended by AI bosses including OpenAI's Sam Altman, Google DeepMind's Demis Hassabis and tech billionaire Elon Musk, saw both the UK and US create AI Safety Institutes which aim to evaluate open and closed-source AI systems.
While things have felt quiet on the AI safety front since the summit, the AI sector itself has been extremely busy.
Competition between the biggest AI chatbots - such as ChatGPT, Gemini, Claude - remains ferocious.
So far the almost exclusively US-based firms behind all of this activity are still cooperating with the concept of regulation, but regulators have yet to curtail anything these companies are trying to achieve.
Similarly, regulators have not demanded access to information the AI firms are unwilling to share, such as the data used to train their tools orthe environmental cost of running them.
TheEU's AI Act is on its way to becoming law and once it takes effect it will require developers of certain AI systems to be upfront about their risks and share information about data used.
This is important, after OpenAIrecently said, externalit would not release a voice cloning tool it developed due to "serious risks" the tech presents, particularly in an election year.
In January, a fake, AI-generated robocall claiming to be from US President Joe Biden urged voters to skip a primary election in New Hampshire.
Currently in the US and UK, AI firms are mostly regulating themselves.
AI concerns
Currently, the majority of AI systems are only capable of performing single, intelligent tasks that would usually be completed by a human.
Known as "narrow" AI, these tasks can range from quickly analysing data or providing a desired response to a prompt.
But there are fears that more intelligent "general" AI tools - capable of completing a range of tasks usually performed by humans - could endanger humanity.
"AI, like chemical science, nuclear science, and biological science, can be weaponised and used for good or ill," Prof Sir Nigel Shadbolt told the BBC's Today programme.
But the University of Oxford professor said fears around AI's existential risk "are sometimes a bit overblown".
"We've got to be really supportive and appreciative of efforts to get great AI powers thinking about and researching what the dangers are," he said.
"We need to understand just how susceptible these models are, and also how powerful they are."
Gina Raimondo, the US commerce secretary, said the agreement will give the governments a better understanding of AI systems, which will allow them to give better guidance.
"It will accelerate both of our Institutes' work across the full spectrum of risks, whether to our national security or to our broader society," she said.
"Our partnership makes clear that we aren't running away from these concerns - we're running at them."
Related topics
- Published29 March
- Published13 March