AI safety is hard to steer with science in flux, US official says
By Jeffrey Dastin
NEW YORK (Reuters) – Policymakers aiming to recommend safeguards for artificial intelligence are facing a formidable challenge: science that is still evolving.
AI developers themselves are grappling with how to prevent abuse of novel systems, offering no easy fix for government authorities to embrace, Elizabeth Kelly, director of the U.S. Artificial Intelligence Safety Institute, said on Tuesday.
Cybersecurity is an area of concern according to Kelly, speaking at the Reuters NEXT conference in New York. Ways to bypass guard rails that AI labs established for security and other topics, called “jailbreaks,” can be easy, she said.
“It is difficult for policymakers to say these are best practices we recommend in terms of safeguards, when we don’t actually know which ones work and which ones don’t,” Kelly said.
Technology experts are hashing out how to vet and protect AI across different dimensions. Another area regards synthetic content. Tampering with digital watermarks, which flag to consumers when images are AI-generated, remains too easy for authorities to devise guidance for industry, she said.
The U.S. AI Safety Institute, created under the Biden administration, is addressing such concerns via academic, industry and civil society partnerships that inform its tech evaluations, Kelly said. She said AI safety is a “fundamentally bipartisan issue,” when asked what will happen to the body after Donald Trump takes office in January.
The institute’s first director, Kelly recently presided over the inaugural gathering of AI safety institutes from around the world, which took place last month in San Francisco.
Asked about the outcome of these meetings, Kelly said the 10 country members were working toward interoperable safety tests with the help of more technical, hoodie-wearing experts than in a typical diplomatic meeting.
“It was very much getting the nerds in the room,” she said.