SAN FRANCISCO -- President-elect Donald Trump has vowed to repeal President Joe Biden's signature artificial intelligence policy when he returns to the White House for a second term.
What that actually means for the future of AI technology remains to be seen. Among those who could use some clarity are the government scientists and AI experts from multiple countries gathering in San Francisco this week to deliberate on AI safety measures.
Hosted by the Biden administration, officials from a number of U.S. allies -- among them Australia, Canada, Japan, Kenya, Singapore, the United Kingdom and the 27-nation European Union -- began meeting Wednesday in the California city that's a commercial hub for AI development.
Their agenda addresses topics such as how to better detect and combat a flood of AI-generated deepfakes fueling fraud, harmful impersonation and sexual abuse.
It's the first such meeting since world leaders agreed at an AI summit in South Korea in May to build a network of publicly backed safety institutes to advance research and testing of the technology.
Hong Yuen Poon, deputy secretary of Singapore's Ministry of Digital Development and Information, said Wednesday that a "helping-one-another mindset is important" between countries when it comes to AI safety, including with "developing countries which may not have the full resources" to study it.
Biden signed a sweeping AI executive order last year and this year formed the new AI Safety Institute at the National Institute for Standards and Technology, which is part of the Commerce Department.
Trump promised in his presidential campaign platform to "repeal Joe Biden's dangerous Executive Order that hinders AI Innovation, and imposes Radical Leftwing ideas on the development of this technology."
But he hasn't made clear what about the order he dislikes or what he'd do about the AI Safety Institute. Trump's transition team didn't respond to emails this week seeking comment.
Tech industry groups -- backed by companies including Amazon, Google, Meta and Microsoft -- are mostly pleased with the AI safety approach of Biden's Commerce Secretary Gina Raimondo, which has focused on setting voluntary standards. They have pushed for Congress to preserve the new agency and codify its work into law.
Some experts expect the kind of technical work happening at an old military officers' club at San Francisco's Presidio National Park this week to proceed regardless of who's in charge.
"There's no reason to believe that we'll be doing a 180 when it comes to the work of the AI Safety Institute," said Heather West, a senior fellow at the Center for European Policy Analysis. Behind the rhetoric, she said there's already been overlap.
Trump didn't spend much time talking about AI during his four years as president, though in 2019 he became the first to sign an executive order about AI. It directed federal agencies to prioritize research and development in the field.
Before that, tech experts were pushing the Trump-era White House for a stronger AI strategy to match what other countries were pursuing. Trump in the waning weeks of his administration signed an executive order promoting the use of "trustworthy" AI in the federal government. Those policies carried over into the Biden administration.
All of that was before the 2022 debut of ChatGPT, which brought public fascination and worry about the possibilities of generative AI and helped spark a boom in AI-affiliated businesses. What's also different this time is that tech mogul and Trump adviser Elon Musk has been picked to lead a government cost-cutting commission. Musk holds strong opinions about AI's risks and grudges against some AI industry leaders, particularly ChatGPT maker OpenAI, which he has sued.