Sixteen global tech giants, including Alphabet’s Google, Meta, Microsoft and OpenAI, have vowed to commit themselves to the safe development of AI.
The pledge, which also saw firms from China, South Korea and the United Arab Emirates sign up, was announced in a UK government statement on Tuesday, as South Korea and Britain co-hosted a global AI summit in Seoul.
The agreement, which comes at a time when the breakneck pace of AI innovation is leaving governments scrambling to keep up, is a step up from the commitments made at the first global AI summit held six months ago, the statement said.
Also on AF: Aussie Who Said he Invented Bitcoin Lied on ‘Grand Scale’: Judge
Zhipu.ai, backed by Chinese tech giants Alibaba, Tencent, Meituan and Xiaomi, as well as UAE’s Technology Innovation Institute, were among the 16 companies pledging to publish safety frameworks on how they will measure risks of frontier AI models.
The firms, also including Amazon, IBM and Samsung Electronics, voluntarily committed to not develop or deploy AI models if the risks cannot be sufficiently mitigated, and to ensure governance and transparency on approaches to AI safety, the statement said.
“It’s vital to get international agreement on the ‘red lines’ where AI development would become unacceptably dangerous to public safety,” said Beth Barnes, founder at METR, a non-profit for AI model safety.
The artificial intelligence (AI) summit in Seoul this week aims to build on a broad agreement at the first summit held in the United Kingdom to better address a wider array of risks.
At the November summit, Tesla’s Elon Musk and OpenAI CEO Sam Altman mingled with some of their fiercest critics, while China co-signed the “Bletchley Declaration,” on collectively managing AI risks alongside the United States and others.
British Prime Minister Rishi Sunak and South Korean President Yoon Suk Yeol oversaw a virtual summit later on Tuesday, followed by a ministerial session on Wednesday.
This week’s summit will address “building… on the commitment from the companies, also looking at how the [AI safety] institutes will work together,” Britain’s Technology Secretary Michelle Donelan said on Tuesday.
AI in Medicine, Finance Concerns
Since November, discussion on AI regulation has shifted from longer-term doomsday scenarios to “practical concerns” such as how to use AI in areas like medicine or finance, said Aidan Gomez, co-founder of large language model firm Cohere.
Industry participants wanted AI regulation that will give clarity and security on where the companies should invest, while avoiding entrenching big tech, Gomez said.
With countries such as the UK and US establishing state-backed AI Safety Institutes for evaluating AI models and others expected to follow suit, AI firms are also concerned about the interoperability between jurisdictions, analysts said.
Representatives of the Group of Seven (G7) major democracies are expected to take part in the virtual summit, while Singapore and Australia were also invited, a South Korean presidential official said.
China will not participate in the virtual summit but is expected to attend Wednesday’s in-person ministerial session, the official said.
South Korea’s foreign ministry said Musk, former CEO of Google Eric Schmidt, Samsung Electronics’ Chairman Jay Y. Lee and other AI industry leaders will participate in the summit.
- Reuters with additional editing by Sean O’Meara
Read more:
Alibaba, Baidu Slash Price of Language Models Used With AI Bots
US Trials of AI-Piloted Fighter Jets Worrying China – ZM
Microsoft Asks China AI Staff if They Want to Shift Abroad
Chinese Firms Close in on High-End AI Memory Chips Coup
AI Hitting Job Markets Like a ‘Tsunami’, Says IMF Chief