fbpx

Type to search

China Joins US, EU in Vow to Tackle ‘Catastrophic’ AI Harm Risk

The three, along with 26 other other signatories signed the “Bletchley Declaration” at the AI Safety Summit in Britain on Wednesday


US Secretary of Commerce Gina Raimondo and British Secretary of State for Science, Innovation and Technology Michelle Donelan listen to China's Vice Minister of Science and Technology Wu Zhaohui speaking on Day 1 of the AI Safety Summit at Bletchley Park in Bletchley, Britain
US Secretary of Commerce Gina Raimondo and British Secretary of State for Science, Innovation and Technology Michelle Donelan listen to China's Vice Minister of Science and Technology Wu Zhaohui speaking on Day 1 of the AI Safety Summit at Bletchley Park in Bletchley, England. Photo: Reuters

 

In its first such collaboration with the West on artificial intelligence (AI), China has agreed to work with the United States and Europe to assess and address the risk of ‘catastrophic’ harm from the technology.

The three, along with 26 other signatories signed the “Bletchley Declaration” at the AI Safety Summit in the Britain on Wednesday aimed at charting a safe way forward for AI.

The summit and the declaration come on the heels of warnings from some tech executives, including Elon Musk, and political leaders that the rapid development of AI poses an existential threat to the world.

 

Also on AF: Big Tech Exaggerating AI’s Threat to Humanity, Expert Says

 

Those warnings have sparked a race by governments world-over and international institutions to design safeguards and regulation.

“We resolve to work together in an inclusive manner to ensure human-centric, trustworthy and responsible AI that is safe, and supports the good of all through existing international fora and other relevant initiatives,” the Bletchley Declaration said.

“Noting the importance of inclusive AI and bridging the digital divide, we reaffirm that international collaboration should endeavour to engage and involve a broad range of partners as appropriate.”

The declaration set out a two-pronged agenda focused on identifying risks of shared concern and building scientific understanding of them, while also developing cross-country policies to mitigate them.

It is a first for Western efforts to manage AI’s safe development.

Wu Zhaohui, China’s vice minister of science and technology joined the US and EU leaders and tech bosses such as Musk and ChatGPT’s Sam Altman at Bletchley Park, home of Britain’s World War Two code-breakers.

Other Asian signatories to the pact included India, Indonesia, Japan, The Philippines, Singapore and South Korea.

 

Questions on China’s presence

Wu Zhaohui, China’s vice minister of science and technology, told the opening session of the two-day summit that Beijing was ready to increase collaboration on AI safety to help build an international “governance framework”.

“Countries regardless of their size and scale have equal rights to develop and use AI,” he said.

China is a key participant at the summit, given the country’s role in developing AI. But some British lawmakers have questioned whether it should be there given the low level of trust between Beijing, Washington and many European capitals when it comes to Chinese involvement in technology.

The United States made clear on the eve of the summit that the call to Beijing had very much come from Britain

“This is the UK invitation, this is not the US,” American ambassador to London Jane Hartley said.

US Vice President Kamala Harris was in attendance at the summit used the British summit to announce the launch of a US AI Safety Institute. The announcement came just days after US President Joe Biden signed an executive order on AI on Monday.

 

 

View this post on Instagram

 

A post shared by Asia Financial (@asiafinancial)

 

ChatGPT-fuelled fears

While the European Union has focused its AI oversight on data privacy and surveillance and their potential impact on human rights, the British summit is looking at so-called existential risks from highly capable general-purpose models called “frontier AI”.

“We are especially concerned by such risks in domains such as cybersecurity and biotechnology, as well as where frontier AI systems may amplify risks such as disinformation. There is potential for serious, even catastrophic, harm, either deliberate or unintentional, stemming from the most significant capabilities of these AI models,” the Bletchley Declaration says.

Mustafa Suleyman, the cofounder of Google Deepmind, told reporters he did not think current AI frontier models posed any “significant catastrophic harms”.

He said, however, it made sense to plan ahead as the industry trains ever larger models.

Fears about the impact AI could have on economies and society took off in November last year when Microsoft-backed OpenAI made ChatGPT available to the public.

Using natural language processing tools to create human-like dialogue, it has stoked fears, including among some AI pioneers, that machines could in time achieve greater intelligence than humans, leading to unlimited, unintended consequences.

 

Next summit in South Korea

Governments and officials are now trying to chart a way forward alongside AI companies which fear being weighed down by regulation before the technology reaches its full potential.

“I don’t know what necessarily the fair rules are, but you’ve got to start with insight before you do oversight,” tech billionaire Musk told reporters at the summit.

A “third-party referee” could be used to sound the alarm when risks develop, he added.

Meanwhile, British digital minister Michelle Donelan said it was an achievement just to get so many key players in one room. She announced two further AI Safety Summits, one to be held in South Korea in six months and another in France six months after that.

“For the first time, we now have countries agreeing that we need to look not just independently but collectively at the risk around frontier AI,” Donelan told reporters.

 

  • Reuters, with additional inputs from Vishakha Saxena

 

Also read:

 

G7 Agree AI Code of Conduct to Limit Tech Threat Risks

 

Beijing Unveils Sweeping Laws to Regulate ChatGPT-Like AI Tech

 

China Wants To ‘Integrate’ ChatGPT-Like Tech in its Economy

 

Foxconn, Nvidia to Build ‘AI Factories’ Producing Intelligence

 

Western Spy Chiefs Warn China Using AI to Steal Tech Secrets

 

OpenAI Boss Urges Regulations to Prevent ‘Harm to the World’

 

$600,000 ‘Deepfake’ Fraud Heats Up AI Debate in China

 

Vishakha Saxena

Vishakha Saxena is the Multimedia and Social Media Editor at Asia Financial. She has worked as a digital journalist since 2013, and is an experienced writer and multimedia producer. As a trader and investor, she is keenly interested in new economy, emerging markets and the intersections of finance and society. You can write to her at [email protected]