Fears over AI dangers are hindering the chance to create a better future

fears over ai dangers are hindering the chance to create a better future

Last week, a breath of optimism swept through the news as the United States and China made headlines for encouraging reasons. The highly anticipated meeting between President Xi Jinping and US President Joe Biden concluded with both nations expressing a shared understanding across a broad spectrum of interests. This included not only traditional sectors such as the economy, trade and agriculture but also burgeoning fields such as artificial intelligence.

The discussion on AI seems to resonate with the appeal for intergovernmental cooperation that Henry Kissinger and Graham Allison proposed last month. Echoing those sentiments, there were hopes the two powers would agree on AI arms control or at least a consensus to exclude AI from nuclear command systems.

The enduring peace among nuclear nations since 1945 underlines a mutual recognition of nuclear perils. AI technology has amplified these risks, making their non-use a matter of course. However, the path to effective AI governance is complex and the theory of knowledge differs vastly from practical epistemology, especially in the unpredictable theatre of war.

AI is more than just a substitute for specific human functions. Its essence is shaped by the data it processes and the complexity of its algorithms. Unlike commercial applications, nuclear weapons in warfare cannot readily amass extensive training data. The confidentiality of enemy information and the dynamic nature of warfare further complicate data acquisition.

Even the most advanced AI algorithms fall short of emulating the nuanced decision-making of an experienced general who bases decisions on a wealth of professional knowledge and real-life experience. This differs significantly from current discussions about the military application of AI, often encapsulated in the “human in the loop” concept.

Algorithms and software are very intricate, so it’s difficult to decide who should be in which loop and how they should interact. Moreover, even in a case where China and the US reach an agreement, another conundrum arises. How could we verify compliance, and would any nation permit external inspection of their nuclear launch system’s software codes?

The discourse on AI risks extends beyond military applications. The prevailing debate in China suggests that technological development is often equated with safety and security. AI, while a captivating and succinct term, isn’t a single technology but a fusion of various technologies.

As such, it’s not appropriate to discuss AI as a generic whole in terms of governance needs. Many AI types and applications don’t threaten human safety and should be actively developed. However, particular safety risks arise at the frontier of AI, a concern endorsed by the 29 signatories of the Bletchley Declaration at the UK AI Safety Summit earlier this month.

Current discussions around AI governance focus on misinformation, intellectual property, privacy and virality. However, these challenges aren’t new. False information has existed in media such as newspapers, radio and television since their inception.

Today’s debates about intellectual property echo those from a decade ago, including fears that electronic books would destroy the publishing industry. AI’s evolution certainly intensifies or updates these issues and could lead to undesirable outcomes such as algorithms amplifying echo chambers in social media.

This leads to the question of whether the advancement of AI by leading firms can protect humanity from the very threats AI poses. In my classroom, when discussing topics such as energy, climate or public health, my socially conscious students often express dismay at corporations prioritising profits over the public interest.

But merely expressing outrage or attempting to shame companies during a few class sessions achieves little. Corporations exist through legal mandates and societal frameworks. What they need are better incentive mechanisms that reward beneficial actions and penalise harmful ones.

What’s more, technology isn’t the source of all evil, and neither should it be a scapegoat for flawed human decision-making. If we outsource moral and political choices to intelligent machines, we ignore the warning of cybernetics pioneer Norbert Wiener – that these responsibilities will ultimately return to us.

From an international governance perspective, if there’s anything to learn from nuclear arms control, it’s that humanity is capable of figuring out how to manage emerging, unpredictable technologies. Given the rapid evolution of AI and the diverse governance approaches worldwide, our focus should shift to adaptable cooperation.

Rather than pursuing a one-size-fits-all standard or approach to governance, the true challenge lies in consistently finding the broadest common ground for ensuring safety amid these risks. This requires a dynamic balance of security, development, prosperity and fairness.

In the realm of technological governance, simplicity is a myth. The challenge with emerging technologies demands a comprehensive strategy involving skill development, robust support systems and cultural assimilation.

Take nuclear energy, automobiles, aviation and pharmaceuticals. These are not mere triumphs of innovation but milestones of a long, arduous journey towards safety and civility. This complex endeavour requires flexible, sustained collaboration across diverse stakeholders. The recent US-China consensus and declaration at the UK AI Safety summit herald promising starts.

However, there’s a crucial aspect eluding our current discourse on AI governance. The rush to establish rules for managing AI’s risks is predicated on the assumption that AI will soon outpace human intellect.

But can an entity surpassing human intelligence ever be fully subjugated by human control? This might not be explained simply by a path-dependent model of technological governance but rather by a deeper reflection that our perception of technology as a tool of power has obscured its true nature.

We’ve been so captivated by AI’s ability to empower specific people and institutions to create new forms of dominance that we might be overlooking an unprecedented opportunity. AI could be the answer to age-old aspirations to forge a future where we coexist with technology in a world of grace and justice.

Dong Ting is a resident research fellow at the Centre for International Security and Strategy, Tsinghua University

News Related

OTHER NEWS

China’s digitalisation push seeks ‘digital energy’ to spark vital tech, economic growth in Greater Bay Area

China is converging its manufacturing, trade and technology strengths in the Greater Bay Area development zone to spearhead digitalisation, aiming to break through technology curbs from the United States and ... Read more »

Pinduoduo ramps up investment in AI with job posts for LLM engineers as it eyes greater adoption

Budget shopping platform Pinduoduo, operated by PDD Holdings, is ramping up investment in artificial intelligence (AI) after posting job adverts for various positions, including algorithm engineers. Pinduoduo is offering up ... Read more »

Baidu forecasts AI revenue growth and limited short-term impact from US chip curbs

Baidu has a stockpile of advanced artificial intelligence (AI) chips to soften the short-term impact of US export curbs, the Chinese tech giant said as it forecast increased AI revenues ... Read more »

Tech war: Nvidia warns of sales hit in China from US chip curbs, as tailor-made products take time to develop

US semiconductor giant Nvidia said on Wednesday it was developing new products for China-based customers, but the process would take time because of the need to formulate chips that meet ... Read more »

Video game streamer Douyu confirms CEO’s arrest in China after weeks of speculation about gambling content

Police in Chengdu, the capital of southwest China’s Sichuan province, said on Wednesday they had arrested a 39-year-old man surnamed Chen on suspicion of opening a casino, a day after ... Read more »

Tech war: China approves Broadcom-VMware merger with conditions, in sign of thaw with US

Broadcom’s US$61 billion takeover of software maker VMware won approval from Chinese regulators, albeit with a list of conditions it must fulfil. Still, the nod from China completes the last ... Read more »

North Korea says it put a military spy satellite into orbit on third try

North Korea says it put a military spy satellite into orbit on third try North Korea says it has put a military spy satellite into orbit on its third attempt ... Read more »
Top List in the World