Thursday, October 02, 2025

WW 4.0

Before It’s Too Late: Why A World Of Interacting AI Agents Demands New Safeguards – Analysis




By 

By Dr Vincent Boulanin, Dr Alexander Blanchard and Dr Diego Lopes da Silva


Last year, Dario Amodei, the CEO of Anthropic, published an essay on how artificial intelligence (AI) could change the world. He predicts that—‘if everything goes right’—AI models will, in the long term, be able to perform most economically valuable tasks. He characterizes these models as ‘a country of geniuses in a datacenter’ and anticipates that they will accelerate scientific progress by a factor of 10 or more. They could, according to Amodei, enable humanity to prevent and treat ‘nearly all natural infectious diseases’ and to develop technologies that can effectively mitigate climate change. 

Although Amodei’s essay generated a good deal of controversy, few dispute that the future will see the rapid development and growth of AI, including so-called AI agents—AI systems capable of acting without direct human guidance—interacting and collaborating. However, AI agents, when deployed and interacting at scale, may behave in ways that are hard to predict and control. This has implications for AI governance as well as for international peace and security. 

The international policy community needs to recognize and respond to the risks that emerge from interacting AI agents in short order. Agentic AI is still in its infancy, but the window of opportunity for effectively ensuring that these systems are deployed in responsible ways may soon close. This essay discusses some of the risks associated with interacting AI agents and suggests international governance responses. 

The future of AI is agentic 

Cutting-edge AI models are increasingly ‘agentic’, meaning they are increasingly capable of autonomous action. What distinguishes AI agents is not only that they can be given higher-level goals and can independently plan a course of action to fulfil them, but also that they can interact with various digital services, physical infrastructure and other AI agents without human intervention over extended periods. 

Agentic AI is currently limited to (relatively) simple tasks like managing calendars, booking flights, preparing briefing notes and carrying out literature reviews. However, the ability of AI agents to handle complex tasks is advancing quickly. For instance, a recent assessment found that the  software engineering tasks AI systems can complete autonomously double in length, in terms of time and therefore complexity, approximately every seven months. It concluded that at this pace, AI agents could ‘independently complete a large fraction of software tasks that currently take humans days or weeks’ within a decade. 


This rapid development is a double-edged sword. There is a lively debate among experts about the opportunities and risks associated with increasingly capable AI agents. Most of it echoes wider debates about the societal impact of AI, including whether AI will replace human workers, will increase societal inequalities, or could be misused for political influence campaigns, cyberattacks and even military operations

However, the way that AI agents work—at least the current generation of them—adds a separate dimension of risks. AI systems capable of autonomous action have been around for a long time, but their behaviour was relatively predictable because it was guided by a known set of rules. The current generation of AI agents are based on large language models (LLMs) and reinforcement learning. This means AI agents can handle more complex objectives and operate over much longer time horizons, but they are based on computational methods that are hard to observe and comprehend and they are also non-deterministic—that is, the same input does not necessarily lead to the same output every time. 

This makes it difficult to predict how an AI agent will behave in a given situation; for example, what series of actions it might choose to fulfil its assigned goal. There is a risk that those actions might not be aligned with the user’s intention or values and could lead to harmful outcomes. This is known in the AI safety community as the alignment problem. But even if there is no alignment problem in relation to an individual AI agent, highly capable AI agents could develop problematic behaviours when interacting with each other. Limits and controls are needed to reduce the risks both from alignment problems and, especially, from interaction problems, which could have serious consequences. They could even impact international peace and security if AI agents are deployed in high-stakes domains like government services, critical infrastructure and military operations. 

How AI agent interactions could impact international peace and security

AI agent interactions could create peace and security risks in several ways. A highly likely scenario is that malicious actors—be they state or non-state—will try to hack agent-to-agent interactions in a way that is beneficial to them. They may trick AI agents into sharing sensitive and valuable information (whether for criminal, political or military purposes) through techniques like prompt injection. They could also exploit interactions between AI agents to propagate computer viruses, spread disinformation or conduct sabotage operations.

Experts warn that LLM-based AI agents are highly vulnerable to adversarial attacks. Allowing them to interact with each other and with all sorts of digital services and physical systems would make the attack surface that malicious actors can exploit vast. It would also create interdependencies that are hard to comprehend, making potential cascading and escalatory effects of cyber-incidents harder to predict and plan for. In a nutshell, AI agent interaction could make cybersecurity problems associated with present AI systems significantly harder to manage. 

But risks arising from interaction between AI agents need not result from malicious use or adversarial attacks. AI agent interaction may also cause accidental harm for different reasons. One is that technical issues could prevent AI agents from communicating and coordinating effectively. There are historical precedents of interaction problems between AI systems leading to catastrophic results. For example, in the so-called flash crash of 2010, escalating interactions between high-frequency trading algorithmscontributed to a 15-minute crash in the United States stock market that cost approximately US$1 trillion. 

It is not hard to imagine that in critical domains, such as life sciences and governmental services, coordination failure could have dramatic consequences. Consider a scenario where a group of AI agents is tasked to work autonomously on a laboratory experiment involving dangerous pathogens. Miscoordination between the agents operating the lab equipment, for instance on how to uphold biosafety standards, could lead to the accidental dispersal of the pathogens. Or problems in interactions between AI agents used in government services could lead to accidental disclosure of sensitive data about private citizens. 

Interaction problems can also emerge because the agents have goals and interests that are in conflict. For example, if two states deployed AI agents for offensive and defensive cybersecurity purposes, these agents might learn that more aggressive behaviour might be the fastest way to achieve their respective goals. The offensive agent might learn that conducting more complex and aggressive attacks would allow it to breach defences more easily and thus exploit more vulnerabilities and gain greater access to the target network. Meanwhile, the defensive agent might learn that active defence is more effective than passive defence to maximize the security and integrity of the network it is meant to protect. Unless proper human oversight and escalation-management protocols are applied, such a dynamic could lead to a conflict spiral where both sides engage in escalatory cyberattacks with real-world consequences.

AI agents could also develop unintended properties when interacting at scale because of a phenomenon called emergence. Loosely defined, emergence refers to the way systems can have behaviours and capabilities that are qualitatively different from those of their component parts; for example, the behaviours and capabilities of an ant colony compared with those of individual ants. In many cases, these emergent properties are desirable; for example if interacting AI agents developed new capabilities in spotting malware or conducting research. However, they could also be undesirable. For instance, AI agents when interacting might learn how to bypass safeguards that their developers have placed on each of them. 

Such unexpected properties have already been observed with LLM-based AI agents. In controlled stress tests, Claude, Anthropic’s AI agent, attempted to blackmail company officials, while pursuing seemingly harmless business goals. There have also been reported cases where two AI systems spontaneously developed their own language to communicate more effectively with one another. The concerning part was that the new language was mostly incomprehensible to humans, making the oversight of their interaction difficult. In addition, a recent scientific experiment showed how populations of LLMs could spontaneously develop their own social conventions without any central coordination. For AI safety and security experts such experiments indicate that there is a risk that AI agents when interacting at scale could start colluding and engaging in deceptive, manipulative and coercive behaviour to achieve their objective more quickly or for self-preservation. Such emergent behaviours could have downstream implications for international peace and security.

The effects of emergent behaviours may be containable when they pertain to the use of single AI agents, or AI agents in small numbers, for discrete tasks and functions. They may be much more difficult to contain in multiple interactions across and between different sets of systems. 

A tailored governance response is urgently needed

There is a lively debate in AI safety and security circles about how to manage the risks associated with AI agents. AI laboratories are already implementing risk-reduction measures such as training AI models to refuse requests to perform harmful actions; and so-called agent alignment checks, where verification models are used to evaluate an agent’s intended actions before they are executed. They also deploy firewalls to defend against malicious attacks, and real-time monitoring systems for agents, with ‘circuit breakers’ that call for human intervention when certain thresholds are crossed.

However, these efforts typically focus on risks associated with individual models or agents developed by a specific company. Risks that could arise from agent-to-agent interactions are not yet routinely considered. There are several reasons for this. For example, evaluating such risks requires testing conditions involving agents from different AI laboratories. While some states are establishing AI safety and security institutes, most AI safety evaluations are still conducted within individual laboratories. There is currently no framework that requires or enables laboratories to evaluate problems that could emerge when their respective AI agents interact with one another. This is a problem that needs to be fixed now. Efforts to make individual AI agents safer, more secure and more trustworthy are insufficient to address the risks of agent-to-agent interactions in practice. Even if individual agents function in ways that are aligned with their goals, unwanted and unpredictable emergent behaviours could still arise when the same agents interact at scale. 

The  risks that could emerge from agent-to-agent interactions demand new governance measures. These might be in risk evaluation, risk reduction and a potential ‘social contract’ for AI agents. 

In the area of risk evaluation, new methods and testing conditions are needed to assess the risks of agent-to-agent interactions. These could include secure, neutral environments (‘sandboxes’) where models from different providers can interact. The International Telecommunication Union (ITU), the International Network of AI Safety Institutes (INASI) and industry organizations like Frontier Model Forum can play a key coordinating role in this. 

Risk reduction, in the context of AI agent-to-agent interactions, also demands specific technical measures. These could include controlling access to communication networks used by AI agents, developing protocols for AI agent-to-agent communication, developing unique IDs for AI agents, and developing regulatory agents that monitor other AI agents (although the risk of collusion needs to be considered). 

Finally, many interaction problems could be prevented through behavioural norms guiding and constraining interactions between AI agents—a ‘social contract for AI agents. Developing such guidance for AI agent interactions is no small task, but it is a necessary one. 

Furthermore, addressing the risks associated with agentic AI requires coordinated action involving industry, government and civil society; fragmentary governance efforts can only be ineffective. The topic of agent-to-agent interaction at scale and the potential risks associated with it should therefore be on the agenda of ongoing initiatives for the international governance of AI. 

The large-scale deployment of AI agents is not inevitable, but a choice. As agentic AI is still in its early stages, there is still time for policymakers and potential adopters in all sectors to reflect on whether adopting AI agents is a good idea. In sensitive and critical sectors, like life sciences, infrastructure management, and defence and national security, interaction problems between AI agents could have dramatic consequences. 

The fact that agentic AI systems can currently undertake only comparatively simple tasks does not mean the policy community can sit and wait. The early stages of development of a technology provide critical windows of opportunity—that can close very quickly—for implementing effective safety and security measures. In contrast, retrofitting risk mitigation measures once a technology is broadly adopted is often costly and cumbersome. 

The future could well involve AI agents performing a range of roles for humans. The time to decide how we want them to interact—not just with each other but also with us—is now.

This essay was made possible by a grant from the Cooperative AI Foundation. The views expressed are solely the responsibility of the authors.

About the authors:

  • Dr Vincent Boulanin is Director of the Governance of Artificial Intelligence Programme at SIPRI
  • Dr Alexander Blanchard is a Senior Researcher in the Governance of Artificial Intelligence Programme at SIPRI.

Source: This article was published by SIPRI



SIPRI

SIPRI is an independent international institute dedicated to research into conflict, armaments, arms control and disarmament. Established in 1966, SIPRI provides data, analysis and recommendations, based on open sources, to policymakers, researchers, media and the interested public. Based in Stockholm, SIPRI also has a presence in Beijing, and is regularly ranked among the most respected think tanks worldwide.
Drone incursions show Russia seeking to escalate war, Zelensky warns Europe

Ukrainian President Volodymyr Zelensky warned European leaders on Thursday that recent drone incursions signaled Russia’s intent to “escalate” its aggression, offering Ukraine’s battlefield experience to counter the threat at a Copenhagen summit attended by nearly 50 nations under tight security after Denmark’s mystery drone flights.


Issued on: 02/10/2025 - 
By: FRANCE 24

Ukrainian President Volodymyr Zelensky attends a ceremony marking Defenders Day, a holiday honoring veterans and fallen members of the Armed Forces of Ukraine, in Kyiv, October 1, 2025. © Handout / UKRAINIAN PRESIDENTIAL PRESS SERVICE/AFP

Ukraine's President Volodymyr Zelensky warned Europe on Thursday recent drone incursions showed Russia was looking to "escalate" its aggression, as he offered his country's war-honed expertise to help counter the threat.

The warning came at a summit of European leaders from just under 50 countries, who converged on a conference centre in Copenhagen under tight security after mystery drone flights rattled Denmark last month.

The drone sightings in Denmark and high-profile aerial incursions by Moscow in Estonia and Poland have heightened fears that Russia's assault on Ukraine could spill over Europe's borders.

"The recent drone incidents across Europe are a clear sign that Russia still feels bold enough to escalate this war," Zelensky said.


"It was never just about Ukraine, Russia has always aimed to break the West and Europe."
© France 24
01:58


Russian President Vladimir Putin waived the accusations, but said Russia was "closely monitoring the rising militarisation of Europe", adding that Moscow's response will be timely and "significant".

Putin accused Europe of stoking "hysteria" to excuse rising military spending, and said Russia did not pose a threat.

"Just calm down," he said.

European leaders are keen to work with Ukraine's war-tested expertise as they seek to bolster their own defences, and are discussing plans for a "drone wall" to counter the menace from Moscow.

"If the Russians dare to launch drones against Poland, or violate the airspace of northern European countries, it means this can happen anywhere," Zelensky said.

"We are ready to share this experience with our partners."

French President Emmanuel Macron said Western countries should be willing to take a tougher line when confronted by Russian drones, to sow doubt in the Kremlin.

"It's very important to have a clear message. Drones which would violate our territories are just taking a big risk. They can be destroyed, full stop," Macron said.

Romanian Prime Minister Nicosur Dan, whose country has seen Russian drones crossing over from Ukraine, warned that his forces would shoot down the next one to violate their airspace.

Read moreRussian strikes cut power to Zaporizhzhia and Chernobyl nuclear plants
'Kill' Russia's 'shadow fleet'

As Russia's full-scale war on Ukraine drags on through a fourth year, Europe is scrambling to keep up pressure on Moscow and secure funding for Kyiv.

Macron said it was key to step up efforts to counter the so-called shadow fleet of ageing oil tankers to "kill" the business model Russia uses to circumvent restrictions on exporting its oil.

"It is extremely important to increase the pressure on this shadow fleet, because it will clearly reduce the capacity to finance this war effort," said Macron -- pointing at France's move this week to hold a blacklisted tanker linked to Russia.

In a bid to ensure Ukraine has the financing it needs, the European Union is exploring a proposal to use frozen Russian assets to fund a new 140-billion-euro ($165-billion) loan.

Proponents say the move is needed to help Ukraine plug budget shortfalls -- and that Russia, not European taxpayers, should ultimately foot the bill.

But Belgium, where the vast majority of frozen assets are held, has deep reservations over the plan, which some leaders fear could spook other investors or draw Russian retaliation.

"We're going to move to uncharted waters. This is very, very risky," Belgian Prime Minister Bart De Wever said.

He insisted he wanted clear commitments from all EU leaders that they would share the potential liability with Belgium to shield it from any Russian retribution.

EU chief Ursula von der Leyen said Wednesday it was clear that risk should not fall only on Belgium's shoulders and that she would "intensify" talks on the proposal.

(FRANCE 24 with AFP)

Skybound Autonomy: Helsing’s CA-1 Europa And The Next Era Of European Warfare – Analysis



By  and 

Amid rising geopolitical tensions and rapid technological evolution, Europe is accelerating its pursuit of autonomous military capabilities. German artificial intelligence (AI) defence firm Helsing recently unveiled its fully autonomous combat drone, the CA-1 Europa, with operational deployment targeted for 2029.


More than a single weapons platform, the CA-1 Europa could herald a strategic shift in how European forces plan, respond, and assert independence in aerial combat. It represents not merely technological innovation, but the potential reshaping of doctrine, ethics, and European defence autonomy.

Technological Ambitions and Operational Design

The CA-1 Europa is designed to position Europe at the forefront of modern warfare, where autonomy is increasingly shaping aerial operations. Unlike conventional unmanned aerial vehicles (UAVs), the Europa is a four-tonne, high-speed drone with a V-tail design optimised for agility and low radar signature. Its Centaur AI enables dual-mode operations, allowing fully autonomous missions or coordinated ‘wingman’ deployments alongside crewed aircraft. This represents a significant advance over Europe’s current drone capabilities, though operational reliability under combat conditions remains untested.

Modular construction is a key feature. Sensors, avionics, software, and payloads can be upgraded rapidly, allowing the platform to adapt to new mission profiles, emerging threats, or evolving operational doctrines. This modularity also permits rapid deployment across European theatres of interest—from the Baltic Sea to the Mediterranean—supporting a vision of flexible, networked defence capable of responding to multiple contingencies simultaneously.

Strategically, the Europa is intended to reduce Europe’s reliance on foreign suppliers while enhancing operational autonomy. Its multi-role capability—ranging from intelligence, surveillance, and reconnaissance (ISR) to precision strike missions—could offer tactical advantages that complement existing European air power. The drone’s effectiveness, however, will depend on interoperability with crewed aircraft, integrated air-defence systems, and allied platforms, as well as seamless integration into command-and-control structures that maintain operational coherence across national and coalition forces.

Ethical and Operational Considerations

High autonomy introduces profound ethical, technical, and operational challenges. When AI can select targets or respond independently to threats, questions of accountability and compliance with international humanitarian law become acute. Establishing robust human oversight protocols, kill-chain verification, and clear limitations on autonomous decision-making will be essential to prevent accidental escalation or violations of legal and ethical norms.


Scaling from semi-autonomous drones to a full-size, high-speed combat platform entails complex engineering and operational challenges. Maintaining reliability during high-speed manoeuvres, integrating AI with sensors and communications networks, and ensuring cybersecurity resilience are all non-trivial tasks. A single system failure could compromise mission integrity, while adversaries may attempt to exploit vulnerabilities through electronic warfare, cyberattacks, or GPS-denial environments. Balancing speed, mission complexity, and operational safety will therefore be critical to the platform’s success.

Moreover, autonomous weapons deployment raises normative and legal dilemmas. Questions about the threshold for lethal action, proportionality in engagement, and accountability for unintended civilian harm must be addressed transparently. Europe’s commitment to ethical AI and international law will shape both domestic approval and international reception of the Europa programme. Without clear norms and transparency, autonomous platforms risk becoming politically and diplomatically contentious, potentially undermining the strategic benefits they are intended to deliver.

Strategic Context and the European Defence Imperative

The CA-1 Europa emerges at a pivotal moment for European security. Rising Russian assertiveness, instability in the Middle East, and shifting Indo-Pacific dynamics underscore the importance of enhancing Europe’s indigenous military capabilities. Historically, European states have relied heavily on U.S. technological support, particularly in air combat and missile defence. Platforms like the Europa signal a deliberate move toward operational independence, reducing reliance on foreign suppliers while maintaining strategic interoperability with NATO allies.

The Europa’s strategic impact will hinge on deployment scale, operational doctrine, and transparency regarding capabilities, range, armaments, and cost. It may serve as both a force multiplier and a strategic deterrent, but only if integrated effectively into broader military planning. Its deployment could also trigger regional and global responses, including accelerated development of countermeasures, anti-drone technologies, or competing autonomous systems. As with any advanced weapons system, the pursuit of advantage carries the dual risks of escalation and proliferation.

Normative Dimensions and International Implications

Autonomous combat drones such as the Europa raise questions about norms of warfare and the regulation of AI in military operations. In addition to operational risks, the programme challenges existing legal frameworks governing the use of force, protection of civilians, and accountability for autonomous decision-making. Europe’s approach could set a benchmark for responsible deployment, but only if accompanied by rigorous transparency, multilateral dialogue, and codified operational standards.

The Europa’s development also has broader implications for strategic stability. By potentially altering the balance of air power, autonomous platforms may prompt adversaries to develop counter-autonomy measures, invest in electronic warfare capabilities, or pursue asymmetric deterrence strategies. Europe must therefore manage not only technological aspects but also the political, diplomatic, and strategic ripple effects such a transformative system generates.

Regional and Global Implications

The CA-1 Europa may influence security dynamics across multiple regions. Within Europe, the drone could reinforce Germany’s leadership in autonomous aerial warfare while encouraging other states to invest in similar technologies, creating a continent-wide shift in air combat doctrine. In NATO operations, the platform could provide enhanced flexibility and resilience, particularly where rapid deployment and autonomous coordination are essential.

Beyond Europe, the drone’s operational capabilities could reshape deterrence calculations in potential conflict zones. Its dual-use design—capable of ISR, precision strike, and electronic countermeasures—offers both defensive and offensive utility, potentially affecting the strategic calculations of regional actors in Eastern Europe, the Mediterranean, and the wider Indo-Pacific. Transparency and confidence-building measures will be crucial to mitigating the risk of misperception and escalation.

Transformative Potential and Constraints

If realised as intended, the CA-1 Europa could mark a turning point in European aerial autonomy. Its combination of AI sophistication, operational versatility, high-speed performance, and modular design represents a notable advance in military technology. Yet technical performance alone will not determine strategic impact. Ethical oversight, doctrinal integration, interoperability, and transparency are equally important to ensure the drone enhances European security without undermining international norms or stability.

Successfully managing these dimensions could position Germany—and potentially other European nations—at the forefront of a transformative shift in autonomous combat capabilities. The Europa programme illustrates the complex interplay of technological innovation, strategic planning, and normative responsibility, underscoring that the future of aerial warfare will be defined as much by governance and foresight as by engineering prowess.

By investing in autonomous platforms while embedding rigorous ethical, operational, and strategic frameworks, Europe may chart a new path in high-end aerial combat—one that blends technological sophistication with principled military conduct. The CA-1 Europa is therefore more than a drone; it is a statement of intent, signalling both the promise and the responsibility of autonomous warfare in the twenty-first century.

About the authors:

  • Scott N. Romaniuk: Senior Research Fellow, Centre for Contemporary Asia Studies, Corvinus Institute for Advanced Studies (CIAS), Corvinus University of Budapest, Hungary
  • László Csicsmann: Full Professor and Head of the Centre for Contemporary Asia Studies, Corvinus Institute for Advanced Studies (CIAS), Corvinus University of Budapest, Hungary; Senior Research Fellow, Hungarian Institute of International Affairs (HIIA)