We seem to be endlessly fascinated by what consumer AI can do – write a poem, create a bizarre image of a prawn that looks like a king, or summarise a meeting you couldn’t be bothered to attend. That’s all well and good, but while we’re distracted by the shiny toys, a far more serious conversation is unfolding in the corridors of power at places like NATO. The real story isn’t about AI replacing your intern; it’s about how it’s fundamentally reshaping national and global security.
The truth is, we are building and deploying technologies with profound security implications faster than we can understand them, let alone regulate them. This isn’t just a tech issue; it’s a geopolitical one, and the questions being asked are fundamental to how we navigate the coming decades.
What Are We Even Talking About?
When we discuss AI cybersecurity challenges, it’s a bit of a two-sided coin. On one side, you have AI as the heroic defender – a tireless digital sentinel sifting through billions of data points to spot a cyberattack before a human analyst has even had their first cup of tea. It can detect anomalies, predict threats, and automate defence mechanisms at machine speed. Many security firms are, quite rightly, betting the farm on this capability.
But flip that coin over, and the picture is much darker. The same AI that powers our defences can be turned into a formidable weapon. Imagine an AI designed not to defend a network, but to attack it. It could learn a system’s vulnerabilities, craft custom malware on the fly, and launch attacks that are so sophisticated and fast that human-led defence teams are simply overwhelmed. This is the dual-use dilemma at the heart of the debate, and it’s a wicked problem.
The Canyons in Our Rulebook
This is where we run into glaring security policy gaps. Our international laws, security doctrines, and treaties were largely written in an analogue or early-digital era. They were designed for a world of tanks, ships, and human soldiers. Trying to apply these rules to autonomous AI agents conducting cyber warfare is like trying to enforce motorway speed limits on drivers in the sky. The frameworks just don’t match the reality.
An recent opinion piece in the Kuwait Times quite rightly pointed out that organisations like NATO are grappling with this very issue. How do you define an “armed attack” when it’s executed by lines of code? How does the principle of proportionality apply when an AI can escalate a conflict in milliseconds, far faster than any human chain of command could react? These are not theoretical questions. The absence of clear answers creates a dangerous ambiguity that malicious actors, whether states or otherwise, are more than happy to exploit. Our policies have holes in them, and those holes are getting bigger with every new AI model that gets released.
The Troubling Ethics of a Robot Army
This leads us directly into the thorny subject of military AI ethics. When military organisations develop AI for offensive cyber operations, it sets a precedent. The technology and techniques pioneered in the classified world of national security have a well-established habit of trickling down into the civilian sphere. The ethical frameworks—or lack thereof—governing an AI used to disable an enemy’s power grid will eventually influence the tools used by sophisticated criminals to attack a bank.
The core ethical questions are unsettling:
– Accountability: If an autonomous AI system makes a mistake and targets a hospital instead of a military command centre, who is responsible? The programmer? The commanding officer who deployed it? The nation-state? Good luck getting a straight answer.
– Control: How can we be certain that these complex systems will always behave as intended? The nature of machine learning means that AI models can evolve in ways their creators didn’t predict. The risk of unintended escalation is very real.
– Bias: AI models are trained on data, and that data can reflect human biases. A military AI trained on flawed intelligence could make catastrophic decisions, perpetuating and even amplifying historical prejudices in a digital battlespace.
This isn’t just a military problem. The ethical shortcuts taken in the name of national security today could easily become the dangerous vulnerabilities in our civil infrastructure tomorrow.
A New Breed of Global Threats
The rise of AI-powered tools is pouring fuel on the fire of global cyber threats. State-sponsored hacking groups and organised cybercriminals now have access to capabilities that were once theoretical. We are seeing:
– AI-Powered Phishing: Forget poorly-worded emails from a long-lost prince. We’re talking about hyper-personalised, context-aware messages crafted by AI that are almost indistinguishable from a real colleague or friend.
– Adaptive Malware: Viruses that can analyse the network they’re in, identify the security software being used, and rewrite their own code to evade detection.
– Automated Reconnaissance: AI agents that can scan the entire internet for unpatched servers or vulnerable devices, compiling lists of targets for hackers at a scale no human team could ever match.
As the Kuwait Times article also notes, this makes attribution incredibly difficult. If an attack is launched by a swarm of decentralised AI agents, tracing it back to its source becomes a forensic nightmare. This ambiguity is a massive strategic advantage for attackers, allowing them to sow chaos with a degree of plausible deniability.
So, What’s the Plan?
Standing still is not an option. Ignoring these challenges is an invitation for disaster. The path forward requires a multi-pronged strategy, moving beyond talk and towards concrete action.
Building a Modern Rulebook
First, we must urgently address the security policy gaps. This means convening international bodies—not just military alliances like NATO, but organisations like the UN—to begin hammering out new treaties for the digital age. We need to define red lines, establish norms for state behaviour in cyberspace, and create mechanisms for de-escalation that work at machine speed. It will be politically messy and painfully slow, but it’s essential work.
Ethics Must Be Coded In, Not Bolted On
Second, we need to champion ethical AI development from the ground up. This means pushing for “Ethics by Design” frameworks, where accountability, transparency, and safety are core components of any AI system, not afterthoughts. The discussion around military AI ethics needs to be brought into the open, with clear and enforceable doctrines that put meaningful human control at the centre of all lethal and critical systems.
A Neighbourhood Watch for the Internet
Finally, collaboration is key. In a world where global cyber threats can originate anywhere and strike everywhere, isolationism is a losing strategy. Nations, private tech companies, and security researchers must build stronger alliances for information sharing and collective defence. When one organisation discovers a new AI-powered attack method, that intelligence needs to be shared rapidly across the network to inoculate everyone else.
We are at a genuine inflection point. The race to develop and deploy AI is also an arms race, whether we choose to call it that or not. We’ve built a machine that is learning exponentially, and we’re now struggling to write the instruction manual for it as it’s running.
The question isn’t whether AI will be part of our security future; it already is. The real question is, are we prepared to make the difficult policy and ethical choices needed to ensure that future is one we can actually live with? What do you think is the single biggest risk we’re currently ignoring?


