The core of the issue lies in the dangerous paradox of AI in security. On one hand, the need for it has never been clearer. A report from Red Canary highlights that a whopping 85% of security leaders fear their teams will be completely overwhelmed by the sheer volume of threats without AI’s help. The modern Security Operations Centre (SOC) is a deluge of alerts, a digital tsunami of data points. Humans simply cannot keep up. SOC automation, powered by AI, is the only way to sift through this noise, separating the genuine threats from the endless false positives. AI models can analyse billions of data points in real-time, detecting patterns of malicious behaviour that would be invisible to the human eye. This is the promise, the glossy brochure version of AI cybersecurity that has fuelled a market projected to rocket from $23.12 billion this year to over $136 billion by 2032, according to an analysis by ResearchAndMarkets.com. It’s an intoxicating narrative of progress and control.
The Monster in the Machine
But here’s the terrifying twist in the tale. The same intelligence that makes AI such a powerful defender also makes it a uniquely vulnerable and dangerous target. We’re not just dealing with software bugs; we’re dealing with something that can be taught to lie. The biggest AI security risks aren’t about crashing the system, but about subtly corrupting its ‘mind’. This leads us to the shadowy world of adversarial attacks, a concept that should be keeping every CISO awake at night. Think of it like this: You’ve trained a sophisticated guard dog (your AI) to recognise burglars by their scent. An adversarial attack is when a clever intruder crafts a special perfume—a single, invisible molecule—that makes them smell exactly like the dog’s owner. The AI, doing precisely what it was trained to do, wags its tail and lets them walk right through the front door with the family silver.
These attacks involve feeding the AI model data that is maliciously designed to be misclassified. A security AI trained to spot malware might be presented with a virus that has been tweaked by just a few pixels or lines of code, making it appear entirely benign to the algorithm, even as it wreaks havoc on the network. Even more sinister is the threat of model poisoning. This is a slower, more insidious form of corruption. Here, the attacker doesn’t just trick the AI at the point of entry; they corrupt its education from the very beginning. By secretly feeding tainted data into the AI’s training set, they can create hidden backdoors in its logic. The AI still functions perfectly 99.9% of the time, but it has been trained to ignore a specific type of attack—the attacker’s signature—or to create a blind spot that the criminals can exploit at will. It’s the ultimate Trojan horse, a weapon hiding inside your last line of defence. The stats are already bearing this out: a sobering 43% of organisations have already experienced a cyber incident directly related to AI.
The Ever-Expanding Battlefield
This internalised threat is made infinitely worse by an external reality: the corporate attack surface is growing at a terrifying rate. According to the Red Canary report, organisations are seeing their attack surfaces—the sum of all their digital entry points, from cloud servers and IoT devices to employee laptops and mobile apps—grow by an astonishing 41% every single year. Trying to defend this sprawling, ever-changing digital estate with human teams alone is like trying to guard a city with no walls using only a handful of watchmen. It’s a fool’s errand.
This explosion in complexity directly fuels higher breach costs, which now average a gut-wrenching $3.7 million per incident. Every new device, every new cloud application, is another potential door or window for an attacker to pry open. We’ve built sprawling glass houses and are only now realising we’re living in a neighbourhood full of expert stone-throwers. This chaotic environment is precisely where AI-driven attacks thrive. Attackers are using AI to automate the process of scanning these vast attack surfaces, identifying vulnerabilities far faster than any human security team could ever patch them.
The End of Malware as We Know It
Perhaps the most profound strategic shift we’re witnessing is the move away from traditional malware towards identity-based attacks. For decades, the primary threat model was a malicious piece of software—a virus, a worm, ransomware. Our defences were built around detecting these foreign invaders. But attackers have gotten smarter. Why bother building a sophisticated battering ram when you can just steal the keys to the gate? Today, stolen credentials are the weapon of choice. The report from Cybersecurity Dive confirms that in an incredible 80% of intrusions, attackers gain their initial foothold using legitimate, stolen credentials.
This changes everything. When an attacker is using valid login details, they don’t look like an intruder. They look like an employee. Traditional signature-based defences are useless against this threat. This is where the AI security risks become acutely painful. Adversaries are now using AI to craft hyper-realistic phishing emails at a scale and quality never seen before. They use generative AI to mimic the writing style of a CEO or finance director, creating urgent requests for wire transfers or credentials that are almost impossible to distinguish from the real thing. They can even use deepfake audio to leave a convincing voicemail, adding another layer of authenticity. Our innate human ability to trust and recognise patterns is being turned against us by machines.
The Human Element: Our Weakest Link and Last Hope
So, what is the answer? Throwing more technology at the problem seems to be the default response, but it’s a deeply flawed strategy. The industry is facing a critical skills shortage in exactly the areas needed to combat these new threats. A massive 75% of organisations report a skills gap in intrusion detection, and 72% are short on incident response talent. We are trying to fight a war of intelligence without enough intelligence officers. This is where the argument for AI as a “force multiplier” comes in, a term used by Red Canary co-founder Brian Beyer. The idea is that AI shouldn’t replace human analysts but augment them, handling the tedious, high-volume work so that the human experts can focus on the complex, creative and novel threats that AI can’t handle.
This sounds sensible, but it conveniently ignores a crucial point: 75% of security executives also worry that over-reliance on AI will erode the critical human problem-solving skills their teams need. If your junior analysts spend all day just clicking “approve” on an AI’s recommendation, how will they ever develop the intuition and deep expertise needed to spot a truly novel attack, like a sophisticated adversarial attack or one of the inevitable zero-day exploits that have no known signature? We are at risk of deskilling the next generation of cybersecurity defenders, leaving them utterly dependent on a technology that we already know can be turned against them.
A Future Bought on Credit
Looking ahead, the picture gets even murkier. We’re pouring billions into a market that is fundamentally a double-edged sword. The same algorithms and hardware powering our defences are available to our adversaries. The innovation cycle is relentless. As we develop AI to detect deepfakes, attackers develop AI to create even better ones. This is a perpetual arms race, and we are not guaranteed to win. Geopolitical tensions add another layer of risk. Recent U.S. tariffs on AI hardware and semiconductors, as noted by Yahoo Finance, are already reshaping global supply chains. This could lead to a world where not all AI is created equal, with certain nations having access to more secure or more powerful hardware, creating a dangerous imbalance.
How can businesses prepare? The answer isn’t to abandon AI. The reality is, we can’t. The 85% of leaders who say they’d be overwhelmed without it are correct. The path forward requires a level of strategic paranoia that most organisations are unprepared for. It means treating your AI systems not as infallible black boxes, but as your most privileged and most vulnerable new employees. It requires continuous red-teaming, where you actively hire experts to try and trick your own AI. It demands a renewed focus on fundamentals: robust identity and access management, a “zero trust” architecture where no user or device is trusted by default, and constant training for your people.
Walking the Tightrope
We are standing at a precarious moment in the history of cybersecurity. We have embraced a technology that offers immense power, but we have been wilfully blind to its inherent fragility. The fact that nearly four out of five security leaders are afraid of their own tools is not just a data point; it’s a desperate cry for help. We’ve built our modern castles on foundations of intelligent, learning sand, and the tide of AI-driven attacks is rising fast. Believing that AI will solve all our security problems is a dangerous fantasy. The reality is that we’ve invited an incredibly powerful, intelligent, and unpredictable new entity into our most sensitive digital spaces. We need to stop treating it like a magical shield and start treating it like what it is: a weapon that can, and will, be used against us.
The question is no longer if our AI defences will be turned against us, but when—and whether we will have the human ingenuity left to fight back when they are.
What is your organisation doing to validate the trustworthiness of its AI security tools? Are we sleepwalking into a disaster of our own making?


