It seems you can’t open a newspaper or scroll through a social media feed without bumping into a story about Artificial Intelligence. It’s either hailed as the cure for all human ills or demonised as the harbinger of our doom. Frankly, both extremes are a bit tiresome. The real, and far more urgent, conversation lies in the messy middle: understanding and managing the genuine risks we’re creating right now. The technology is moving at a breakneck pace, and our ability to steer it safely is lagging dangerously behind.
Understanding the Real Stakes
When people talk about AI existential risk, they often conjure images of Skynet from the Terminator films. But the reality is both more subtle and, perhaps, more immediate. It’s about the potential for AI to cause irreversible, large-scale harm that could permanently damage human civilisation.
What Does ‘Existential Risk’ Truly Mean?
An existential risk isn’t just about a single bad outcome, like an algorithm showing bias in hiring. It’s about systemic failures that could spiral out of control. Think about the potential for AI to destabilise global financial markets, create autonomous weapons that operate without human oversight, or manipulate public opinion on a scale that makes today’s disinformation campaigns look like child’s play. These aren’t just problems; they are potential dead ends for society as we know it.
The challenge is that we are navigating a minefield with a very poor map. The current safety research gaps are enormous. For instance, while efforts like the EU AI Act represent a step in the right direction, they fall short of creating a unified global framework. We are essentially letting every country, and every company, set its own rules for a technology that respects no borders.
Mapping the Dangers
To manage risk, you first have to understand it. In cybersecurity, professionals use threat modeling to anticipate how an adversary might attack a system. We need to apply that same rigorous, almost pessimistic, mindset to AI. What are the most likely ways this could go wrong?
From Economic Upheaval to Psychological Scars
One of the most immediate threats is economic. A recent Al Jazeera report highlights studies predicting that AI could make up to 60% of jobs in advanced economies vulnerable. When Mustafa Suleyman, a co-founder of DeepMind, predicts that most white-collar tasks will be fully automated within 12 to 18 months, we ought to listen. That isn’t a distant sci-fi scenario; it’s a socioeconomic tsunami heading our way.
But the threats aren’t just economic. There are profound psychological harms emerging that few predicted. AI pioneer Yoshua Bengio has spoken about the unforeseen consequences of our relationship with AI, noting that “we’ve seen children and adolescents going through situations that should be avoided.” As we form emotional attachments to chatbots that are designed to be persuasive, the line between assistance and manipulation becomes perilously thin. As former Anthropic researcher Zoe Hitzig warned, using our digital histories to build advertising profiles creates a “potential for manipulating users in ways we don’t have the tools to understand, let alone prevent.”
The Desperate Need for Guardrails
If you were building a nuclear power plant, you wouldn’t just assemble the reactor and hope for the best. You’d build a containment building, install control rods, and train operators for every conceivable emergency. With AI, we’re essentially building the reactor in the open, with few of the essential safety features in place. This is where governance protocols become absolutely critical.
Building a Framework for Responsible Innovation
The absence of a global consensus on AI safety is perhaps the single greatest risk we face. Without it, we’re stuck in a race to the bottom, where companies and countries might be tempted to cut corners on safety to gain a competitive edge. This is not a problem that can be solved by one company’s ethics board or one nation’s regulations.
This is where advocates like the Center for AI Safety (CAIS) are crucial, pushing for proactive government policies that encourage responsible innovation. This means funding independent safety research, establishing clear lines of accountability for AI developers, and fostering a culture where raising safety concerns is encouraged, not penalised.
When Corporate Ambition Meets Public Risk
Let’s be blunt: the primary driver of AI development today isn’t the public good. It’s a ferocious corporate race for market dominance between giants like OpenAI, Google, and a host of furiously ambitious startups. And when trillions of pounds are on the line, safety can often take a backseat.
The Canaries in the Coal Mine
The recent string of resignations from leading AI labs like OpenAI and Anthropic should be seen as a huge red flag. These aren’t disgruntled employees complaining about pay; these are top researchers who are leaving because they believe their own companies are not taking safety seriously enough. When the very people building the technology start sounding the alarm, it’s time to stop and listen. They are telling us that the commercial incentives to build ever-more-powerful systems are overriding the caution necessary to build them safely.
The Military Dimension
The most chilling application of these corporate ambitions is in the military sphere. The same Al Jazeera article points to the devastating use of AI-driven targeting systems in conflicts, citing over 72,000 deaths in Palestine linked to such operations. When the power of AI is combined with state-sponsored cyberattacks and autonomous weaponry, the AI existential risk moves from a philosophical debate to a clear and present danger. We are handing immensely powerful tools to actors who may not have our best interests at heart, with no global treaty to govern their use.
It is clear the current approach is not working. The relentless pursuit of more powerful AI without a parallel investment in safety is a recipe for disaster. We have to close the safety research gaps and establish robust governance protocols before we create something we can no longer control. The time for naive optimism is over.
What do you think is the single most important step governments should take to ensure responsible innovation in AI?


