We’ve all seen the films. The grizzled agent in a dimly lit room, staring at a screen, yelling, “Enhance! Enhance!” as a blurry image magically resolves into a crystal-clear face. It used to be science fiction. Now, it’s a £177,000 ($225,000) line item in a government contract, and the database isn’t a curated collection of mugshots; it’s your holiday photos, your LinkedIn profile, and that embarrassing picture from a decade ago you thought was long forgotten.
The US Customs and Border Protection (CBP) has just signed a one-year deal with Clearview AI, a company whose entire business model would have been considered dystopian fantasy not long ago. This move slams the accelerator on a debate we’ve been lazily circling for years: the thorny world of AI surveillance ethics. This isn’t just another tech procurement; it’s a fundamental test of where we draw the line between security and an Orwellian reality.
The Algorithm Never Blinks: AI Meets Surveillance
What exactly is AI surveillance? It’s not just about putting up more cameras. It’s about teaching machines to watch, understand, and identify what they see at a scale no human team could ever manage. This technology powers everything from your phone’s face unlock to sophisticated security systems.
At the heart of this revolution is biometric identification. This is the science of using your unique physical characteristics—your face, fingerprints, even the way you walk—as a form of ID. It’s the ultimate password, one you can never change. And companies like Clearview AI have become the self-appointed librarians of this data, hoovering up what it claims are over 60 billion images from the public internet to build a colossal, searchable face-database.
A Clearview into the CBP’s New Playbook
So, what does the CBP’s new contract actually entail? According to a report by Wired, the deal gives Border Patrol’s intelligence units and the National Targeting Center access to this immense facial recognition library. The official purpose is for tactical targeting and “strategic counter-network analysis.” Let’s be clear about what that means: it’s about identifying people of interest to disrupt security threats.
This isn’t the CBP’s first foray into high-tech tools, but this partnership represents a significant escalation. The agency is moving beyond traditional law enforcement databases and tapping into a vast, unregulated ocean of public data. Think of it like this: a standard police database is a carefully curated photo album of known individuals. Clearview AI’s database is the entire internet, scraped and indexed by face. The CBP technology is now plugged directly into that index.
The deal’s terms are shrouded in nondisclosure agreements, leaving critical questions unanswered. Will US citizens be fair game in these searches? How long will data on individuals be retained? The lack of transparency here isn’t a bug; it feels like a feature.
Your Face, Their Database: The Looming Privacy Concerns
This brings us to the core of the issue: the staggering privacy concerns. You never consented to have your face included in a law enforcement search engine. You uploaded a photo to a social network, not a digital line-up. Civil liberties groups are, rightly, sounding the alarm. This practice obliterates any meaningful notion of consent and personal privacy.
The ethical dilemmas are profound.
– Lack of Consent: The entire system is built on images taken without permission for this purpose.
– Potential for Misuse: A tool designed for “tactical targeting” at the border could easily be repurposed for domestic surveillance. Where does it stop?
– The Chilling Effect: If you know any photo of you could end up in a government database, does that change how you behave online? Or even in public?
This isn’t some slippery slope fallacy; it’s the logical progression of unchecked technological deployment. We are normalising a level of surveillance that would have been unthinkable just a few years ago.
The Lawmakers Wake Up
Thankfully, this isn’t going unnoticed in the halls of power. Some lawmakers are pushing back hard against this techno-authoritarian creep. Senator Ed Markey, for example, has been a vocal opponent, introducing legislation to ban federal agencies like CBP and ICE from using facial recognition entirely.
Markey’s argument is simple: the technology is invasive, often inaccurate, and operates without a clear legal framework or public consent. His efforts signal a growing unease among policymakers who are starting to realise that we can’t simply trust agencies to self-regulate when it comes to tools this powerful. The legislative landscape is lagging far behind the technology, creating a dangerous grey area where AI surveillance ethics are an afterthought.
The Machine Gets It Wrong
Perhaps the most compelling argument against the widespread use of this technology has nothing to do with ethics and everything to do with competence. The simple fact is, these systems are deeply flawed.
The US National Institute of Standards and Technology (NIST), the gold standard for this kind of testing, has found that facial recognition systems can have alarmingly high error rates. As Wired noted in their coverage, in non-ideal conditions—like poor lighting, an angled face, or a low-quality image—error rates “often in excess of 20 percent” are common.
More worryingly, when a system is configured to always return a potential match (even if the person isn’t in the database), it can be “100 percent wrong.” It will simply serve up the ‘best’ available, but incorrect, match. This isn’t a minor bug; it’s a catastrophic flaw when someone’s liberty is at stake. The consequences of a false match aren’t just an administrative headache; they can lead to wrongful arrests and destroyed lives. We’re effectively outsourcing critical decisions to an algorithm that’s only right some of the time.
Where Do We Go From Here?
The CBP’s contract with Clearview AI is a watershed moment. It forces a conversation we desperately need to have about the kind of society we want to live in. Do we want one where our every online presence is catalogued for government scrutiny, using biometric identification tools that are both ethically dubious and technically unreliable?
The lure of security is powerful, but it cannot come at the cost of our fundamental rights to privacy and due process. This isn’t about being anti-technology; it’s about being pro-foresight. We need to demand transparency, establish clear legal guardrails, and have an honest public debate about the implications of these tools before they become irrevocably embedded in our society.
This is more than just another government contract. It’s a choice about our future. So, the question I leave you with is this: who should have the right to your face? And what are you willing to do to protect it?


