The Immediate Danger: Alex Stamos on AI’s Real Threat to Cybersecurity

1 week ago
12 min read

The Immediate Danger: Alex Stamos on AI’s Real Threat to Cybersecurity

In an era increasingly defined by rapid technological advancement, the discourse surrounding Artificial Intelligence often bifurcates into two starkly different narratives: one of existential, apocalyptic threats posed by superintelligent machines, and another of more immediate, human-driven risks. Alex Stamos, a preeminent figure in computer security and former head of security at Facebook during the tumultuous 2016 Russian hacking incidents, firmly aligns himself with the latter. In a recent interview on “Frankly Fukuyama,” Stamos articulated a nuanced yet urgent perspective: the true threat of AI lies not in sentient machines taking over, but in malicious human actors leveraging AI to amplify their capabilities in cyber warfare.

Stamos, currently a leading voice at Stanford’s Cyber Policy Center and instructor of the popular “Hack Lab” course, dismisses the “AI doomer” narrative as largely unfounded, pointing to a growing skepticism among experts regarding the near-term feasibility of Artificial General Intelligence (AGI). Instead, he warns of a perilous “race” where cyber attackers are already automating their “kill chains” at an alarming pace, forcing defenders into a complex and constrained battle for automation.

Dispelling the AGI ‘Doomer’ Prophecy: A Matter of Understanding

A significant portion of the public debate around AI revolves around the concept of AGI – artificial general intelligence – a hypothetical AI that can understand, learn, and apply intelligence to any intellectual task that a human being can. Proponents of the “existential threat” theory, often citing figures like Geoffrey Hinton and Elon Musk, envision a future where AGI surpasses human intelligence, potentially leading to unforeseen and uncontrollable consequences for humanity. Stamos, however, offers a grounded counter-argument.

“I’m much more of a short-term worried about humans person. I’m not an existential threat guy,” Stamos states, reflecting a sentiment increasingly shared within expert circles. He notes that over the past year, skepticism towards AGI has grown, with many experts pushing its timeline “post-2030, even further out.” Crucially, Stamos argues that current Large Language Models (LLMs), the bedrock of today’s AI revolution, are fundamentally limited in their capacity to achieve AGI.

The core of this limitation, Stamos explains, lies in LLMs’ lack of “true understanding” of the physical world. “LLMs don’t have true understanding of the world in a way that human beings do,” he emphasizes. He illustrates this with a compelling analogy: an LLM “knows that an apple fell on Newton’s head, but it doesn’t actually understand gravity.” This distinction is critical. LLMs excel at language – one of the most complex cognitive functions – but struggle with basic physical interaction. “LM’s can write poetry, but pretty much it’s impossible to train an LM if you gave it control of a robot to get your Frisbee,” Stamos quips, referencing his golden doodle’s ability to catch a Frisbee despite its inability to write verse.

This “novelist’s view of the world,” as Stamos terms it, means LLMs process information primarily through the written word, lacking an innate, 3D comprehension of space and physics that even animals possess. Achieving AGI, he suggests, would likely require “completely different types of models” – physical, physics-based, and vision models – or a complex integration of various models, akin to the multi-faceted nature of the human brain.

Beyond cognitive limitations, Stamos highlights a fundamental absence in AI that precludes existential threat: “AI doesn’t want anything.” Unlike living organisms, AI models are “a box of numbers,” devoid of “wants, instincts, or mammalian desires.” They don’t seek to reproduce, kill, or self-perpetuate. Any “bad” action attributed to AI, Stamos asserts, is “because somebody made it to want to do it.” This underscores the central role of human intention and agency, even in the most advanced AI scenarios.

Simulated Creativity vs. Genuine Innovation: The Security Implications

Another crucial distinction Stamos draws is between AI’s ability to “simulate creativity” and genuine human innovation. While LLMs can generate novel text or code, their output is fundamentally derived from patterns in their training data. This limitation has significant implications for cybersecurity.

Stamos warns security professionals that while AI can be immensely useful for defensive purposes – automating routine tasks, identifying known threats, and processing vast amounts of data – it cannot “foresee new attacks.” An AI system, trained on past attack patterns, can emulate and defend against known threats. However, it “is not going to think, ‘oh well, what might a human do in the future? What creative things could happen?’”

This vulnerability has already manifested in the “trust and safety world,” where companies over-reliant on AI for content moderation have been “outsmarted by human bad guys who end up just changing their approach a little bit and then getting around the AI.” AI-driven protections, while “quick, very broad, very cheap” and operating “at scale,” are also “very brittle” because “it only takes a little bit of creativity to get around them.” This implies that the human element, both in offense and defense, remains indispensable for true adaptability and innovation in the cyber realm.

The Immediate Threat: AI-Powered Cyber Kill Chains

While the threat of AGI remains distant, Stamos emphasizes that the “real problem is bad people using these technologies for bad purposes.” He reveals that “over the last six months, we’ve seen a real increase in the use of AI for offensive purposes.” This isn’t theoretical; it’s an observable, escalating trend in the cyber landscape.

At the heart of this escalation is the automation of the “cyber kill chain,” a concept borrowed from the military, outlining the sequential steps an attacker takes to achieve their objective within a target network. These steps typically include:

  • Reconnaissance: Gathering information about the target.
  • Weaponization: Pairing an exploit with a backdoor or payload.
  • Delivery: Transmitting the weapon to the target.
  • Exploitation: Triggering the weapon’s code to run on the target system.
  • Installation: Installing a backdoor or persistent access.
  • Command and Control (C2): Establishing remote control over the compromised system.
  • Actions on Objective: Achieving the attacker’s goal (e.g., data theft, system disruption, ransomware).

Attackers are “systematically figured out how to use AI in more and more parts of the kill chain,” Stamos warns. This automation offers several critical advantages to malicious actors:

  1. Force Multiplier: AI allows attackers to “trade human beings for computers,” meaning fewer skilled human hackers can achieve more. This is invaluable for both state-sponsored actors and financially motivated groups, who consistently face a shortage of talent.
  2. Parallel Operations: A smaller human team can “supervise a bunch of AI agents doing their work at the same time,” dramatically increasing the scale and breadth of attacks.
  3. Speed Kills: “If an attacker is really fast, that’s a real problem for you,” Stamos quotes Rob Joyce of the NSA. Defenders rely on “tripwires” – alerts at each stage of the kill chain – to detect and respond to attacks. However, if AI automates the entire kill chain, “the whole thing could be over” in the 15 minutes it takes a human defender to wake up, log in, assess an alert, and initiate a response. This speed advantage shifts the balance of power significantly towards the aggressor.

The Defender’s Dilemma: Automation Under Constraint

The stark reality presented by AI-driven offensive capabilities necessitates an equally robust defensive response. “Attackers are now automating all the parts of the kill chain. And so defenders have to do that too,” Stamos asserts. However, defensive automation faces unique and complex hurdles that attackers often do not.

“The problem for defenders is we have bosses. We have Sarbanes-Oxley letters. We work for corporations that have to live up to rules. We have auditors,” Stamos explains. Regulatory compliance and corporate governance create a cautious environment, making it “much more dangerous for us to do things like hook up all the parts of our network to an AI system that can just shut parts of it down at any point if it feels bad.”

Stamos employs a vivid metaphor from the film “Bridge over the River Kwai,” where the protagonists rig a beautifully constructed bridge with explosives for its eventual destruction. “That’s what it’s like to be a security person,” he says. “You have this beautiful infrastructure and you have to rig it with explosives because in the end to stop an attack you usually have to break the infrastructure in some way.” This could mean cutting off firewalls, dropping internet transit, or shutting down servers – actions that, in extreme cases, involve telling a CEO “it’s time to shut the company down.”

Delegating such “bridge-blowing” authority to an autonomous AI agent is a profound challenge. While an AI might not be entrusted with turning off all internet access, it may need to suspend accounts, shut down individual virtual machines, or create firewall rules. These actions, even on a smaller scale, carry “extreme risk in a corporate environment” and “production environment.” The “big race right now” is for defenders to implement “automation within constraints” – systems that are effective yet responsible, capable of swift action without causing catastrophic collateral damage or violating compliance mandates.

The Agentic AI Problem: Unconstrained Delegation and Liability

Beyond the immediate cyber battlefield, Stamos delves into a broader, systemic issue that AI exacerbates: the “delegation problem” with “agentic AI.” Agentic AI refers to AI systems designed to operate autonomously to achieve goals, often by interacting with other systems and making decisions. The fundamental challenge, Stamos notes, is that “we just don’t have good ways of delegating authority to agents overall.”

There are currently no robust, standardized mechanisms to grant AI agents “temporary constrained delegation.” Stamos gives a relatable example: “If you want to tell an agent go, you’re allowed to be me for the purposes of shopping on Amazon for, and you can buy up to, you know, I want you to buy a gift for my wife and it can cost up to $50. Go buy something nice for her.” Such a nuanced instruction, with built-in constraints, is currently impossible to implement securely. An agent given your credit card “can go buck wild.” The same applies to corporate networks; an AI agent given email access could “write an email to tell your boss to go f himself.”

The absence of “semantics” for constrained delegation, coupled with a lack of standardized “sandboxes” for AI agents, leaves a critical vulnerability. While some proprietary solutions are emerging, an open, interoperable framework is desperately needed. This technical challenge has profound legal implications, as Stanford Law School colleague Mark Lemley is actively exploring.

Stamos humorously (and grimly) predicts a boom for “older lawyers like Mark,” as “the legal issues here are huge.” AI agents acting autonomously, signing contracts, and engaging in unauthorized actions will undoubtedly lead to a “multiple level appellate process.” The DocuSign analogy – where the ease of digital signatures raises questions of true authorization – foreshadows a future where AI-signed agreements will present even greater legal complexities.

The Geopolitical Dimension: China’s AI Cyber Prowess

The race for AI dominance and its application in cyber warfare takes on a significant geopolitical dimension. Stamos highlights China as the “first” and most concerning adversary in this space. He cites a recent report from Anthropic, an AI safety-focused company, which “threw like a huge grenade into this whole world” by claiming to have caught a Chinese intelligence group using their systems to automate parts of the cyber kill chain.

While acknowledging some controversy and a desire for more technical details from Anthropic, Stamos trusts the report’s veracity. He points to existing open-source tools from Chinese groups that demonstrate similar capabilities. One such tool, utilizing DeepSeek (an open-source Chinese LLM), is trained on Kali Linux – a popular penetration testing distribution – to “automatically do a bunch of hacking for you.” This tool allows users to “ask it, ‘hey, go hack this network for me,’ and it will go do a bunch of the hacking automatically for you.”

The proliferation of sophisticated open-source AI models from Chinese labs like DeepSeek and QM is a critical factor. These models, while perhaps “70-80% as good” as closed models from OpenAI or Anthropic, offer a crucial advantage to malicious actors: anonymity and customizability. “If you do that, you are not leaving logs at Anthropic or OpenAI for the FBI to get, for the NSA to get,” Stamos explains. Furthermore, unlike commercial models that have built-in protections against misuse, open-source models can be “intentionally train[ed] … to do hacking.”

Stamos dismisses the notion that the US is “way ahead of China in AI” as “foolish.” China has pursued a “plan for decades to catch up to the United States in fundamental sciences, including in computer science.” This strategy, involving sending students to top US institutions like Stanford, has been effective in building their domestic AI capabilities. The academic nature of much AI research, with discoveries openly published until the commercialization of GPT-3, further facilitated knowledge transfer.

Crucially, Stamos emphasizes that high-level AI capabilities are “not at all” limited to state-level actors. Unlike nuclear technology, where control over fissile material was key, “controlling the compute is effectively impossible.” He notes that US export controls on GPUs are partially circumvented by countries like Singapore and the UAE, which import Nvidia GPUs “because they are effectively exporting that compute to China.” The final AI model, a “single thumb drive” of data, can be “zipped … in seconds over fiber optic cables.”

This accessibility means that “if this is something that individuals can really master, it seems to me the threat level then goes way up because you’re living in a world where there are a lot of people with bad intentions.” The democratization of advanced AI tools lowers the barrier to entry for sophisticated cyberattacks, empowering a broader range of malicious actors beyond state intelligence services or large criminal organizations.

Conclusion: A Call for Defensive Innovation and Responsible Governance

Alex Stamos’s insights paint a clear picture: the immediate, tangible threat of AI is not a sci-fi dystopia, but a rapidly evolving cyber battleground where human ingenuity, amplified by AI, is being weaponized. The “doomer” narrative distracts from the pressing need to address the very real, short-term challenges posed by AI in the hands of malicious actors.

The “race” for automation between cyber attackers and defenders is already underway, with attackers gaining a significant speed and scale advantage. For defenders, the challenge is not merely technical, but also organizational, regulatory, and ethical. Developing AI-driven defensive systems that are both effective and accountable, capable of swift, surgical action without jeopardizing critical infrastructure or violating compliance, is paramount. Simultaneously, the fundamental problem of constrained delegation for AI agents requires urgent attention from technologists, policymakers, and legal experts to prevent widespread misuse and liability issues.

The geopolitical landscape further complicates matters, with nations like China rapidly developing and deploying sophisticated AI tools, some of which are openly accessible and easily repurposed for offensive cyber operations. The democratization of AI capabilities means that the threat is no longer confined to state-level actors, but extends to a wider array of individuals and groups with harmful intentions.

Ultimately, Stamos’s analysis is a powerful call to action: rather than fearing a hypothetical AGI takeover, society must focus on building resilient defenses, establishing robust governance frameworks for AI delegation, and fostering international cooperation to mitigate the very real and present danger of AI-enabled cyber warfare. The future of cybersecurity hinges not on stopping AI itself, but on wisely managing the human element that wields its formidable power.


Source: The real threat posed by AI: an interview with Alex Stamos (YouTube)

Leave a Comment