AI Governance Library

Defense Against the AI Dark Arts: Threat Assessment and Coalition Defense

The United States and its partners must get ready for what the worst people in the world might do with the most advanced AI models.
Defense Against the AI Dark Arts: Threat Assessment and Coalition Defense

⚡ Quick Summary

This Hoover Institution report reframes AI governance from a product safety problem into a national and international security challenge. Instead of focusing on how companies build safe models, it asks how states can defend against malicious use of advanced AI—especially by adversaries. The authors argue that current approaches are too narrow and reactive. They propose a new “coalition defense” model that combines governments, private companies, and allied nations into a coordinated system for threat detection, evaluation, and countermeasures. The report’s core message is clear: AI risk is not just about alignment or misuse—it is about strategic surprise. To address this, governments must rapidly build capabilities to evaluate frontier threats, partner deeply with industry, and create new international structures for AI security.

🧩 What’s Covered

The report is structured around building a comprehensive defense architecture for AI risks. It begins by distinguishing two categories of threats: misuse by malicious actors (e.g., bio, cyber, or strategic attacks) and loss of control over highly capable systems. It emphasizes that governments must develop independent capabilities to assess both, rather than relying on private companies.

A central concept is the creation of a “coalition security enterprise”—a coordinated system across allied nations that performs four key functions: threat assessment, countermeasure development, preparedness planning, and managing technological uncertainty. This includes building institutions (like AI Safety Institutes), expanding secure compute infrastructure, and conducting pre-deployment evaluations of frontier models.

The report then expands the scope beyond current safety practices. It critiques the focus on product-level evaluations and introduces a broader mandate: anticipating adversarial AI systems that may not resemble commercial models at all. This requires governments to experiment with capability elicitation—actively testing how models could be weaponized.

A major section is devoted to international coordination. The authors propose a layered structure: a small core coalition for sensitive defense work, a broader suppliers’ group managing compute and infrastructure risks, and an even wider global dialogue on safety standards.

Finally, the report explores the role of public-private partnerships. It argues that governments lack the resources to match frontier AI development and must rely on a handful of leading companies. This creates complex trade-offs around access to compute, data, talent, and infrastructure. The report also addresses open-weight models, highlighting their irreversible risks and the need for pre-release evaluation.

💡 Why it matters?

This report marks a shift from “AI safety” to “AI security.” It moves the conversation from internal model risks to geopolitical competition and systemic threats. For practitioners, it highlights that governance frameworks focused only on compliance, risk management, or ethics are incomplete.

The key implication is that AI governance will increasingly resemble national security policy. Organizations will need to think in terms of threat modeling, resilience, and adversarial use—not just responsible deployment. The emphasis on coalition-based approaches also signals that AI governance will not be purely regulatory but deeply tied to alliances, supply chains, and infrastructure control.

❓ What’s Missing

The report is strong on strategy but lighter on operational detail. It does not provide concrete implementation pathways for companies or regulators outside the U.S.-centric ecosystem. The proposed coalition model raises significant governance questions—who leads, how decisions are made, and how conflicts between allies are resolved—but these are not fully addressed.

There is also limited discussion of civil liberties, accountability, or democratic oversight, especially given the expansion of government capabilities and classified evaluations. Finally, while the report acknowledges uncertainty, it does not deeply explore alternative governance models beyond the security paradigm.

👥 Best For

This resource is best suited for policymakers, national security professionals, and senior AI governance leaders working on strategic risk. It is particularly valuable for those designing cross-border AI frameworks, public-private partnerships, or high-risk AI oversight mechanisms.

📄 Source Details

Hoover Institution report (December 2024), authored by Philip Zelikow, Mariano-Florentino Cuéllar, Eric Schmidt, and Jason Matheny.

📝 Thanks to

Philip Zelikow, Mariano-Florentino Cuéllar, Eric Schmidt, and Jason Matheny

About the author
Jakub Szarmach

AI Governance Library

Curated Library of AI Governance Resources

AI Governance Library

Great! You’ve successfully signed up.

Welcome back! You've successfully signed in.

You've successfully subscribed to AI Governance Library.

Success! Check your email for magic link to sign-in.

Success! Your billing info has been updated.

Your billing was not updated.