Pentagon Insider: What's Next For Anthropic and The Department of War — With Michael Horowitz

Summary of Pentagon Insider: What's Next For Anthropic and The Department of War — With Michael Horowitz

by Alex Kantrowitz

48mMarch 4, 2026

Overview of Big Technology Podcast — “Pentagon Insider: What's Next For Anthropic and The Department of War”

Host Alex Kantrowitz interviews Professor Michael Horowitz (Univ. of Pennsylvania; former Deputy Assistant Secretary of Defense for Force Development and Emerging Capabilities). The conversation examines the sudden breakdown between Anthropic and the U.S. Department of Defense (referred to on the show as Department of War): what sparked it, how Anthropic’s Claude is actually being used inside the military, the meaning and risk of the Pentagon’s “supply chain risk” designation, and broader implications for AI in warfare.

Key takeaways

  • The Anthropic–DoD conflict is driven as much by personalities, politics, and trust as by substantive technical disagreements. A triggering incident (questions raised about Anthropic tech being referenced in a Maduro extraction operation) exacerbated tensions.
  • The immediate contractual dispute centers on wording around prohibited uses (e.g., “mass surveillance” vs. “all lawful uses”) and on whether a vendor can carve out certain downstream uses of its product.
  • Anthropic was one of the first frontier AI labs willing to do classified work for the U.S. military; its models (Claude) were integrated through platforms like Palantir’s Maven Smart System as a decision-support input—not as an autonomous targeting tool.
  • Current military use of LLMs is primarily for decision support, open-source/media synthesis, simulations, and experimental prototype work—human review layers remain central.
  • The Pentagon treats AI vendors more like traditional weapons/hardware suppliers (who don’t dictate use), while Anthropic sees its models as continuously-updated services that merit ongoing governance over use cases.
  • The Pentagon’s “supply chain risk” designation (reserved historically for entities like Huawei) and talk of the Defense Production Act create legal, political, and commercial confusion and could chill industry-government cooperation.
  • On AI and warfare: Horowitz frames AI as a general-purpose technology with three primary military impact buckets—administrative/enterprise efficiency, intelligence/surveillance/decision-support, and battlefield/autonomous systems. Readiness and risks vary greatly by bucket.

Background: what happened and what triggered it

  • Anthropic had a contract relationship with the Pentagon (classified and unclassified work). After the U.S. operation to remove Venezuela’s Maduro (reported trigger), Anthropic asked whether its tech had been used (via Palantir/Maven integration). The Pentagon reportedly took offense that Anthropic inquired.
  • The Pentagon updated AI contracting policy to require an “all lawful uses” clause for new AI contracts. Negotiation over contract language (mass surveillance carve-outs and autonomous-weapons exclusions) escalated amid eroding trust.
  • The DoD canceled the contract, designated Anthropic a supply chain risk (restricting DoD vendors from commercial dealings with Anthropic), and publicly signaled aggressive enforcement—while also reportedly using Anthropic tech in ongoing operations (demonstrating a contradiction).

How the Pentagon is using Anthropic’s tech (per Horowitz)

  • Claude is typically one input in integrated dashboards (e.g., Palantir’s Maven Smart System) used by combatant commanders for situational awareness.
  • Uses include: querying open-source information and media, synthesizing intelligence chatter, rapid simulation and scenario generation, and experimental prototyping.
  • Horowitz doubts Claude is being used for autonomous battlefield targeting. Any AI outputs are likely subject to layered human review before operational effect.
  • The most forward-leaning command in experimenting with such tools has been U.S. Central Command.

Core policy and terminology issues

  • “All lawful uses” vs. vendor-specific carve-outs: Pentagon sees such carve-outs as an unprecedented attempt by a vendor to limit how the U.S. can use technology it procures. Anthropic wants assurances against certain downstream uses (e.g., mass surveillance, autonomous targeting).
  • “Fully autonomous weapons” is an imprecise public phrase; the policy term is “autonomous weapon systems” (weapon systems that, after activation, select and engage targets without further human intervention). Many forms of autonomy have long existed in weapons (e.g., homing munitions, close-in weapon systems).
  • Technical readiness: Anthropic’s claim that current frontier models aren’t ready for frontline autonomous weapons is plausible—autonomy at the edge requires different models and reliability guarantees than LLMs trained on broad public data.

Legal / commercial consequences and risks

  • The DoD supply-chain-risk label historically applies to entities deemed national-security threats (e.g., foreign adversary vendors inserting backdoors). Applying it to Anthropic is unusual and legally contestable.
  • Potential outcomes are messy and contradictory: supply-chain designation blocks commercial ties, while invoking the Defense Production Act could compel work—these pull in opposite directions.
  • Short-term effects: several federal agencies (State, Treasury, HHS) were reportedly told to stop using Claude for unclassified tasks. The classified environment is harder to replace quickly because Anthropic was an early classified integrator.
  • Longer-term risk: Other AI firms may hesitate to enter classified programs if refusal to accept certain DoD terms risks total blacklisting. That could chill innovation and reduce competition.

Where this leaves Anthropic and the DoD

  • Worst-case: a protracted legal/political battle and sustained decoupling—damage to Anthropic’s government business and reputational harm in public sector channels.
  • Best-case: bargaining/renegotiation, a mediated compromise on contract language, or alternative classified arrangements (e.g., multi-vendor consortia) that restore cooperation.
  • Political dynamics matter—administration changes or internal DoD politics could rapidly change the posture.

Broader implications for AI and warfare (Horowitz’s framework)

  • Think of AI as a general-purpose technology with three main military roles:
    1. Administrative and logistical modernization (paperwork, procurement, payroll).
    2. Intelligence, surveillance, reconnaissance (ISR) and decision support—speeding analysis, reducing information overload, improving commander decision timelines.
    3. Battlefield applications including autonomous systems and algorithmic operational planning—here, different algorithms and rigorous testing/validation are required, and LLMs may not be the right tool for edge autonomy.
  • Key point: treat AI in military use case–specific terms rather than a monolithic “AI will/do X” framing. Readiness, ethical risks, and governance needs vary by use case.

Notable quotes / concise paraphrases

  • “Personalities and politics masquerading as a policy dispute.” — Horowitz on the Anthropic–DoD split.
  • “Anthropic was the first frontier AI lab willing to do classified work.” — explains why this split is uniquely disruptive.
  • “From the Pentagon’s perspective, what Anthropic is asking for is unprecedented” — vendors usually don’t tell the military how to use procured tools.
  • “Anthropic is not wrong that their tech isn’t ready for prime time for autonomous weapon systems.” — on LLM limitations for targeting at the edge.
  • The supply-chain-risk label “is not good for American innovation or the American economy” if applied in this manner.

Practical implications / recommendations (for different audiences)

  • For policymakers: avoid ad hoc, politically driven branding (e.g., supply chain risk) without clear legal and technical rationale—use case–specific risk assessments and transparent processes reduce chilling effects.
  • For AI companies: expect the DoD to seek broad usage rights; consider early legal and policy strategies for government work (clarify carve-outs, plan for classification requirements, and contingency scenarios).
  • For defense practitioners: continue emphasizing rigorous testing & evaluation (T&E) and human-in-the-loop safeguards when integrating LLMs; be explicit about distinction between decision-support and edge-autonomy.
  • For the public and journalists: evaluate claims about “AI targeting strikes” skeptically—current evidence and expert testimony indicate LLMs are more likely being used for intelligence synthesis and decision support rather than direct autonomous targeting.

Final observation

This episode highlights a fragile intersection of technology, national security, and politics. The Anthropic–DoD clash is less about immediate battlefield capability and more about trust, contract governance, institutional roles, and how a government accustomed to buying hardware adapts to purchasing ever-evolving AI services. The resolution will shape future industry willingness to partner with the U.S. government and influence how AI is fielded across the three buckets Horowitz outlines.