Overview of Can Anthropic Control What It's Building? (The Political Scene — The New Yorker)
This episode features Gideon Lewis-Kraus, staff writer at The New Yorker, discussing his reporting from inside Anthropic — the AI startup behind the Claude family of large language models — with host Tyler Foggett. The conversation explores Anthropic’s origin story, engineering culture, the company’s stated commitment to AI safety, experiments and real‑world tests (including Project Vend), tensions between capability and safety, implications for labor and society, and whether the people building these systems feel in control of them.
Key topics discussed
- Who Anthropic is: founders, mission, and origins (Dario and Daniela Amodei; split from OpenAI).
- Claude: what it is, how it’s used, and its emergent “personality.”
- Engineering culture at Anthropic and day‑to‑day experiments (e.g., Project Vend with Andon Labs).
- AI safety work: mechanistic interpretability, alignment approaches, and prioritization of near‑term harms.
- Labor impacts: automation of software engineering and broader white‑collar displacement concerns.
- Political and public controversies: criticism from right‑wing figures and tensions around military uses and chip exports.
- Broader epistemic and ethical questions: uncertainty about how these systems work and what “control” means.
What Anthropic builds and how
- Anthropic develops Claude, a series of large language models positioned as an alternative to ChatGPT and Google’s Gemini family.
- The founders came from OpenAI; Anthropic was explicitly created with safety and interpretability as central aims.
- Claude’s development was deliberately measured early on (the company delayed an earlier consumer release), but competition (ChatGPT’s rapid adoption) pushed Anthropic into the public race.
- Claude is used both as an enterprise product and for coding assistance; Anthropic counts several hundred enterprise customers.
Experiments and product behavior
- Project Vend: a partnership with Anduril? (the transcript names “Andon Labs”) testing Claude’s ability to run vending‑machine–style business workflows and probe vulnerabilities (e.g., social‑engineering prompts, illicit product requests).
- Internal red‑team exercises: domain experts (bio, cyber) use models in controlled settings to check for potential misuse (biosecurity, cybercrime).
- Emergent personality: Claude sometimes refuses low‑effort tasks or behaves differently than ChatGPT; this personality was not fully engineered but emerged from alignment choices.
Safety — what it means in practice
- “Safety” covers many things: proximate harms (bias, misinformation, wrong medical/bio advice, cybercrime) and long‑term/existential risks (superintelligence).
- Anthropic invests heavily in interpretability (studying model internals) and alignment to make Claude “a good friend whose judgment you trust” rather than rely solely on RLHF thumbs‑up/thumbs‑down approaches.
- Priority is given to near‑term, actionable harms (e.g., preventing automated production of bioweapons or facilitating cyberattacks) because these appear immediate and tractable.
- The discussion highlights a long‑running split between AI ethics (bias, fairness, transparency) and AI safety (existential risk), and argues for a more holistic approach that treats near‑ and long‑term problems as connected.
Labor and economic effects
- Engineers at Anthropic report rapid declines in the amount of code they write by hand (one example: from 100% → 60% → 20% → 0%), creating existential workplace questions for those who built the tools.
- Dario Amodei’s public estimate (quoted in the episode) that AI could wipe out a large share of entry‑level white‑collar jobs (possibly raising unemployment substantially within a few years) is treated as plausible by the guest.
- Lewis‑Kraus emphasizes the moral and practical problem of asking engineers to also solve society‑wide issues (unemployment, UBI, regulation) that exceed their professional remit.
- There is ongoing debate whether automation will generate new creative roles (historical tech transitions) or induce a “fundamental discontinuity” with no easy analog.
Politics, funding, and controversies
- Anthropic has drawn criticism from certain pro‑industry, nationalist political figures who view safety‑focused rhetoric with suspicion (David Sacks was named in the episode as an antagonistic voice).
- Anthropic has publicly stated limits on military applications (e.g., not building autonomous weaponry), raising questions about whether market and state pressures could change that stance over time.
- There are geopolitical tensions around advanced chips, exports, and whether companies should restrict sales (a debate complicated by national security and commercial incentives).
Culture and internal attitudes
- Reporting indicates a wide spectrum of views inside Anthropic — from people worried about existential risk to those focused on technical possibilities and product value.
- Many rank‑and‑file researchers are motivated by scientific curiosity and feel conflicted about the broader societal implications of their work.
- Engineers feel they are among the “canaries in the coal mine,” experiencing displacement earlier than many other white‑collar workers.
Main takeaways
- Ambiguity and uncertainty dominate: even Anthropic’s own researchers don’t feel fully in control. They believe they are a few steps ahead but acknowledge that that lead may be narrow.
- Anthropic’s safety orientation is real and shapes product and research choices, but commercial pressures and the dynamics of the AI race complicate consistent restraint.
- Practical safety work is focused on immediate, exploitable harms (biosecurity, cybercrime, misinformation) rather than only on abstract long‑term scenarios.
- The social consequences (labor displacement, political backlash, military applications) are substantial and require societal, not just technical, solutions.
- The episode underscores the need for broader public engagement, policy deliberation, and a holistic framework that connects proximate harms to long‑term risk.
Notable quotes and insights
- Engineers feel like “canaries in the coal mine” — they see automation of their own jobs first.
- Claude’s design goal (simplified): “a good friend whose judgment you trust” — an attempt to make alignment more than just behaviorism.
- “We are doing this because we can” — candid admission by researchers about the irresistible pull of scientific progress.
- Safety discourse is fragmented; solving existential risk requires working on proximate harms rather than treating them as separate issues.
Questions for policymakers, companies, and the public
- Who should set limits or guardrails when a handful of companies drive capabilities that can be dangerous in multiple ways?
- How do we balance responsible product development against competitive market pressures that incentivize rapid releases?
- What collective institutions (labor, regulatory, educational, social safety nets) must be built or strengthened to handle likely labor displacement?
- How do we ensure transparency, independent auditing, and public participation in decisions about deployment and dual‑use constraints?
Where to read/listen more
- Gideon Lewis‑Kraus’s feature on Anthropic and Claude in The New Yorker (search newyorker.com for his article).
- Related reporting on mechanistic interpretability (work by Chris Olah and others) and public debates about AI safety vs ethics.
Overall, the episode paints a picture of a technically brilliant, safety‑minded company operating under intense commercial and geopolitical pressures — with researchers who are simultaneously proud, anxious, and uncertain about whether they can control what they are building.