Head of Claude Code: What happens after coding is solved | Boris Cherny

Summary of Head of Claude Code: What happens after coding is solved | Boris Cherny

by Lenny Rachitsky

1h 27mFebruary 19, 2026

Overview of Head of Claude Code: What happens after coding is solved (Lenny Rachitsky feat. Boris Cherny)

This episode features Boris Cherny (Head of Cloud Code at Anthropic) in conversation with Lenny Rachitsky. They reflect on the last year of rapid AI-driven change in software development, the origin and impact of Cloud Code (aka “QuadCode”), how agentic AI (Cowork) is expanding beyond coding into general knowledge work, and what builders and organizations should do to succeed and stay safe as these tools scale.

Episode summary — what was discussed

  • Origin story of Cloud Code: Boris prototyped a terminal-based coding agent, iterated quickly, launched internally, then publicly; adoption accelerated with model improvements.
  • How Cloud Code/Cowork work: agents that can use tools (run commands, access Slack/Gmail/browser, edit repos) and act on behalf of users—not just chatbots.
  • Concrete impact: rapid adoption internally and externally; fast growth and compounding usage metrics.
  • Product and engineering principles used at Anthropic to build Cloud Code: latent demand, minimal scaffolding, under-resource early work, and betting on future model capabilities.
  • Practical tips for teams and individuals on building with and using agents (including specific product usage advice).
  • Safety & alignment approach: three-layer model (mechanistic interpretability, evals, real-world testing) and open-source tooling to sandbox agents.
  • Broader implications: coding is “largely solved” for many tasks; roles will blur (more builders/PMs), and teams should prepare culturally and operationally.

Key metrics & impact (claims from the episode)

  • A semi-analysis report cited: ~4% of public GitHub commits now authored by Cloud Code; the report predicted rapid growth (up to ~20% by year-end).
  • Boris: many engineers (including himself) now rely entirely on Cloud Code—he claims 100% of his code has been generated by Cloud Code since November and he ships 10–30 PRs/day.
  • Reported internal gains: productivity per engineer up ~200% (measured in PRs), and Cloud Code daily active users continue to accelerate (Boris: DAUs recently doubled month-over-month for a period).
  • At Anthropic, Claude reviews 100% of pull requests (automated reviews with human checkpoints for safety).

Note: some funding/revenue numbers mentioned in the conversation are likely imprecise in the transcript; take dollar figures with caution.

What Cloud Code / Cowork enable (agent capabilities)

  • Agents can:
    • Use tools and interact with systems (run bash, open browser, read/write Slack, Gmail, Google Docs, spreadsheets).
    • Create code, run tests, open PRs, and do code review.
    • Perform non-coding tasks (fill forms, process emails, PM workflows, analyze data, automate repetitive browser workflows).
  • “Agent” defined technically: an LM that can use tools—hence it acts, not just answers.

Product & engineering principles Boris emphasizes

  • Latent demand: observe how users misuse or repurpose tools; productize those paths rather than forcing new workflows.
  • Build for the model six months from now (bet on improvements, not only current limits).
  • Under-resource early prototypes: give small teams ownership and speed; constrained resourcing forces clarity and quick shipping.
  • Don’t over-orchestrate the model: provide tools + goals; let the model decide how to use them rather than rigidly prescribing step-by-step workflows.
  • Bet on general models (the “bitter lesson”): more general, capable models tend to outperform narrowly engineered systems over time.
  • Give engineers freedom to experiment (e.g., generous tokens) — early experimentation yields novel ideas.

Practical tips and recommendations (for teams and individual users)

For builders / product teams:

  • Give your team tokens and permission to experiment early; optimize costs later only if the idea scales.
  • Prioritize exposing the model to useful tools instead of heavy pre-curated workflows.
  • Release early (research preview) to learn about real-world behavior and safety issues.
  • Look for “latent demand” to expand product form factors: bring the product to where users already work (terminal, IDE, Slack, desktop app, mobile).

For Cloud Code / Cowork users (Boris’s pro tips):

  • Use the most capable model available (Boris: Opus 4.6 when recording).
  • Enable “plan mode” often: have the agent outline the plan before executing (reduces iteration).
  • Try different interfaces (terminal, desktop app, mobile, Slack, Chrome extension) to fit your workflow.
  • Point an agent at a Slack feedback thread or issue feed—let it propose PRs and fixes; rapid response drives more feedback and adoption.
  • Run multiple agents in parallel for different tasks (project management, bug triage, automation).

Safety & alignment approach (how Anthropic thinks about risk)

  • Three-layer approach:
    1. Mechanistic interpretability — peek “inside” models (neuronal activations, circuits) to understand internal representations and planning.
    2. Evals — laboratory-style tests across scenarios to quantify behavior.
    3. Real-world testing — controlled releases / research previews to observe agent behavior in practical contexts.
  • Anthropic open-sources safety tooling like sandbox environments to constrain agent access and encourage a “race to the top.”
  • Continuous feedback loop: findings from production feed back into model and product improvements.

Implications for jobs, careers and organizations

  • Short/medium term: coding for many practical tasks is becoming automated. Engineers’ workflows shift toward supervision, architecture, product thinking and safety.
  • Longer term: roles will blur—more “builders,” more product-minded people who also can code via agents. Product managers, designers, data scientists and many knowledge workers will be impacted as agents handle tool-based work.
  • Career advice Boris offers:
    • Experiment constantly with agent tools and become AI-native.
    • Grow as a generalist who crosses disciplines (product ↔ engineering ↔ design ↔ business).
    • Focus on higher-level thinking (what to build, user understanding, system design) rather than low-level implementation.

Notable quotes

  • “100% of my code is written by Cloud Code. I have not edited a single line by hand since November.”
  • “Coding is largely solved.”
  • “The model is starting to come up with ideas… a little more like a coworker.”
  • “Latent demand is the single most important principle in product.”
  • “Don’t try to box the model in.”
  • “Bet on the more general model.”

Actionable next steps (for listeners)

  • If you’re a builder: prototype with agents, expose the model to tools, and ship early to find latent demand. Give engineers tokens to experiment.
  • If you’re an engineer / PM / designer: try an agent in your workflow (code tab, cowork, Chrome extension). Start with plan mode and the best model available.
  • If you’re a leader: rethink roles and hiring (favor curiosity and cross-disciplinary skills), invest in observability/safety, and plan for rapid model-driven productivity changes.

Additional resources Boris recommends

  • Technical / reading:
    • Functional Programming in Scala (technical taste/best technical book he recommends)
    • Accelerando (by Charles Stross) — fast-paced sci‑fi relevant to exponential change
    • Wandering Earth (collection) — Chinese sci‑fi perspective (Si Xin Liu)
  • Podcasts/products he likes: Acquired podcast; Cowork (Anthropic product) and Cloud Code (QuadCode).
  • Where to follow Boris: Twitter / X or Threads (he asks listeners to send bugs/feature requests there).

If you want a concise checklist to act on from this episode: 1) try an agent on a real low-value task (emails, PM reminders), 2) give one engineer tokens and a week to prototype, 3) measure impact and iterate based on user feedback (latent demand), and 4) instrument safety/evals before broader rollout.