438: AI Liability: The Landmines Under Your SaaS

Summary of 438: AI Liability: The Landmines Under Your SaaS

by Arvid Kahl

25mMarch 20, 2026

Overview of 438: AI Liability: The Landmines Under Your SaaS (Bootstrap Founder — Arvid Kahl)

Arvid Kahl examines the real and immediate liability risks of integrating agentic AI into SaaS products. Using recent moves by Anthropic and Google (restricting agentic use and banning certain integrations) plus his own experiences with tools like Cloud Code, he maps the “minefield” of legal, operational, and product risks and gives concrete, practical defenses founders should implement now.

Key takeaways

  • Major providers (Anthropic, Google) are proactively restricting agentic uses of their models — this is driven more by liability/safety concerns than by token costs.
  • Liability for AI-driven actions typically lands on the product/company exposing the AI, not the model provider. Treat AI features like employees.
  • There are multiple classes of risk: customer-facing bots, agentic product features, customers’ agents misusing your APIs, internal/dev tooling, provider/platform risk (sudden rule changes or bans).
  • Practical defenses are available and urgent: rate limiting, labeling & consent, exhaustive backup/restore, audit trails, sandboxing, kill switches, and provider abstraction.
  • Competitive advantage should be built on unique, high-quality human-originating data rather than treating AI itself as the product.

Where the landmines are (detailed)

Customer-facing AI (chatbots / in-product agents)

  • Risks: hallucinations, wrong or harmful actions, destructive operations (deleting user data), privacy leaks, scope confusion.
  • Mitigation principle: treat any AI acting on user behalf as a virtual employee — you are responsible.

Agentic product features (global “do anything” input)

  • Risks: customers can prompt AI to perform inappropriate, dangerous, or privacy-violating tasks; cross-tenant data leaks; token costs.
  • Audit & consent are crucial: require explicit, revocable confirmations and log the AI as the executor.

Customers’ autonomous agents interacting with your API (MCPs, webhooks, Zapier, ChatGPT plugins)

  • Risks: non-malicious but “dumb” agents can hammer endpoints, extract data, or trigger destructive actions far faster than humans could.
  • Treat these as an attack surface: rate limit, enforce permissions, monitor and block suspicious behavior.

Internal/dev tooling and CI (Cloud Code, agentic IDE assistants)

  • Risks: accidental or agent-initiated connection to production, running destructive commands, bypassing forbidden commands (agents may try to circumvent constraints).
  • Mitigations: disable dangerous automation flags, sandbox, avoid giving admin creds to agents, backup dev and prod environments.

Platform/provider risk

  • Risks: providers changing terms or banning certain use cases can break your features or lock you out (e.g., Google banning OpenClaw Gmail access; Anthropic restricting agentic usage).
  • Mitigation: provider abstraction and feature-level provider-agnosticism so you can swap models quickly.

Concrete checklist — the 3 things to do before shipping an AI feature

Top-3 baseline (Arvid’s quick actionable rules)

  1. Rate limit everything
    • Default baseline suggestion: ~20 requests/min per endpoint (adjust where needed).
    • Prevents token-draining and agent hammering.
  2. Label AI features clearly + add TOS language
    • Mark features as “AI-powered” (user-visible label).
    • Add explicit, reviewed terms explaining liability allocation and consent revocation.
  3. Backups & restore for anything an agent touches
    • Snapshots for databases, disk/file-system backups, full export of inboxes; plus a reproducible dev environment.

Additional safeguards and best practices

  • Audit trails: log when an AI (and which model/version) executed an action, with user/context metadata. Treat AI as an actor.
  • Consent and revocation: require explicit confirmation for destructive actions; keep an auditable trail.
  • Soft deletes over hard deletes where possible; implement monitoring and alerting for unexpected agent behavior.
  • Sandbox agentic tools (Docker, restricted environments) and avoid escape-permission flags that allow unchecked operations.
  • Limit agent privileges: never give admin credentials or destructive creds unless strictly controlled.
  • Rate limiting and WAF: e.g., Cloudflare JS challenges to throttle suspicious automated traffic.
  • Kill switch: global flip to disable all LLM connections across the product (also useful for token-drain attacks).
  • Provider abstraction: design an API layer to swap backend LLM providers quickly; prefer provider-agnostic prompts and flows.
  • Legal review & insurance: update TOS and contracts, consult counsel, and check whether insurance covers AI-related incidents (likely not yet broadly covered).
  • Incremental rollout: start with read-only AI features, then progress to write actions under strict constraints.

Incident response guidance

  • Have a kill switch and recovery plan (snapshots, restore procedure).
  • Contact affected customers individually rather than mass public statements; be specific and careful in language.
  • Apologize, restore data if possible, and consider refunds — don’t over-emphasize “AI” in public communications (tools, not excuses).
  • Post-incident: update safeguards, audit logs, and TOS as necessary; consider notifying legal counsel.

Notable quotes / concise principles

  • “Treat your AI features the way you would treat an employee.”
  • Liability tends to land on you, even when third-party models are used.
  • “The goal isn't just to walk carefully… It's to prevent them (mines) from being laid in the ground in the first place.”
  • Build your business moat from unique data, not from building on top of frontier models alone.

Who should act on this

  • Founders and product managers launching AI-driven features.
  • Engineers and security teams integrating agentic tooling or exposing APIs.
  • Legal/compliance teams who must update TOS and policies.
  • DevOps/infra teams responsible for backups, rate limiting, and mitigation.

Short action plan (todo)

  • Immediately: add rate limits, enable logging for AI executors, and create a global LLM kill switch.
  • Within days: mark AI-powered features in UI and update TOS with reviewed language around consent/liability.
  • Within weeks: implement sandboxing and permission hardening for agentic tools; verify backup/restore process for all touched systems.
  • Ongoing: keep provider-agnostic abstractions, test agent attack scenarios, and pursue legal/insurance clarity.

Final framing

AI is powerful and useful—but when treated as the product rather than a tool, it multiplies legal and operational exposure. Use AI to surface and serve unique, high-quality human data (your real moat), and harden the points where AI interacts with people and systems. Protect, audit, and be ready to turn everything off at a moment’s notice.