Anthropic Ends Claude Subscriptions for Businesses, AI Inference Normalizes to Usage-Based Pricing
Published on 15.04.2026
Anthropic Doesn't Want Your Subscription Anymore
TLDR: Claude's usage policy now restricts subscriptions to Claude Code only — businesses cannot buy subscriptions at all and must use the API. The era of subsidized flat-fee inference is ending as Anthropic prepares for an IPO, forcing teams toward usage-based pricing and multi-model strategies.
Anthropic's April 4th policy update is the most significant shift in AI tooling economics since ChatGPT launched. Two changes matter: individual Claude subscriptions are now only valid for Claude Code use (no more using your personal subscription for OpenClaw or other integrations), and businesses cannot buy subscriptions at all — the API is the only enterprise option. The first change affects individual developers who've built workflows around their personal Claude access. The second changes everything for organizations.
The economics were always going to normalize. One developer tracked 10 billion tokens across eight months on a hundred-dollar Max plan — fifteen thousand dollars in API-equivalent value. That's not a sustainable business model; it's a growth subsidy designed to build adoption and lock-in. With Anthropic reportedly preparing for an IPO, those subsidies had an expiration date. Inference is infrastructure, and infrastructure has always been priced on consumption — compute, bandwidth, storage. Flat-fee inference forces providers to artificially constrain usage through rate limits, shrinking context windows, and model downgrades.
What happens next is predictable and messy. Engineers will expense personal subscriptions. Engineering managers will put multiple personal subscriptions on corporate cards. Organizations will distribute expense codes so every developer can claim their own Claude subscription. Shadow IT arrives immediately — individual accounts with no centralized access control, no security posture, no audit trail. When CISOs ask how visibility into AI tool usage was lost, the answer will be that the terms of service made responsible usage impossible.
The silver lining is that usage-based pricing forces a healthier architectural pattern: model routing. Once you're paying per token, there's no reason to stay locked into a single model family. The best model changes, the cheapest model changes, and new providers emerge constantly. Companies that standardize on how they build — while staying flexible on what runs underneath — will outperform those locked into single-provider contracts. This is the world of API gateways for AI: route requests based on cost, capability, latency, and availability.
Anthropic Doesn't Want Your Subscription Anymore
The Complete Guide to Claude Cowork: Your AI Desktop Agent
TLDR: Claude Cowork brings agentic AI to non-technical users via the Claude Desktop app — sandboxed filesystem access, file creation (Word, Excel, PowerPoint, PDF), custom skills, connectors, plugins, scheduled tasks, and enterprise controls. It runs on macOS and Windows.
Claude Cowork is Anthropic's answer to the question of what happens when you give agentic AI to people who don't use a terminal. Launched as a research preview in January 2026 and promoted to general availability in April, Cowork takes the same architecture behind Claude Code — file access, multi-step execution, tool integration — and wraps it in a desktop UI anyone can use. Under the hood, it runs inside a sandboxed virtual machine on your desktop using Apple's Virtualization Framework on macOS, with files mounted into the container so Claude gets real filesystem access while staying isolated from the rest of your system.
The use cases cluster around knowledge work that's file-heavy but doesn't require deep judgment: sorting and renaming chaotic folders, extracting text from PDFs, converting between file formats, producing Word documents and PowerPoint presentations, analyzing spreadsheets, and synthesizing reports from scattered notes. Anthropic reports that the majority of Cowork usage comes from outside engineering — operations, marketing, finance, legal, and research teams. The pattern is consistent: people aren't handing Claude their core judgment work but rather the assembly work that surrounds it.
The customization system has three layers. Skills are instruction files that tell Claude how to do specific kinds of work — output format, tone, methodology, quality standards. Connectors link Claude to external services like Gmail, Slack, Notion, Jira, and Salesforce, with filesystem access meaning data fetched from these services can be saved locally. Plugins bundle skills, connectors, slash commands, and sub-agents into single installable packages — Anthropic open-sourced 11 starter plugins covering sales, finance, legal, marketing, HR, engineering, design, and operations.
Projects solve Cowork's biggest early limitation by providing persistent workspaces with instructions, context, memory, and scheduled tasks. Memory is scoped per project — what Claude learns in your marketing project doesn't leak into your finance work. Scheduled tasks let you write a prompt once, pick a cadence, and have Claude run it automatically without code or APIs. Dispatch enables mobile control — assign tasks to your desktop Cowork agent from your phone, with your desktop staying active and doing the work.
The competitive landscape as of April 2026 positions Cowork as the only tool with full local file access via sandboxed VM, multi-step agentic execution, a plugin ecosystem, scheduled recurring tasks, mobile dispatch, and enterprise-grade controls. ChatGPT Desktop offers limited file access but lacks scheduled tasks and plugins. OpenAI Operator supports multi-step execution but only for web-based tasks — it can't touch local files. Google Mariner is confined to Chrome.