AI Agents Doing Your Newsletter Homework While You Chat
Published on 30.03.2026
How to Run a Competitive Analysis on Substack Newsletters with Claude Cowork
TLDR: Claude Cowork — Anthropic's visual wrapper around the Claude Code agentic engine — autonomously produced a 17-page competitive analysis of a Substack newsletter while the human presenter just talked. No terminal. No copy-pasting. No babysitting.
Summary:
The demo at the center of this piece is genuinely striking, not because the technology is magic, but because of how mundane the friction has become. During a live recording of the One Shot Show, the author pointed Claude Cowork at a competitor's Substack — GenAI Unplugged by Dheeraj Sharma — gave it a goal, and walked away into a conversation with a guest. By the time they looked back, a 17-page Word document was sitting there: audience profiles, writing style DNA extraction, SEO gap mapping, growth strategy, ten specific content ideas. The agent browsed, scraped, analyzed, and composed. The human contributed exactly nothing to that process.
The article frames this as the latest stop on a tool evolution journey that goes: Claude chatbot (you are the middleman), Claude Projects (better context, same bottleneck), Claude Code (mid-2025, terminal-native agentic loops), and now Claude Cowork (January 2026, the same engine wrapped in a GUI with a folder panel, a chat pane, and an output pane). The author is careful to delineate the tradeoffs — Cowork is slower than Code, lacks what appears to be the 1M token context window, and can only touch folders you explicitly grant it. But it requires zero terminal literacy, which is a genuinely meaningful unlock for a large segment of knowledge workers who've been watching agentic AI from the outside.
The workflow the author describes has a satisfying fault-tolerance built in. When Firecrawl MCP credits ran out mid-scrape, the agent didn't stop and ask for help — it silently fell back to Claude's built-in web fetch and continued. That's the thing that separates an agent from an assistant: the assistant waits; the agent reroutes. The CLAUDE.md concept at the heart of Cowork is also worth dwelling on — the idea that you drop a structured file into a folder that encodes your audience, your voice, your content guidelines, and every subsequent task inherits that context automatically. That's not a feature. That's a memory architecture.
What the author surfaces at the end — almost as an aside — is actually the most interesting loop in the whole piece: the output of a competitive analysis (audience profile, voice characteristics, writing DNA of a competitor) can be fed directly back into your own CLAUDE.md as training context. You analyze your competitor and the output becomes the input for mimicking, differentiating, or positioning against them. The agent is eating its own tail in the most productive possible way, and the author barely lingers on how philosophically weird that is.
There's one honest critique worth surfacing here: the article spends considerable time explaining what Claude Cowork is and tracing the evolution from chatbot to agent, which feels slightly like marketing scaffolding around what is actually a workflow tutorial. The five-phase analysis structure — homepage recon, archive dive, content scraping, pattern analysis, report generation — is the genuinely valuable payload, and it gets somewhat buried under the product positioning. The insight that each phase contextualizes the next, so the order is not arbitrary, is the architectural lesson worth extracting and applying to any multi-step agentic workflow, not just newsletter analysis.
Key takeaways:
- Claude Cowork equals Claude Code's agentic engine with a visual GUI — no terminal needed, at the cost of some speed and flexibility
- The agent self-recovered from a tooling failure (Firecrawl credit exhaustion) without human intervention — a meaningful signal of real agentic behavior
- CLAUDE.md files act as persistent project memory that survives across tasks — this is the architectural primitive worth paying attention to
- The competitive analysis output (audience profile, voice DNA) can feed directly back into your own CLAUDE.md, creating a compounding loop
- Five-phase task decomposition (context first, then depth) is a transferable pattern for multi-step agentic workflows
- Power users should stay in Claude Code; non-technical users get most of the same capability through Cowork
Why do I care: As someone who thinks in systems and has watched a decade of productivity tools promise to remove toil while quietly creating new kinds of it, this one actually lands differently. The CLAUDE.md pattern is the piece every frontend architect should pay attention to — it's essentially a project-scoped system prompt with real persistence, and it starts to resemble what professional AI tooling actually needs: not a smarter autocomplete, but a collaborator that already knows the conventions, the audience, and the constraints before you type the first word. The agent autonomy here is still narrow — it's browsing and writing, not reasoning about business strategy — but the self-recovery behavior and the output-feeding-back-into-context loop suggest we're closer than most teams realize to AI that genuinely compounds its own usefulness over time.
How to Run a Competitive Analysis on Substack Newsletters with Claude Cowork