In this one-off interactive, gamified workshop, we’ll simulate real-world work scenarios at your organisation via a board game, helping you identify and eliminate bottlenecks, inefficient processes, and unhelpful feedback loops.
Workshop Details

Last Tuesday I told Claude: "Find my last three email threads with the automotive client, check Google Drive for the latest SAP migration status deck, cross-reference the open items in Notion, and draft a progress update email to the project sponsor. Use formal German."
Then I went to make coffee.
When I came back, the email was sitting in the compose window. It had pulled the right threads, found the right deck, identified the three open items from Notion that actually mattered, and written a status update in the exact tone I use with that client. I read it, changed one word, and sent it.
That's not a chatbot answering a question. That's an agent doing work.
The word "agent" is everywhere right now, and most of the conversation is hype. But underneath the noise, something real has changed in how Claude operates — and if you've been following this series, you've already seen the building blocks. Memory gives Claude context. Projects give it boundaries. Artifacts let it build things. Extended Thinking lets it reason. Agents are what happens when you put all of that together and point it at a real task.
Most of the architectural thinking behind this comes from Anthropic's own research — particularly their "Building Effective Agents" piece and their learning platform. What follows is how those ideas actually play out when you're a consultant using Claude every day.
There's an important difference that most people miss when they hear "AI agent." Anthropic draws a clean line between two things:
Workflows are sequences where the steps are predefined. You design the path, Claude follows it. First do this, then check that, then produce this output. Predictable, consistent, and you control the logic.
Agents are systems where Claude decides the steps itself. You give it a goal, and it figures out which tools to use, in what order, and when to stop. Flexible, adaptive, and you trust the model to navigate.
| Workflow | Agent | |
|---|---|---|
| Who decides the steps? | You do. Predefined code paths. | Claude does. Dynamic, based on context. |
| Predictability | High. Same input → same path. | Lower. Adapts to what it finds. |
| Best for | Well-defined, repeatable tasks. | Open-ended problems where you can't predict the path. |
| Risk | Low. Fails predictably. | Higher. Errors can compound. |
| Example | "Translate this text, then format it as a document." | "Research this topic across my tools and draft a recommendation." |
The key insight from Anthropic's research: the most successful implementations use the simplest approach that works. Don't reach for an agent when a workflow will do. Don't build a workflow when a single prompt is enough.
Here's what's interesting. If you've followed this series and configured Claude with memory, projects, and connectors, you're already using agentic patterns — you just might not have the vocabulary for them yet.
| Pattern | What It Means | What It Looks Like in Practice |
|---|---|---|
| Prompt Chaining | Break a task into a sequence. Each step feeds the next. | "Write a blog outline, check it against my style guide, then write the full post." Claude does each step in order. |
| Routing | Classify the input, then send it down the right path. | Claude reads your email and decides: this is a scheduling request → check calendar. This is a client question → search Drive. |
| Parallelization | Run multiple things simultaneously, then combine results. | "Search my Drive for the proposal AND check Notion for open items AND find the latest email thread." Claude does all three, then synthesizes. |
| Orchestrator-Workers | One Claude breaks the problem down, delegates sub-tasks, synthesizes results. | "Prepare a quarterly client review." Claude decides what data it needs, gathers it from multiple sources, and assembles the deliverable. |
| Evaluator-Optimizer | Generate a response, evaluate it, improve it in a loop. | "Write this proposal, then review it for tone and completeness, then revise." Claude iterates on its own output. |
That email I described at the top? That was an orchestrator-workers pattern. Claude broke the task into sub-tasks (search email, search Drive, check Notion), executed each one, reasoned about the combined results, and produced the deliverable. I didn't design that workflow. I gave it a goal and it figured out the steps.
The temptation is to go full agent on everything. Don't.
Anthropic's own recommendation — and my experience backs this up — is to start simple and add complexity only when it measurably improves the result. A single well-crafted prompt beats a complex agentic workflow for most tasks. Here's how I decide:
| Complexity Level | When to Use It | Example |
|---|---|---|
| Single prompt | Task is clear, self-contained, no external data needed. | "Rewrite this paragraph in a more direct tone." |
| Prompt + retrieval | Task needs context from your files or the web. | "Summarize the key decisions from last week's meeting notes." |
| Workflow (chained steps) | Task has multiple sequential steps you can define in advance. | "Draft the newsletter, apply the brand voice, format as HTML." |
| Agent (dynamic steps) | Task is open-ended, requires multiple tools, and you can't predict the path. | "Prepare a client status update from across all my tools." |
The progression matters. Each level adds latency, cost, and the chance of compounding errors. An agent that takes three wrong turns burns time and tokens. A well-scoped prompt that hits the target in one shot is always better — when the task allows it.
After months of using Claude agentically across client engagements, here's what I've found actually matters:
Good tools beat good prompts. This is counterintuitive, but Anthropic found the same thing building their own agents: they spent more time optimizing the tools than the prompts. When Claude has well-connected, well-documented tools — a properly configured Google Drive, a structured Notion workspace, a clean Figma file — the agentic behavior is dramatically better. Garbage tools produce garbage agent loops.
Context is everything. An agent without memory, project instructions, and uploaded reference documents is just a general-purpose AI clicking through your tools randomly. The entire customization stack we've covered in this series — memory, projects, preferences — is what makes an agent effective. It's the difference between handing a task to a stranger and handing it to someone who's worked with you for six months.
Human checkpoints matter. I don't let Claude send emails, publish content, or modify client documents without reviewing the output. The agent does 90% of the work. The last 10% — the judgment call — stays with me. This isn't a limitation. It's the design.
Simplicity wins. The most complex system I've seen fail was a multi-agent pipeline someone built with four chained models and a custom orchestration layer. The most effective agentic behavior I use daily is a single Claude conversation with memory, project context, and connected tools. No framework. No abstraction layer. Just a well-configured AI with access to where my work lives.
I'll be honest: we're early. The agentic capabilities in Claude today are genuinely useful for my work, but they're not autonomous. Claude doesn't run my consulting business while I sleep. It doesn't replace the judgment calls, the relationship management, or the strategic intuition that clients actually pay for.
What it does is eliminate the grind. The research phase. The context-gathering. The first-draft assembly. The cross-referencing across tools. The formatting and document creation. The repetitive tasks that used to eat two hours of every day and required zero expertise — just patience.
That's what agents mean in practice right now. Not artificial general intelligence. Not autonomous businesses. Just a well-configured AI that can string together multiple steps, use your actual tools, and deliver a result that's 90% of the way there.
The remaining 10% is your job. And honestly, that's the part you should be spending your time on anyway.

What Artifacts Actually Are Most people think of Artifacts as "that side panel where Claude shows you code previews." That's like describing a Swiss Army knife as "that thing with the toothpick."

This article breaks down how to choose the right Claude model without overthinking it. It comes down to three variables — capabilities, speed, and cost — and two approaches: start cheap and upgrade, or start smart and optimize later. Includes a comparison of the current model lineup (Opus 4.6, Sonnet 4.6, Haiku 4.5) and a quick decision guide

Extended Thinking didn't change my workflow as dramatically as Memory or Artifacts did. Those features changed what Claude is to me. Extended Thinking changed how well it performs at the things it was already doing. But "how well" matters. A lot.
Not Sure Where to Start?
In this one-off interactive, gamified workshop, we’ll simulate real-world work scenarios at your organisation via a board game, helping you identify and eliminate bottlenecks, inefficient processes, and unhelpful feedback loops.
Workshop Details
MOVE FAST & FIX THINGS.