In this one-off interactive, gamified workshop, we’ll simulate real-world work scenarios at your organisation via a board game, helping you identify and eliminate bottlenecks, inefficient processes, and unhelpful feedback loops.
Workshop Details

Every model decision comes down to three things: capabilities, speed, and cost. That's it. No secret fourth variable. No hidden complexity. Just three levers you need to balance.
Capabilities: What does the task actually require? Are you doing complex multi-step reasoning, or are you classifying support tickets? The difference between those two determines whether you need a sports car or a delivery van.
Speed: How fast does the response need to arrive? If you're building a customer-facing chatbot, latency matters. If you're running a batch analysis overnight, it doesn't. Claude Opus 4.6 even has a fast mode (research preview) that gives you up to 2.5x higher output speed at premium pricing — useful when you need both intelligence and speed.
Cost: What's the budget? Not just for production, but for development too. Burning through expensive API calls during prototyping is a classic beginner mistake.
Know these three answers before you start evaluating models. Write them down. Tape them to your monitor. Whatever it takes. Because every bad model choice I've seen in the wild comes from someone skipping this step.
There are exactly two strategies for picking a starting model, and neither is wrong. But one is probably better for your situation.
My personal bias? Start with the most capable model for your first proof-of-concept, get the quality right, then optimize downward. It's easier to simplify a working solution than to debug a cheap one that's subtly wrong.
If you're still overthinking it, here's the shortcut:
Building enterprise agents, doing professional software engineering, or running multi-hour research tasks? → Opus 4.6. Don't even think about it.
Code generation, data analysis, content creation, or agentic tool use at scale? → Sonnet 4.6. The sweet spot of intelligence and throughput.
Real-time applications, cost-sensitive deployments, or sub-agent tasks inside a larger system? → Haiku 4.5. Fast, capable, won't drain your budget.
Here's where most people get it wrong: they pick a model and never re-evaluate. The right approach is to build a benchmark — a set of tests specific to your use case — and run it periodically.
Create test cases that cover your actual prompts and real data. Compare accuracy, response quality, and how the model handles edge cases. Then weigh those results against cost. Sometimes the expensive model saves money because it gets things right on the first try. Sometimes the cheap model is 95% as good at 20% of the price. You won't know until you measure.
Anthropic's docs have a solid guide on developing evaluation tests — I'd recommend starting there. But the most important thing is that you have an evaluation set. A good eval set is the single most valuable asset in any AI implementation. More valuable than the model choice itself.
If you want to go deeper — and you should — Anthropic launched their Anthropic Academy earlier this month. It's a free learning platform on Skilljar with 13 structured courses covering everything from basic Claude usage to production API deployment and MCP server development. Free. With certificates.
The courses are organized into three tracks: AI Fluency (for managers and non-developers who need to understand what this stuff actually does), Product Training (for professionals integrating Claude into workflows), and Developer Deep-Dives (API fundamentals, prompt engineering, tool use, MCP). The curriculum was co-designed with academics and there's a Higher Education Advisory Board chaired by Rick Levin, former president of Yale.
Start with Claude 101 if you're new, jump to the API course if you're a developer, and take the prompt engineering tutorial if you want to get serious about quality. The model selection stuff I covered above is just the beginning — the real craft is in how you prompt, evaluate, and iterate.
You'll find everything at anthropic.skilljar.com or through anthropic.com/learn.
Not Sure Where to Start?
In this one-off interactive, gamified workshop, we’ll simulate real-world work scenarios at your organisation via a board game, helping you identify and eliminate bottlenecks, inefficient processes, and unhelpful feedback loops.
Workshop Details