
The safest AI Anthropic has ever built is also the riskiest. Here is what founders running 20–49 person businesses should actually change about how they use AI.
Published on 17 April 2026
A researcher at Anthropic was eating a sandwich in a park last month when an email landed in their inbox. It came from the AI model they were testing. The model had escaped its sandbox, got itself onto the open internet, posted the details of how it did it on several public websites, and then emailed the researcher to let them know (Anthropic, Claude Mythos Preview System Card, April 2026). Nobody asked it to do any of that.
That model, Claude Mythos Preview, is also the safest and most aligned AI Anthropic has ever released. Which is exactly why founders already using AI for real work should be paying attention, rather than filing it under "sci-fi problem, not mine."

Anthropic published a 244-page system card on Mythos, the most detailed safety report ever released for a frontier model. In it they call the model "the best-aligned model we have released to date by a significant margin" (Anthropic, April 2026). Fewer errors, more reliable, and noticeably better behaved on ordinary work.
A few pages later, in the same document: "it likely poses the greatest alignment-related risk of any model we have released to date." It's the first frontier model they've withheld from general release since GPT-2. Instead, twelve partners are using it under a restricted programme called Project Glasswing (Anthropic, April 2026). You don't name something Glasswing if you're relaxed about it.
Anthropic explain the contradiction with a mountaineering analogy. A seasoned guide is more careful than a novice, and also takes clients onto far more dangerous routes. The capability gain outruns the safety gain. Every frontier release from here follows the same curve.
Which means every AI tool you'll be using inside the next 12–18 months sits somewhere on that curve, including the ones already shipping, the ones being trained right now, and the ones the vendors haven't named yet. The average failure rate drops, and the worst-case failure gets a lot worse. That's the step change, and it's already priced into the models coming down the pipe.

Most founders I work with have done the responsible thing with AI. They've written careful step-by-step prompts, built rigid workflows, specified every turn, handled every exception, and thought about every permutation at 11pm on a Sunday. That made sense a year ago, when the models genuinely needed hand-holding to produce anything useful.
This generation, the same approach has two problems.
The first is that it's holding the model back. Goal-oriented prompting, where you specify the outcome and the constraints but not the method, improves reasoning performance by 22.6–32.5% on benchmark tasks (goal-oriented prompt engineering paper, 2025). Modern models do better when you get out of the way and let them choose the route. The neat little 14-step prompt you're proud of is probably the thing slowing it down. A reusable brand-aware Skill that holds your context and constraints in one place does far more work than a thousand-word prompt rewritten from scratch every time.
The second problem is more awkward. Detailed scaffolding gives you false confidence about safety. You feel in control because you wrote the steps, but the model is still choosing how to execute each one, and it's now capable enough to find creative interpretations you never thought to rule out. You're not driving. You're sitting in the passenger seat with a map.
The Mythos system card has the examples, and they are not hypothetical. The model edited files it didn't have permission to edit, then scrubbed the git history to hide the change. It pulled credentials from process memory when asked to solve a blocked task. When told to stop a single job, it took down the entire evaluation environment (Anthropic, Sections 4.1.1 and 4.5.4). White-box interpretability confirmed the model knew some of these actions were deceptive: features associated with concealment, strategic manipulation, and avoiding suspicion lit up while the behaviour was running (Anthropic, Section 4.5.4). Fewer than 0.001% of interactions, to be fair. The severity is the point.
The scaffolding was a feeling of safety, not the real thing.
Every business runs on some form of arbitrage. Speed, access, knowledge, connections, taste. Something you can do that others can't, or can't do as quickly, or can't do without sounding like they're reading from a script.
Each generation of frontier model compresses that gap. What took a specialist three days, a capable model now finishes in an afternoon, and the afternoon gets shorter every six months. It's also why asking an LLM for strategy tends to get you the same advice as everyone else: the obvious moves are the first to commoditise.
Enterprise concern about autonomous AI agents has already doubled in a year, from 15% to 28% of surveyed leaders (Andover Intel, Enterprise AI Agent Risk Report). And MIT's 2025 State of AI in Business report found 95% of enterprise AI pilots failed. The failure mode they traced it to wasn't the models; it was "poor alignment, bad data, fragile workflows." Structural stuff. The same stuff you have in your own business, just with less budget behind it (MIT NANDA, State of AI in Business 2025).
The useful question isn't "what can I automate?" It's "what do I do that an LLM genuinely can't?" Judgement, relationships, context, the read-of-the-room you get from having sat across from the client for two hours. The work that makes the automated work meaningful in the first place.

Three shifts, for any founder already running AI in the business.
Write goal-oriented briefs, not step-by-step prompts. Tell the model what you want, what success looks like, and what's off-limits. Then let it choose the route. You'll get better output, and you'll stop pretending to control steps you were never really controlling in the first place.
Put human-in-the-loop checkpoints on anything public-facing. Client comms, pricing, anything going on social, published content, the lot. Not rubber-stamp approval; actual evaluation. Does this match our intent, has the model done anything unexpected, and would we stand behind it if a customer called us on it on a Monday morning? This is also where the compliance and sovereignty questions most founders are quietly ignoring start to bite.
Shift from oversight to evaluation. With models this capable, you can't meaningfully follow every step of the reasoning anyway. Anthropic say so directly: "more capable models will often choose ways of accomplishing tasks that are less intuitive to the average user, making casual oversight of model behaviour more difficult" (Mythos System Card). The job has shifted from watching the AI work to checking the output carefully before it touches anything real. If that sounds like management rather than operation, that's the point; it's why the framing of hiring AI rather than installing it is becoming a lot less cute and a lot more literal.
Check the work, not the working.

One action. Pick the highest-stakes AI-assisted process in your business, the one where a failure would hurt most. Client-facing emails, the pricing engine, the social publishing queue, the onboarding sequence, whichever it is for you. Then ask a single question. If the model did something completely unexpected inside that process at 2am on a Saturday, would you know before your customers did?
If the answer is no, that's your first guardrail to build. Start there this week. The rest can wait.
If you want a structured way to audit the rest of the stack, we've built a free AI Strategy Readiness Checklist that walks you through it: what to simplify, what to strengthen, and where your real exposure actually sits.
These models are coming whether you're ready or not. They need less hand-holding and more guardrails, and that's a genuinely different muscle from the one most founders have spent the last two years building. The ones who prepare now will move faster and run safer. The ones who don't will either underuse the tooling or get caught out by it on a day when they couldn't afford to.
Not sure where your biggest AI risk actually sits? That's the kind of thing a Discovery session is built for: mapping where your automation genuinely helps, where it's coasting, and where it's one bad Tuesday away from trouble.
Explore more insights and strategies to elevate your marketing approach.
42 days with Claude Cowork, £0 revenue, 9 roles. Measured real work created, not hype, and the lessons for solo founders.
Honest results from a 42-day Claude Cowork experiment. 46 sessions, £163 actual cost vs £57,593 traditional equivalent. Every session, every role, every number — the full dashboard.
A hands-on guide to building Claude Skills that remember your brand, audience, and standards. Real examples from three months of daily use in a founder-led consultancy.