What Anthropic's leaked code says about how they lead and work together
By Paul Musters, founder of emaho.
A software leak usually gets attention for the technical secrets inside. What's more interesting is what the code reveals about how people actually work with each other at Anthropic. How they communicate. What leadership has decided matters. What they've clearly had to learn the hard way.
They think about the person who comes after them
Throughout the codebase, developers leave notes explaining why they made a specific call. That comes from a team that assumes the next person will continue their work. Knowledge in people's heads walks out when those people leave. Knowledge in the system stays.
People have real ownership, and the system is built to support it
New features go out in the off position. Anyone can build and release without immediately affecting all users. That's a leadership decision about trust. The trust ladder shows five levels from founder decides everything to founder freedom, where the team runs without the founder in every decision.
Nobody marks their own work complete
Anthropic built a Verification Agent into their internal tooling. After any non-trivial implementation, an independent agent verifies the work before completion can be reported. The rule is precise: "You cannot self-assign PARTIAL. Only the verifier assigns a verdict." This is internal only, used by their own team. Most teams still rely on self-reporting. Anthropic made the separation between doing and verifying structural.
The version they build for themselves is stricter
When an Anthropic employee uses Claude Code, additional instructions activate. The AI is told to push back on misconceptions, report failures directly, and verify work before claiming it's done. There's also a type called I_VERIFIED_THIS_IS_NOT_CODE_OR_FILEPATHS, where every analytics call requires using the full type name, making data safety confirmation a conscious micro-ritual at every call site. The standard they hold themselves to internally is higher than what they ship externally.
The culture that made them fast also made them exposed
Anthropic optimises heavily for speed and continuous shipping. The same internal material ended up visible to the public twice in a year. Every cultural strength has a shadow. Speed and openness are real strengths. They're also how you end up on the front page twice for the same reason.
Frequently asked questions
What does Anthropic's leaked code reveal about their company culture?
The leaked Anthropic codebase shows a team that writes documentation for the person who comes next, not to cover mistakes. Developers embed the reasoning behind decisions directly into the work. Features ship in the off position so people can build and test without affecting all users. Accountability is structural: no one can mark their own work complete. The code reveals a culture where trust is built into the system itself.
How does Anthropic handle accountability across their team?
Anthropic built an internal verification system where any meaningful piece of work must be reviewed by an independent check before anyone can report it complete. The rule is explicit: the person who did the work cannot close the loop themselves. This applies to their own internal team only. It is a structural answer to the problem that when someone judges both whether their work is done and whether it is good, those two judgements tend to blend together.
What does it mean that Anthropic holds its own team to a higher standard?
When an Anthropic employee uses Claude Code, the tool activates additional behavioral instructions. It is told to flag requests based on wrong assumptions, report failures plainly rather than softening them, and confirm work actually runs before saying it does. The same tool also requires developers to use a label called I_VERIFIED_THIS_IS_NOT_CODE_OR_FILEPATHS whenever they log data, making the compliance confirmation happen at the exact moment of the decision rather than in a separate policy document.
What is the culture trust ladder and which level does Anthropic operate at?
The trust ladder is a five-level model that describes how companies evolve from founder-controlled decisions to self-sustaining culture. Level 1 (Campfire): the founder decides everything. Level 2 (Wild West): autonomy without clarity. Level 3 (Blueprint): decisions mapped to roles. Level 4 (Engine): the system enables independent action and mistakes stay contained. Level 5 (Ecosystem): culture runs without the founder in daily decisions. Based on the code evidence, Anthropic operates at Level 4.
Why did the Anthropic code leak happen twice?
Anthropic optimises heavily for speed and continuous shipping. Internal and public work are closely woven together because of that structure. When the first leak happened, the fix addressed the symptom rather than the structural cause. The Direction lever only pointed toward speed. What was less defined was what internal means in practice at every step of that process. When the same thing surfaces twice, the answer is in the structure, not a patch on top of the existing setup.
What are the five culture levers that determine how fast a team develops?
The five culture levers are Direction, Safety, Trust, Growth, and Learning speed. Direction is whether the team knows what matters and why. Safety is whether people feel able to speak up, flag problems, and try things without fear. Trust is whether the system is designed around people being reliable. Growth is whether the organisation learns faster than it scales. Learning speed is whether hard-won lessons outlast the people who learned them. Anthropic shows clear evidence of all five levers working at a high level.
How do you separate the doing of work from the verification of work?
At early stage, founders typically handle both building and checking quality. That works at small scale. As a team grows, the founder can no longer be the sole judge of whether something is good enough. Making that separation structural means building it into how work moves, not leaving it to individual judgement. Anthropic's answer is a mandatory independent verification pass. Most growing teams still rely on self-reporting, which is where quality gaps emerge as headcount increases.
What is psychological safety and why does it affect team performance?
Psychological safety is the condition under which people feel able to speak up, flag errors, share ideas, and ask for review without fear of judgement or consequence. Research at Harvard by Amy Edmondson showed that teams with high psychological safety reported three times more errors and improved faster as a result. When a system catches errors neutrally, people become less afraid to make them. That is the connection between structural accountability and a team that keeps getting better.
