1
1
When the creator of the world’s most advanced coding agent speaks, Silicon Valley doesn’t just listen — it takes notes, recognizing a potential paradigm shift in how software is built. For the past week, the global engineering community has been meticulously dissecting a viral thread on X from Boris Cherny, the visionary creator and head of Claude Code at Anthropic. What began as a seemingly casual sharing of his personal terminal setup has rapidly escalated into a powerful manifesto on the future of software development, with industry insiders hailing it as a watershed moment for the burgeoning AI startup.
The impact has been immediate and profound. "If you’re not reading the Claude Code best practices straight from its creator, you’re behind as a programmer," declared Jeff Tang, a prominent and influential voice within the developer community, underscoring the urgency of adopting these new methodologies. Kyle McNease, another keen industry observer, went even further, proclaiming that with Cherny’s "game-changing updates" and innovative approach, Anthropic is "on fire," potentially poised to experience "their ChatGPT moment" – a reference to the explosive impact OpenAI’s flagship product had on public consciousness and technological adoption.
This wave of excitement stems from a striking paradox inherent in Cherny’s methodology: his workflow is surprisingly elegant and simple in its core components, yet it empowers a single human developer to operate with the staggering output capacity typically associated with a small, dedicated engineering department. This exponential leap in productivity has resonated deeply. As one user aptly noted on X after implementing Cherny’s described setup, the experience "feels more like Starcraft" than traditional coding. This analogy vividly captures the shift from the laborious, line-by-line typing of syntax to a strategic role, akin to commanding an array of autonomous units to execute complex tasks in parallel. This isn’t just about faster typing; it’s about a fundamentally different way of interacting with the development process itself.
Cherny’s revelations offer a rare, direct glimpse into the mind of an architect who is not just using AI, but is redefining the human-AI collaboration in software engineering. His workflow represents a blueprint for a future where developers are no longer merely coders, but rather orchestrators of intelligent, highly efficient digital workforces. Here is a detailed analysis of the workflow that is currently reshaping how software gets built, straight from the architect himself.
How Running Five AI Agents at Once Transforms Coding into a Real-Time Strategy Game
Perhaps the most striking and immediately impactful revelation from Cherny’s candid disclosure is his radical departure from traditional linear coding practices. In the conventional "inner loop" of development, a programmer typically focuses on a single task: writing a function, meticulously testing it, debugging, and only then moving sequentially to the next discrete piece of work. Cherny, however, transcends this linear approach, instead adopting the role of a fleet commander, overseeing multiple simultaneous operations.
"I run 5 Claudes in parallel in my terminal," Cherny wrote, illustrating his core strategy. "I number my tabs 1-5, and use system notifications to know when a Claude needs input." This seemingly simple setup, powered by advanced terminal emulators like iTerm2, allows him to effectively manage five distinct work streams concurrently. While one AI agent might be diligently running a comprehensive test suite to validate recent changes, another could be actively refactoring a legacy module to improve its efficiency or maintainability. Simultaneously, a third Claude might be drafting critical technical documentation, while others tackle bug fixes or explore new feature implementations. This parallel processing dramatically reduces idle time and maximizes throughput.
Beyond his local terminal, Cherny further amplifies his capacity by running "5-10 Claudes on claude.ai" in his browser. He employs a sophisticated "teleport" command, enabling seamless handoffs of sessions and context between his web-based agents and his local machine. This integrated approach allows for incredible flexibility and scale. This methodology directly validates the "do more with less" strategy championed by Anthropic President Daniela Amodei earlier this week. While competitors like OpenAI pursue massive, trillion-dollar infrastructure build-outs, Anthropic is proving that superior orchestration and innovative human-AI workflows with existing, powerful models can yield exponential productivity gains, challenging the conventional wisdom of raw compute power.
The Counterintuitive Case for Choosing the Slowest, Smartest Model
In an industry perpetually obsessed with optimizing for speed and minimizing latency, Cherny’s choice of AI model is profoundly counterintuitive yet deeply insightful. He revealed that he exclusively uses Anthropic’s heaviest, most sophisticated, and consequently, slowest model: Opus 4.5.
"I use Opus 4.5 with thinking for everything," Cherny explained. "It’s the best coding model I’ve ever used, and even though it’s bigger & slower than Sonnet, since you have to steer it less and it’s better at tool use, it is almost always faster than using a smaller model in the end." This statement encapsulates a critical shift in understanding AI efficiency. The perceived slowness of Opus 4.5 is a trade-off for its superior reasoning capabilities and reduced need for human intervention. A less capable model might generate tokens faster, but if those tokens frequently require correction, debugging, or extensive human refinement, the overall development cycle lengthens considerably.
For enterprise technology leaders and engineering managers, this presents a crucial insight: the true bottleneck in modern AI-assisted development isn’t merely the generation speed of an individual token. Instead, it’s the invaluable human time spent identifying, understanding, and correcting the AI’s mistakes. Cherny’s workflow brilliantly illustrates that paying the "compute tax" for a smarter, more capable model upfront ultimately eliminates or significantly reduces the far more costly and time-consuming "correction tax" later in the development process. This approach prioritizes intelligence and autonomy over raw speed, leading to a net gain in efficiency and quality.
One Shared File Transforms Every AI Mistake into a Permanent Lesson
A significant challenge with standard large language models (LLMs) is their inherent "amnesia." They typically do not "remember" specific company coding styles, architectural decisions, or preferred patterns from one interaction or session to the next. This lack of persistent institutional memory can lead to repetitive errors and a constant need for human re-instruction.

To ingeniously address this persistent problem, Cherny’s team maintains a single, highly critical file named CLAUDE.md, strategically located in their git repository. "Anytime we see Claude do something incorrectly we add it to the CLAUDE.md, so Claude knows not to do it next time," he wrote. This simple yet powerful practice transforms the codebase itself into a dynamic, self-correcting organism.
The mechanism is elegant: when a human developer reviews a pull request and identifies an error or a deviation from established best practices, they don’t just fix the code. Crucially, they also tag the AI, updating its internal instructions within CLAUDE.md to prevent that specific mistake from recurring. This creates a continuous feedback loop. As Aakash Gupta, a prominent product leader analyzing the thread, insightfully noted, "Every mistake becomes a rule." The longer the team collaborates using this system, the more intelligent, context-aware, and aligned with the team’s specific requirements the AI agent becomes, fostering an environment of perpetual learning and improvement.
Slash Commands and Subagents Automate the Most Tedious Parts of Development
The "vanilla" workflow that garnered significant praise for its efficiency and elegance is not merely about running multiple agents; it’s profoundly empowered by rigorous automation of repetitive and often tedious tasks. Cherny extensively leverages slash commands – custom shortcuts that are themselves checked into the project’s repository. These commands allow him to initiate complex operations with a single, streamlined keystroke, significantly reducing cognitive load and manual effort.
He highlighted one particularly impactful command: /commit-push-pr. Cherny reveals he invokes this command dozens of times daily. Instead of the laborious manual process of typing out a series of git commands, crafting a detailed commit message, and then navigating to a web interface to open a pull request, this single slash command orchestrates the entire sequence autonomously. This effectively offloads the "bureaucracy of version control" to the AI agent, freeing the developer to focus on higher-level problem-solving and creative work.
Furthermore, Cherny strategically deploys subagents – specialized AI personas – each meticulously designed to handle specific phases of the development lifecycle. For instance, he utilizes a dedicated "code-simplifier" subagent to meticulously clean up and optimize architectural patterns after the core development work is completed, ensuring maintainability and adherence to best practices. Another critical subagent, the "verify-app" agent, is tasked with running comprehensive end-to-end tests, providing a crucial layer of quality assurance before any code is shipped to production. These specialized agents act as an intelligent, automated development pipeline, dramatically enhancing both speed and reliability.
Why Verification Loops Are the Real Unlock for AI-Generated Code
If there is one singular reason why Claude Code has reportedly achieved the significant milestone of $1 billion in annual recurring revenue so rapidly, it is likely the sophisticated implementation of verification loops. This capability fundamentally transforms the AI from a mere text generator into a robust, self-validating testing and development partner.
"Claude tests every single change I land to claude.ai/code using the Claude Chrome extension," Cherny detailed. "It opens a browser, tests the UI, and iterates until the code works and the UX feels good." This comprehensive approach means the AI isn’t just generating code; it’s actively engaging in the critical processes of testing, validation, and iterative refinement.
Cherny argues passionately that equipping the AI with the ability to verify its own work – whether through automated browser interactions, executing bash commands, or running extensive test suites – improves the quality of the final result by an astounding "2-3x." This capability is the true "unlock" for AI-generated code, as it moves beyond simply producing output to ensuring the output is functional, reliable, and meets desired specifications. The agent doesn’t just write code; it proactively proves the code works, drastically reducing the burden on human QA and increasing developer confidence in AI-assisted solutions.
What Cherny’s Workflow Signals About the Future of Software Engineering
The overwhelming and enthusiastic reaction to Cherny’s thread signifies a pivotal and irreversible shift in how developers conceptualize their craft. For many years, the concept of "AI coding" primarily equated to advanced autocomplete functions within a text editor – essentially, a faster way to type. However, Cherny’s groundbreaking demonstration illustrates that AI can now function as an intricate operating system for labor itself, transforming the very nature of software development.
As Jeff Tang succinctly summarized on X, "Read this if you’re already an engineer… and want more power." This sentiment perfectly captures the essence of Cherny’s contribution: he’s not just offering tips, but a blueprint for amplifying human capabilities to an unprecedented degree.
The tools and methodologies to multiply human output by a factor of five or more are not a distant dream; they are already here and actively being implemented. Their effective deployment, however, requires more than just technical proficiency. It demands a fundamental shift in mindset – a willingness to stop thinking of AI merely as an assistant that helps with tasks and to start treating it as a powerful, autonomous workforce capable of executing complex development cycles. The programmers who embrace this mental leap first won’t just experience incremental gains in productivity. They will be playing an entirely different game, operating at a strategic, orchestrating level, while everyone else remains stuck in the traditional, character-by-character typing paradigm. This is not just an evolution of coding; it’s a revolution in how software is engineered.