1
1
1
2
3
The artificial intelligence coding revolution, while promising unprecedented efficiency, has introduced a significant hurdle: its often prohibitive cost. Anthropic’s Claude Code, a sophisticated terminal-based AI agent designed to write, debug, and deploy code autonomously, has undoubtedly captured the imagination of software developers globally. However, its tiered pricing structure—ranging from $20 to $200 per month depending on usage—has ignited a growing backlash among the very programmers it aims to empower, pushing many to seek more accessible solutions.
In response to this rising frustration, a compelling free alternative has rapidly gained traction. Goose, an open-source AI agent developed by Block, the financial technology company formerly known as Square, offers functionality nearly identical to Claude Code. Crucially, Goose operates entirely on a user’s local machine, eliminating subscription fees, cloud dependencies, and restrictive rate limits that typically reset every few hours. This fundamental difference in architecture underpins its core appeal: complete control and unparalleled privacy.
"Your data stays with you, period," affirmed Parth Sareen, a software engineer, during a recent livestream demonstration of Goose. This statement encapsulates the tool’s primary draw, highlighting its commitment to data sovereignty. Developers gain full command over their AI-powered workflow, including the invaluable ability to work offline, a feature particularly attractive for those who often code on the go, even from an airplane. The project’s explosive popularity underscores this demand, boasting over 26,100 stars on GitHub, the premier code-sharing platform, with 362 contributors and 102 releases since its launch. The rapid development pace, evidenced by the latest version 1.20.1 shipping on January 19, 2026, rivals that of many commercial products, showcasing a vibrant and dedicated community. For developers increasingly disillusioned by Claude Code’s escalating pricing and stringent usage caps, Goose represents a refreshing and increasingly rare offering in the AI industry: a genuinely free, no-strings-attached option for serious, professional work.
To fully grasp the significance of Goose, one must understand the unfolding controversy surrounding Claude Code’s pricing and usage policies. Anthropic, the San Francisco-based artificial intelligence company founded by former OpenAI executives, integrates Claude Code into its subscription tiers. The free plan, paradoxically, offers no access to Claude Code whatsoever. The Pro plan, priced at $17 per month with annual billing or $20 monthly, imposes severe limitations, granting users access to just 10 to 40 prompts every five hours—a constraint that experienced developers can exhaust within mere minutes of intensive coding. The Max plans, at $100 and $200 per month, provide more generous headroom, offering 50 to 200 prompts and 200 to 800 prompts respectively, alongside access to Anthropic’s most powerful model, Claude 4.5 Opus. Yet, even these premium tiers are not immune to restrictions, which have significantly inflamed the developer community.
In late July, Anthropic introduced new weekly rate limits, further complicating the value proposition. Under this revised system, Pro users are allocated 40 to 80 hours of Sonnet 4 usage per week. Max users on the $200 tier receive 240 to 480 hours of Sonnet 4, complemented by 24 to 40 hours of Opus 4. Nearly five months later, the widespread frustration among developers remains unabated. The core issue lies in the ambiguity of these "hours," which do not represent actual time but rather token-based limits that fluctuate wildly based on codebase size, conversation length, and the complexity of the code being processed. Independent analyses suggest that the practical per-session limits translate to approximately 44,000 tokens for Pro users and around 220,000 tokens for the $200 Max plan. "It’s confusing and vague," one developer wrote in a widely circulated analysis. "When they say ’24-40 hours of Opus 4,’ that doesn’t really tell you anything useful about what you’re actually getting."
The backlash across platforms like Reddit and various developer forums has been vociferous. Numerous users have reported hitting their daily limits within 30 minutes of intensive coding sessions. Others have resorted to canceling their subscriptions entirely, lambasting the new restrictions as "a joke" and "unusable for real work." Anthropic has publicly defended these changes, asserting that the limits affect fewer than five percent of users and are specifically aimed at those running Claude Code "continuously in the background, 24/7." However, the company has yet to clarify whether this figure refers to five percent of its Max subscribers or five percent of its entire user base—a distinction that carries immense implications for understanding the true impact of the policy.
Goose, by contrast, adopts a radically different philosophy to address the same challenges. Built by Block, the payments company helmed by co-founder Jack Dorsey, Goose is what engineers refer to as an "on-machine AI agent." Unlike Claude Code, which transmits user queries to Anthropic’s remote servers for processing, Goose is designed to run entirely on a user’s local computer. This is achieved by leveraging open-source language models that users can download and manage themselves, ensuring that all processing occurs client-side. The project’s official documentation succinctly describes its capabilities as going "beyond code suggestions" to "install, execute, edit, and test with any LLM." That final phrase—"any LLM"—serves as Goose’s defining differentiator. It is designed to be model-agnostic, providing unparalleled flexibility.
Developers can configure Goose to connect to Anthropic’s Claude models via API access, integrate with OpenAI’s GPT-5, Google’s Gemini, or route requests through services like Groq or OpenRouter. However, the most compelling aspect, and where Goose truly shines, is its ability to operate entirely locally using tools such as Ollama. Ollama simplifies the process of downloading and executing open-source models directly on a user’s own hardware. The practical implications of this local setup are profound: it eliminates all subscription fees, usage caps, and rate limits. Crucially, it eradicates any concerns about proprietary code or sensitive data being transmitted to external servers, as all conversations with the AI remain securely on the user’s machine. "I use Ollama all the time on planes—it’s a lot of fun!" Sareen noted during his demonstration, underscoring how local models liberate developers from the constraints of internet connectivity and cloud reliance.
Beyond simple code suggestions, Goose operates as either a command-line tool or a desktop application, capable of autonomously executing complex development tasks. It can initiate entire projects from scratch, write and execute code, debug failures, orchestrate workflows across multiple files, and seamlessly interact with external APIs—all without requiring constant human oversight. This advanced capability relies heavily on what the AI industry terms "tool calling" or "function calling," which is the language model’s ability to request and trigger specific actions from external systems. When a developer instructs Goose to create a new file, run a test suite, or check the status of a GitHub pull request, the agent doesn’t merely generate text describing the desired outcome; it actively executes those operations within the local environment.
The effectiveness of this functionality is largely dependent on the underlying language model. According to the Berkeley Function-Calling Leaderboard, which ranks models on their proficiency in translating natural language requests into executable code and system commands, Claude 4 models from Anthropic currently demonstrate superior performance in tool calling. Nevertheless, newer open-source models are rapidly closing this gap. Goose’s documentation highlights several robust options with strong tool-calling support, including Meta’s Llama series, Alibaba’s Qwen models, Google’s Gemma variants, and DeepSeek’s reasoning-focused architectures. Furthermore, the tool integrates with the Model Context Protocol (MCP), an emerging standard designed to connect AI agents to external services. Through MCP, Goose can access databases, search engines, file systems, and a myriad of third-party APIs, significantly extending its capabilities beyond the base language model.
For developers seeking a completely free, privacy-preserving setup, the process of configuring Goose with a local model involves three primary components: Goose itself, Ollama (the tool for running open-source models locally), and a compatible open-source language model.
Step 1: Install Ollama. Ollama is an open-source project that dramatically simplifies the complex task of running large language models on personal hardware. It handles the intricate processes of downloading, optimizing, and serving models through a user-friendly interface. To begin, download and install Ollama directly from ollama.com. Once installed, models can be pulled with a single command. For coding-centric tasks, Qwen 2.5 is a highly recommended option due to its strong tool-calling support: ollama run qwen2.5. The model will download automatically and commence running on your machine.
Step 2: Install Goose. Goose is available as both a desktop application, offering a more visual experience, and a command-line interface (CLI), which appeals to developers who prefer working exclusively in the terminal. Installation instructions vary by operating system but typically involve downloading pre-built binaries from Goose’s GitHub releases page or utilizing a package manager. Block provides binaries for macOS (both Intel and Apple Silicon), Windows, and Linux.
Step 3: Configure the Connection. In Goose Desktop, navigate to Settings, then Configure Provider, and select Ollama. Confirm that the API Host is set to http://localhost:11434, which is Ollama’s default port, and click Submit. For the command-line version, execute goose configure, select "Configure Providers," choose Ollama, and enter the model name when prompted. With these steps completed, Goose is successfully connected to a language model running entirely on your hardware, ready to execute complex coding tasks without incurring any subscription fees or relying on external cloud services.
An inevitable question arises: what kind of computer is necessary to run these systems? Running large language models locally demands significantly more computational resources than typical software applications. The primary constraint is memory, specifically RAM on most systems, or VRAM if leveraging a dedicated graphics card for acceleration. Block’s documentation suggests that 32 gigabytes of RAM provides "a solid baseline for larger models and outputs." For Mac users, this refers to the computer’s unified memory, which acts as the main bottleneck. For Windows and Linux users with discrete NVIDIA graphics cards, GPU memory (VRAM) becomes more critical for accelerating model inference. However, expensive, high-end hardware isn’t strictly a prerequisite for getting started. Smaller models with fewer parameters can operate effectively on more modest systems. For instance, Qwen 2.5 is available in multiple sizes, and its smaller variants can perform well on machines equipped with 16 gigabytes of RAM. "You don’t need to run the largest models to get excellent results," Sareen emphasized, advocating for starting with smaller models to test workflows before scaling up as needed. To provide context, an entry-level Apple MacBook Air with 8 gigabytes of RAM would likely struggle with most capable coding models. Conversely, a MacBook Pro with 32 gigabytes of unified memory—an increasingly common configuration among professional developers—can handle these models comfortably.
While Goose with a local LLM offers compelling advantages, it is important to acknowledge that it is not a perfect, drop-in substitute for Claude Code. The comparison involves genuine trade-offs that developers must consider.
Model Quality: Claude 4.5 Opus, Anthropic’s flagship model, arguably remains the most capable AI for sophisticated software engineering tasks. It demonstrates exceptional prowess in comprehending complex codebases, adhering to nuanced instructions, and generating high-quality code with minimal iteration. While open-source models have made dramatic advancements, a performance gap persists, particularly for the most challenging and abstract coding problems. One developer who transitioned to the $200 Claude Code plan described the difference succinctly: "When I say ‘make this look modern,’ Opus knows what I mean. Other models give me Bootstrap circa 2015."
Context Window: Claude Sonnet 4.5, accessible via API, boasts a massive one-million-token context window. This capacity is sufficient to load entire large codebases without the need for complex chunking or context management strategies. Most local models are typically limited to 4,096 or 8,192 tokens by default, though many can be configured for longer contexts at the expense of increased memory usage and slower processing speeds.
Speed: Cloud-based services like Claude Code operate on dedicated server hardware specifically optimized for AI inference, resulting in faster processing times. Local models, running on consumer laptops, inherently process requests more slowly. This speed difference can be a critical factor for iterative development workflows that demand rapid changes and immediate AI feedback.
Tooling Maturity: Claude Code benefits from Anthropic’s substantial engineering resources, leading to highly polished and well-documented features such such as prompt caching (which can reduce costs by up to 90 percent for repeated contexts) and structured outputs. Goose, despite its active development with 102 releases to date, relies on community contributions and may exhibit less refinement in specific areas compared to its commercially backed counterparts.
Goose enters an already crowded market of AI coding tools, yet it carves out a distinctive and compelling niche. Cursor, a popular AI-enhanced code editor, charges $20 per month for its Pro tier and $200 for Ultra, mirroring Claude Code’s Max plans. However, Cursor’s Ultra tier provides approximately 4,500 Sonnet 4 requests per month, representing a substantially different allocation model compared to Claude Code’s hourly resets. Other open-source projects, such as Cline and Roo Code, offer AI coding assistance but with varying levels of autonomy and tool integration. Many of these projects primarily focus on code completion rather than the more agentic task execution capabilities that define both Goose and Claude Code. Enterprise-grade offerings like Amazon’s CodeWhisperer and GitHub Copilot, along with solutions from major cloud providers, target large organizations with complex procurement processes and dedicated budgets, making them less relevant to individual developers and small teams seeking lightweight, flexible tools. Goose’s unique combination of genuine autonomy, model agnosticism, local operation, and zero cost creates an exceptionally strong value proposition. The tool is not attempting to outcompete commercial offerings on sheer polish or raw model quality; instead, it competes fiercely on the principle of freedom—both financial and architectural.
The AI coding tools market is experiencing rapid evolution. Open-source models are improving at an astonishing pace, continually narrowing the performance gap with proprietary alternatives. Notably, Moonshot AI’s Kimi K2 and z.ai’s GLM 4.5 now benchmark near the levels of Claude Sonnet 4—and they are freely available. If this trajectory continues, the quality advantage that currently justifies Claude Code’s premium pricing may eventually erode. Should this occur, Anthropic would face increasing pressure to compete on features, user experience, and seamless integration rather than solely on raw model capability.
For the present, developers are confronted with a clear choice. Those who demand the absolute best model quality, who can comfortably afford premium pricing, and who are willing to accept usage restrictions may continue to prefer Claude Code. However, for a growing segment of developers who prioritize cost-effectiveness, data privacy, offline access, and architectural flexibility, Goose presents a compelling and genuine alternative. The very existence of a zero-dollar open-source competitor offering comparable core functionality to a $200-per-month commercial product is remarkable in itself. This phenomenon reflects both the rapid maturation of open-source AI infrastructure and a burgeoning appetite among developers for tools that genuinely respect their autonomy and control.
Goose is not without its imperfections. It demands a more technical setup process than commercial alternatives, and its effective operation relies on hardware resources that not every developer possesses. Furthermore, while its model options are improving rapidly, they may still trail the best proprietary offerings when tackling the most complex and nuanced tasks. Yet, for an expanding community of developers, these limitations are considered acceptable trade-offs for something increasingly rare in the contemporary AI landscape: a tool that truly belongs to them, fostering an environment of innovation without prohibitive barriers.
Goose is available for download at github.com/block/goose. Ollama is available at ollama.com. Both projects are free and open source.