AI coding platform built into developers’ workflow
Build and automate custom workflows without the usual friction
Includes basic inline completions and chat, with access to the Grok Code Fast Model in the VS Code experience. This is enough to test the workflow, but not enough to judge the full product if you care about top models or larger context windows. Unlocks frontier and open-source models such as Claude Opus-4.6, GPT-5.2, Gemini-3, Grok-4, Llama, and Mistral, plus extended context. For many individual developers, this looks like the real starting point rather than the free tier. Positioned for AI engineering teams with broader shared usage and expanded capabilities. If multiple teammates are actively using multi-agent workflows, this is likely where actual spending starts to make sense. Adds priority support and higher-end access. This tier is for heavier users who want the best response times and fewer limits. Includes volume discounts for 10+ seats, on-prem deployment, advanced security controls, custom SLAs, and training opt-out by default. Enterprise buyers should expect the real cost conversation to center on security, deployment model, and support requirements, not just seat price. The main pricing story is that BLACKBOX AI is cheap to begin with compared with many AI coding products. That said, our research also surfaced complaints about billing and cancellation, so teams should keep an eye on account management and procurement flow before rolling it out widely. If you only test the free plan, you will not see the full value, because many of the headline model choices and context benefits sit behind paid tiers.
Free
$0
Pro
$10/month
Pro Plus
$20/month
Pro Max
$40/month
Enterprise
Custom pricing
Install on macOS via Homebrew: `brew install --cask block-goose` (desktop) or `brew install block-goose-cli` (CLI) Linux packages available as DEB, RPM, and Flatpak. Windows supports direct download and WSL. Configure your preferred LLM provider with an API key or connect an existing Claude, ChatGPT, or Gemini subscription. License: Apache 2.0 (fully open source) GitHub: github.com/block/goose Documentation: goose-docs.ai
| Feature | BLACKBOX AI | Goose (Block) |
|---|---|---|
| Pricing | Free | Free |
| Access to 300+ models and major frontier providers | The platform supports Claude, GPT, Gemini, Grok, Llama, Mistral, DeepSeek, and BLACKBOX’s own models across plans and surfaces. This gives teams flexibility when one model is better at reasoning, another is faster for autocomplete, and another is cheaper for high-volume work. | Run Goose as a native desktop app (macOS, Linux, Windows), a command-line tool, or an embedded API in your own applications |
| VS Code extension with large adoption | The VS Code extension has passed 4.2 million installs and brings inline completions, chat edits, and multi-agent execution into an editor many developers already use daily. Adoption at that scale suggests the product is not asking users to abandon their setup just to try the tool. | Enable or disable extensions during a conversation without restarting, and Goose auto-detects which extensions a task needs |
| Support for 35+ IDEs and desktop environments | BLACKBOX AI integrates with more than 35 development environments, including VS Code, PyCharm, IntelliJ, Android Studio, and Xcode. That breadth matters for teams with mixed stacks, where one AI tool often fails because it only fits one editor culture. | Switch between Anthropic, OpenAI, Google, Ollama, OpenRouter, Azure, Bedrock, and more without changing your workflow |
| Security and enterprise controls | Communication uses TLS 1.3, and enterprise plans include end-to-end encryption, zero-knowledge architecture, on-premise deployment, and file exclusion controls. For teams working with sensitive IP or regulated environments, those controls are often the difference between "interesting demo" and "approved tool." | Set boundaries for what Goose can do on its own, from fully autonomous operation to manual approval for each action |
| Multi-agent coding | BLACKBOX AI can run the same task through multiple agents and models in parallel, then present the outputs as selectable diffs. In practice, this means a developer can compare different implementations of a payment flow or refactor instead of accepting one AI answer blindly, which is a meaningful difference from single-model assistants. | — |
| Specialized development agents | BLACKBOX AI lists agents for refactoring, migration, test generation, deployment, code review, documentation, security analysis, performance optimization, scaffolding, language translation, rollback management, lint fixes, canary deployment, and schema management. That specialization matters because users are not just asking a general chatbot to "help with code," they are invoking workflows tuned for specific parts of the software lifecycle. | — |
| CLI for natural language project generation | The command-line interface lets developers describe a project in plain English and generate a working codebase with dependencies and structure. For developers who live in the terminal, this keeps the workflow inside familiar tools while reducing setup time on greenfield projects. | — |
| AI-native IDE and visual app building | BLACKBOX AI’s own IDE and Builder product can generate full-stack apps from prompts, including frontend, backend, database, and deployment-ready structure. This is especially useful for teams that want to move from idea to a working prototype quickly, or for non-engineers using Builder to create internal tools and product mockups. | — |
| Code extraction from videos and images | BLACKBOX AI can pull usable code from tutorial videos and screenshots. This sounds niche until you remember how much developer learning still happens through YouTube and conference clips, where copying code manually is slow and error-prone. | — |
| OpenAI-compatible API | The API is designed so existing OpenAI SDK integrations can work by changing the base URL. That reduces migration effort for teams already building internal AI workflows and lowers the switching cost compared with providers that require a full rewrite. | — |
| 70+ MCP Extensions | — | Connect to external tools and services through the Model Context Protocol open standard, with built-in extensions for development, web scraping, memory, and data visualization |
| Bring Your Own Subscription | — | Use existing Claude, ChatGPT, or Gemini subscriptions through ACP (Authentication and Connection Protocol) instead of managing separate API keys |
| Session Memory | — | The built-in memory extension lets Goose retain your preferences and context across conversations |
| Built-in Security Scanning | — | Extensions are automatically scanned for malware before activation |
BLACKBOX AI is an AI coding platform built to sit inside the way developers already work, not beside it. Founded in 2020 and headquartered in San Francisco, the company has grown fast without outside funding, reaching more than 12 million total users, roughly 10 million monthly active users, and an estimated $31.7 million in annual revenue with about 180 employees. We found that its identity is broader than "code autocomplete." BLACKBOX AI positions itself as software that builds software, with an ecosystem that spans a native IDE, VS Code extension, desktop app, CLI, browser tools, API, Slack integration, and a no-code Builder product. What makes the product interesting is the architecture behind it. Instead of tying users to one model, BLACKBOX AI orchestrates more than 300 AI models and surfaces access to Claude, GPT, Gemini, Llama, Mistral, Grok, and its own models depending on plan and context. That matters because coding work is uneven. One task needs fast inline suggestions, another needs careful reasoning across a codebase, another needs a second opinion. BLACKBOX AI leans into that reality with a multi-agent system that can send the same task to several models at once and let developers compare the results. The company’s pitch is speed, but the product story is really about control. Developers can use it for a single completion, a refactor, a migration, a test suite, a deployment workflow, or a whole app generated from a natural language prompt. Enterprises can run it with on-premise deployment and zero-knowledge security controls, while individuals can start free and upgrade cheaply. That range helps explain why BLACKBOX AI has shown up in both solo developer workflows and large-company environments, including reported use by Meta, Google, IBM, and Salesforce.
Goose is an open source AI agent by Block that runs locally on your machine as a desktop app, CLI, or embeddable API. It goes beyond code suggestions to handle research, writing, automation, data analysis, and general problem-solving through natural language instructions. Built in Rust for speed and portability, Goose works with 15+ LLM providers and connects to 70+ extensions through the Model Context Protocol (MCP) standard. The project recently moved to the Agentic AI Foundation at the Linux Foundation, and the GitHub repository has over 41,000 stars.