Google shut down Project Mariner on May 4, 2026, ending a 17-month experiment in building an AI agent that browsed the web like a human. No press conference, no blog post. The landing page now reads: “Thank you for using Project Mariner. It was shut down on May 4th, 2026 and its technology voyaged to other Google products.”

That quiet farewell masks a significant story. Project Mariner was Google DeepMind’s flagship agentic AI product, revealed in December 2024 as part of the company’s push into autonomous agents powered by Gemini 2.0. It could navigate Chrome, fill out forms, search job listings, and book travel across sites like Expedia, all by taking continuous screenshots and using visual recognition to identify buttons, text fields, and links. A later update added the ability to handle up to 10 tasks simultaneously.

Seventeen months later, it’s gone. And it’s not the only one. OpenAI shut down Operator, its own standalone browser agent, on August 31, 2025. Both products followed the same trajectory: launch with impressive demos, struggle in production, get absorbed into broader platforms. The pattern tells you something about where the agent industry landed in 2026.

What Mariner Actually Did

Project Mariner’s architecture was unusual. Rather than reading page data directly through DOM access or API calls, it processed screenshots in real time to understand what appeared on screen. According to Yaabot, the agent used Gemini 2.0’s multimodal capabilities to interpret visual browser data, identifying interactive elements through computer vision and then acting on them.

The approach had a clear advantage: it worked on any website without requiring special integration. There were no APIs to configure, no plugins to install. The agent saw what you saw and interacted with it the same way you would.

The tradeoff was everything else. Visual processing at scale demanded significant compute. According to Digital Trends, the screenshot-based approach was “resource-intensive,” required “significant computing power to process visual data in real time,” and resulted in “slow performance and occasional errors, such as selecting incorrect options.”

Every action involved taking a screenshot, processing it through the multimodal model, identifying the correct element, generating a click or keystroke, taking another screenshot to verify the result, and repeating. A human books a flight in 20 clicks. Project Mariner needed 20 screenshots, 20 inference calls, and 20 verification steps to do the same thing.

The Internal Collapse

The first public signal of trouble came in March 2026, almost two months before the shutdown. Wired’s Maxwell Zeff reported that Google had begun reassigning staffers away from the Project Mariner team. As Android Authority noted, “that’s typically not a sign of a healthy project.”

By the time the landing page updated on May 4, the decision had clearly been made weeks earlier. Google’s language was careful: Mariner’s technology “voyaged to other Google products,” including the Gemini API, Gemini Agent, and AI Mode in Search. The Mariner landing page itself suggests users try Gemini Agent for “complex tasks.”

This is the standard playbook for killing a product without admitting failure. The technology lives on in a more capable product. The standalone experiment gets a dignified retirement.

The Operator Precedent

Google isn’t the first frontier lab to reach this conclusion. OpenAI launched Operator in January 2025 as a standalone browser agent that could navigate websites, fill forms, and complete multi-step tasks. It used a Computer-Using Agent (CUA) model built on GPT-4o to take screenshots and interact with web pages visually.

Operator lasted seven months. By July 2025, OpenAI had integrated its functionality into ChatGPT as “agent mode,” combining Operator’s browser capabilities with deep research, code execution, and file handling. The standalone product shut down on August 31, 2025.

The resulting ChatGPT agent is described by OpenAI as “a unified agentic system” that brings together “Operator’s ability to interact with websites” with code execution, analysis, and document creation. The browser component survived, but only as one tool among many in a system that defaults to code-level execution whenever possible.

Two frontier labs. Two standalone browser agents. Two shutdowns within nine months. The market delivered a verdict.

Why Code-Level Agents Won

The architectural gap between browser agents and code-level agents isn’t subtle. Browser agents process visual data to interact with interfaces designed for humans. Code-level agents work directly with files, APIs, and system calls. The efficiency difference is structural.

According to Yaabot’s analysis, code-level agents like Claude Code and OpenClaw are “35x more token-efficient” than browser agents and “faster, more reliable, and capable of more complex tasks.” The comparison table in their analysis makes the gap concrete: browser agents carry higher compute costs (real-time image processing versus structured data), higher error rates (visual misidentification versus deterministic API calls), and lower task complexity ceilings (form filling versus code and file modification).

The reason is architectural, not incremental. A browser agent trying to fill out a Jira ticket needs to: navigate to the page, screenshot it, identify the title field, click it, type text, screenshot again to verify, find the description field, click it, type more text, find the submit button, click it, screenshot to confirm. A code-level agent calls the Jira API directly: one structured request, one response.

Every layer of visual interpretation adds latency, cost, and error surface. The browser agent does in 12 steps what the code agent does in one. And the browser agent’s 12 steps each carry a nonzero probability of failure, compounding into reliability problems that worsen as task complexity increases.

Digital Trends summarized the shift: “Tools that operate at the file and code level, rather than the visual browser level, have become the dominant model. They are faster, cheaper to run, and more capable of handling complex, multi-step tasks.”

The Remaining Browser Agents

Browser-based interaction hasn’t disappeared entirely. OpenAI kept browser capabilities inside ChatGPT agent mode as one tool alongside code execution. Anthropic ships computer use in Claude Cowork, where Claude can “open apps on your computer, navigate a web browser and fill in spreadsheets”, according to CNBC. Perplexity launched Comet, a browser with built-in agent capabilities.

But none of these position browser interaction as the primary agent architecture. They treat it as a fallback for cases where no API, CLI, or direct integration exists. Claude Cowork uses browser navigation “when there is no direct integration.” ChatGPT agent mode navigates websites when it can’t accomplish the task through code execution or structured API calls.

The hierarchy is clear: API first, code second, browser last.

The Gemini Agent Gamble

Google’s consolidation into Gemini Agent raises a question: can browser-based capabilities survive as a feature inside a larger product, or will they atrophy the way voice assistants stagnated after smartphones absorbed their functionality?

The Verge noted that Google also showed off “auto-browse,” a Chrome feature that can perform multi-step tasks like researching flight costs. Google I/O 2026 starts May 19, and new Gemini Agent capabilities are expected. The browser layer isn’t dead inside Google, but it’s no longer the product. It’s a feature.

The risk is that features without a team tend to stagnate. Project Mariner had a dedicated team at DeepMind working on the hard problems of visual page understanding, error recovery, and multi-step task completion. Distributing that technology across Gemini API, Gemini Agent, and AI Mode means no single team owns the browser agent problem anymore.

What This Means for Agent Architecture

Project Mariner’s shutdown is clarifying for anyone building agent infrastructure. The two architectures that competed in 2024 and 2025, browser-first and code-first, are no longer competing. Code-first won.

The implications are concrete. Teams building agent systems should default to API and CLI integrations over browser automation. Browser interaction should be a fallback for legacy applications that don’t expose APIs, not the primary design pattern. Token budgets should assume code-level efficiency, not visual-processing overhead.

For platforms like OpenClaw that already operate at the code level (modifying files, running shell commands, interacting with APIs through skills), the Mariner shutdown validates the architectural bet. For enterprise teams evaluating agent frameworks, it simplifies the decision: any framework that relies primarily on screen scraping or visual browser automation is building on architecture the market has rejected.

The browser agent era lasted roughly 18 months, from Operator’s January 2025 launch to Mariner’s May 2026 shutdown. Code-level agents, from Claude Code to OpenClaw to OpenAI Codex, now define how autonomous AI systems interact with the world. The visual layer was a bridge. The industry crossed it.