Anthropic's new product, powerful enough to make the AI Agent Infrastructure team unemployed?
Original Title: "Anthropic Today Released a New Product That May Cause a Wave of AI Agent Infrastructure Teams to Lose Their Jobs"
Original Author: Bayu, AI Engineer
This product is called Claude Managed Agents. In a nutshell: you tell Anthropic what kind of AI agent you want, and it helps you run it in the cloud, including all infrastructure, with usage-based pricing. Sentry used it to go live with end-to-end automated bug fixing in a few weeks, while Rakuten deployed a specialized agent in a week. Previously, these tasks would require an entire engineering team working for months.

Meanwhile, Anthropic's annual recurring revenue has just surpassed $30 billion, triple that of December last year. Most of the growth comes from enterprise customers. Wall Street has started to get nervous, with the WSJ stating that investors are becoming increasingly cautious about the stock prices of traditional SaaS companies, fearing that products like Anthropic's could make some traditional software services obsolete.
What exactly is this product? How does it differ from the Claude Code you are already using? How was it achieved technically?
What Is It? How Does It Differ from Claude Code?
If you have used Claude Code, you know how AI agents work: you give them a task, and they autonomously plan steps, use tools, write code, modify files, and complete the task step by step.
Claude Code runs on your own computer and is a command-line tool for personal developer use. It stops running when you shut down your computer.
Managed Agents run on Anthropic's cloud and are an API service for enterprise use. They can run continuously 24/7, retain progress even if disconnected, and your product can directly embed AI agent capabilities.
This is how Notion operates: users assign tasks to Claude agents within Notion, the agents work in the background, complete the tasks, and return the results, all without users having to leave Notion.

Several Typical Use Cases:
· Event-triggered: The system discovers a bug, automatically assigns a bot to fix it and raise a pull request, with no human intervention in between.
· Scheduled: Automatically generates a GitHub activity summary or team work brief every morning.
· Fire-and-forget: Assign a task to a bot in Slack, it completes the task and returns the document, PowerPoint, or app.
· Long-Running Task: Running a deep research or code refactoring task for several hours.
What's the Difference Between Cloud-hosted Bots and In-house Bots?
You could self-host, but it's costly and slow.
An intelligent bot that can go live requires much more than just "calling an API": a sandbox environment (an isolated secure space where AI can run code, modify files, tinker without affecting the real external system, like providing AI with a dedicated virtual machine), credential management, state recovery, permission control, end-to-end tracing, and more.
Many enterprise customers used to need an entire engineering team dedicated to these tasks. Now, it's plug and play, freeing up engineers to focus on the core of the product.
However, the pain points solved by Managed Agents go beyond just saving labor.
Matt Dongslee (@dongxi_nlp) has a succinct summary:

There's a specific example in the Anthropic Engineering Blog:
When Claude Sonnet 4.5 nears the context window limit, it "panics" and hastily ends the task. They added context reset in the scheduling framework to address this. However, with Claude Opus 4.5, this issue disappeared, and the previous patch actually became a burden.
If you build your scheduling framework, you have to update it with every model upgrade. Delegate it to Anthropic; they optimize it for you, essentially optimizing what they sell to you.

Who's Using It? How?
Notion allows users to offload tasks such as coding, creating PPTs, and organizing spreadsheets directly to Claude within the workspace, running dozens of tasks in parallel, with the whole team collaborating on the same output. Notion Product Manager Eric Liu said users can delegate open-ended complex tasks directly without leaving Notion.

Sentry implemented a fully automated process "from bug discovery to code fix submission." Their AI debugging tool Seer, after identifying the root cause, allows Claude to write patches directly, open PRs (pull requests). Engineering Director Indragie Karunaratne said they were able to launch in a few weeks, saving the ongoing maintenance cost of self-built infrastructure.
Atlassian integrated it into Jira, enabling developers to directly assign tasks to Claude AI.
Asana created AI Teammates, adding AI collaborators in project management who can take on tasks and deliverables.
General Legal (legal tech company) has the most interesting approach: their AI can temporarily create tools to search data based on user queries. Previously, each user query had to be anticipated, and a retrieval tool developed in advance, but now the AI generates them on-demand. The CTO said development time has been reduced by 10x.
Rakuten deployed specialized AI agents in engineering, product, sales, marketing, and finance departments, each going live within a week, receiving tasks via Slack and Teams and delivering tangible outputs such as spreadsheets, PPTs, and apps.
Technical Principle: Decoupling the Brain from the Hands
The Anthropic engineering team wrote a tech blog post titled Scaling Managed Agents: Decoupling the brain from the hands, discussing the architectural evolution behind Managed Agents.


Initially, they shoved everything into one container: AI's inference loop, code execution environment, and session log, all together. The benefit was simplicity, but the downside was that all eggs were in one basket—if the container went down, the entire session was lost, and individual parts could not be replaced separately.
Later, they made a key split:
· The "Brain" is Claude and its scheduling framework, responsible for thinking and decision-making.
· The "Hand" is the sandbox and various tools, responsible for executing specific operations.
· The "Memory" is an independent session log, recording everything that happens.
The three are independent, and if one goes down, it does not affect the other two.
This split brought several practical benefits:
Speed
Not every task needs to start the full sandbox environment. Now, the sandbox is only launched on-demand when AI truly needs to run code. The median first response latency decreased by about 60%, and in extreme cases, it dropped by over 90%.
Security
Code generated by AI runs in the sandbox, while credentials to access external systems are stored in a secure vault outside the sandbox, with physical isolation on both sides. For example, to access a Git repository, the system clones the code during initialization, and AI interacts with git push/pull normally, but the Token itself is not visible to AI. For services like Slack and Jira, they are accessed via the MCP protocol, where requests go through a proxy layer, the proxy layer retrieves the credentials from the vault to call the service, and AI never handles the credentials throughout the process.
Flexibility
The Brain doesn't care what the Hand is. There's an interesting phrase in the engineering blog: the scheduling framework doesn't know if the sandbox is a container, a mobile phone, or a Pokémon emulator. It just needs to adhere to the "input a name, get a string out" interface.
This also means that multiple Brains can share the Hand, and one Brain can hand over the Hand to another Brain, laying the foundation for multi-agent collaboration.
Limitations
Managed Agents are not all-powerful. There are several points to note:
Some features are still in the research preview stage. Abilities such as multi-agent collaboration, advanced memory tools, and self-assessment iteration (allowing the agent to judge its own task completion quality and iteratively improve) are not fully open yet and require application for access.
Platform Lock-in. Opting for Managed Agents means your agent infrastructure is tied to the Anthropic ecosystem. If you plan to switch models or platforms in the future, migration costs should not be overlooked.
Context management remains a challenge. While session logs are stored independently, deciding which information to retain or discard during long tasks still involves irreversible decisions. This is an ongoing challenge, and their current approach separates context storage from context management: storage ensures preservation, while management policies adjust with model evolution.
Cost Predictability. $0.08 per session hour may sound reasonable, but for complex tasks requiring the agent to run for several hours, considering token consumption and runtime costs, the overall cost may not be insignificant. Enterprises need to assess their budgets accordingly.
Managed Agents indicate that most enterprises still have a long way to go before they can "fully rely on AI agents for work."
While the infrastructure barrier has been lowered, Managed Agents cannot assist with defining good tasks, designing workflows, or establishing trust to allow AI to access core business data.
The "AWS Moment" of AI Agent Infrastructure
Managed Agents seem to be following the path AWS took in its early days: first providing computing power, then encapsulating the runtime environment.
Ten years ago, enterprises debated whether to "move to the cloud"; now, the debate is whether to "self-host Agent infrastructure or go with managed services." Historical experience tells us that most enterprises eventually choose managed services because infrastructure is never a core competency. OpenAI has also launched its own Agent platform, Frontier, and the competition in this space is just beginning.
From a technological perspective, the "separation of brain and hand" architectural approach is worth noting. It allows each part of the system to evolve independently: upgrade the model, change the brain; need a new tool, add a hand; alter the storage solution, replace the memory layer.
A good analogy from an engineering blog: the read() command of an operating system doesn't care whether it's dealing with a 1970s disk or a modern SSD; the abstraction layer is stable, allowing the underlying implementation to be swapped out easily.
From a usage perspective, if you are an enterprise developer looking to embed AI agent capability in your product, Managed Agents might save you several months of infrastructure work.
Six languages (Python, TypeScript, Java, Go, Ruby, PHP) are supported by SDKs. If you are already using Claude Code, update to the latest version, type /claude-api managed-agents-onboarding to get started.
If you are a casual AI enthusiast, the most immediate impact you might feel is: in the SaaS products you use, more and more AI agents will be working in the background to assist you, with these agents likely running on Managed Agents.
Pricing Reference: Token costs are based on the Anthropic API standard pricing, with a runtime cost of $0.08 per session hour (idle time is not billed) and $10 per thousand web searches.
Do you think that the infrastructure for AI agents will eventually be dominated by a few major players, similar to how cloud computing is today?
You may also like

Prediction Markets Under Bias

Stolen: $290 million, Three Parties Refusing to Acknowledge, Who Should Foot the Bill for the KelpDAO Incident Resolution?

ASTEROID Pumped 10,000x in Three Days, Is Meme Season Back on Ethereum?

ChainCatcher Hong Kong Themed Forum Highlights: Decoding the Growth Engine Under the Integration of Crypto Assets and Smart Economy

Why can this institution still grow by 150% when the scale of leading crypto VCs has shrunk significantly?

Anthropic's $1 trillion, compared to DeepSeek's $100 billion

Geopolitical Risk Persists, Is Bitcoin Becoming a Key Barometer?

Annualized 11.5%, Wall Street Buzzing: Is MicroStrategy's STRC Bitcoin's Savior or Destroyer?

An Obscure Open Source AI Tool Alerted on Kelp DAO's $292 million Bug 12 Days Ago

Mixin has launched USTD-margined perpetual contracts, bringing derivative trading into the chat scene.
The privacy-focused crypto wallet Mixin announced today the launch of its U-based perpetual contract (a derivative priced in USDT). Unlike traditional exchanges, Mixin has taken a new approach by "liberating" derivative trading from isolated matching engines and embedding it into the instant messaging environment.
Users can directly open positions within the app with leverage of up to 200x, while sharing positions, discussing strategies, and copy trading within private communities. Trading, social interaction, and asset management are integrated into the same interface.
Based on its non-custodial architecture, Mixin has eliminated friction from the traditional onboarding process, allowing users to participate in perpetual contract trading without identity verification.
The trading process has been streamlined into five steps:
· Choose the trading asset
· Select long or short
· Input position size and leverage
· Confirm order details
· Confirm and open the position
The interface provides real-time visualization of price, position, and profit and loss (PnL), allowing users to complete trades without switching between multiple modules.
Mixin has directly integrated social features into the derivative trading environment. Users can create private trading communities and interact around real-time positions:
· End-to-end encrypted private groups supporting up to 1024 members
· End-to-end encrypted voice communication
· One-click position sharing
· One-click trade copying
On the execution side, Mixin aggregates liquidity from multiple sources and accesses decentralized protocol and external market liquidity through a unified trading interface.
By combining social interaction with trade execution, Mixin enables users to collaborate, share, and execute trading strategies instantly within the same environment.
Mixin has also introduced a referral incentive system based on trading behavior:
· Users can join with an invite code
· Up to 60% of trading fees as referral rewards
· Incentive mechanism designed for long-term, sustainable earnings
This model aims to drive user-driven network expansion and organic growth.
Mixin's derivative transactions are built on top of its existing self-custody wallet infrastructure, with core features including:
· Separation of transaction account and asset storage
· User full control over assets
· Platform does not custody user funds
· Built-in privacy mechanisms to reduce data exposure
The system aims to strike a balance between transaction efficiency, asset security, and privacy protection.
Against the background of perpetual contracts becoming a mainstream trading tool, Mixin is exploring a different development direction by lowering barriers, enhancing social and privacy attributes.
The platform does not only view transactions as execution actions but positions them as a networked activity: transactions have social attributes, strategies can be shared, and relationships between individuals also become part of the financial system.
Mixin's design is based on a user-initiated, user-controlled model. The platform neither custodies assets nor executes transactions on behalf of users.
This model aligns with a statement issued by the U.S. Securities and Exchange Commission (SEC) on April 13, 2026, titled "Staff Statement on Whether Partial User Interface Used in Preparing Cryptocurrency Securities Transactions May Require Broker-Dealer Registration."
The statement indicates that, under the premise where transactions are entirely initiated and controlled by users, non-custodial service providers that offer neutral interfaces may not need to register as broker-dealers or exchanges.
Mixin is a decentralized, self-custodial privacy wallet designed to provide secure and efficient digital asset management services.
Its core capabilities include:
· Aggregation: integrating multi-chain assets and routing between different transaction paths to simplify user operations
· High liquidity access: connecting to various liquidity sources, including decentralized protocols and external markets
· Decentralization: achieving full user control over assets without relying on custodial intermediaries
· Privacy protection: safeguarding assets and data through MPC, CryptoNote, and end-to-end encrypted communication
Mixin has been in operation for over 8 years, supporting over 40 blockchains and more than 10,000 assets, with a global user base exceeding 10 million and an on-chain self-custodied asset scale of over $1 billion.

$600 million stolen in 20 days, ushering in the era of AI hackers in the crypto world

Vitalik's 2026 Hong Kong Web3 Summit Speech: Ethereum's Ultimate Vision as the "World Computer" and Future Roadmap

On the same day Aave introduced rsETH, why did Spark decide to exit?

Full Post-Mortem of the KelpDAO Incident: Why Did Aave, Which Was Not Compromised, End Up in Crisis Situation?

After a $290 million DeFi liquidation, is the security promise still there?

ZachXBT's post ignites RAVE nearing zero, what is the truth behind the insider control?

Vitalik 2026 Hong Kong Web3 Carnival Speech Transcript: We do not compete on speed; security and decentralization are the core






