What it means for an AI to take action

Dr Claude DelormeHead of Research, moccet

The leap from a system that produces text to a system that produces effects is the central engineering challenge of this generation of AI. Reliability is a different property from intelligence, and most agent products have underestimated the gap.

An AI that takes action produces effects in the world rather than text in a chat. The leap from a system that produces text to a system that produces effects is the central engineering challenge of this generation of AI. The leap requires reliability, not just intelligence, and four architectural commitments. Anthropic published research in February 2026, titled Measuring AI agent autonomy in practice, finding that production agents pause for clarification more than twice as often as users interrupt them on complex tasks. moccet is being built around the same discipline.

This essay explains what taking action actually requires, why most products that claim it have not built it, and how to tell the difference.

Why is taking action different from answering questions?

A system that produces text needs to be smart. The output is the product, so the output must be good. Engineering effort is concentrated in the underlying model and in the prompts that elicit good behaviour. If the text is wrong, the user reads it, recognises that it is wrong, and either corrects it or discards it. The system does not have to be reliable in the strong sense. The user supplies the reliability by serving as a check on the output before anything irreversible happens.

A system that produces effects needs to be smart and reliable. Reliability is a different property from intelligence. A brilliant agent that succeeds 80 percent of the time and fails catastrophically 20 percent of the time is worse than a less brilliant agent that succeeds 95 percent of the time and fails gracefully when it does. Catastrophic failures are not bad text the user can ignore. Catastrophic failures are emails sent to the wrong people, meetings booked at the wrong times, payments dispatched to the wrong accounts.

The gap between intelligence and reliability is what most agent products underestimate. Improving the underlying language model raises capability. Improvement does not, on its own, raise reliability. Reliability requires sandboxing, confirmation steps, observability, idempotent operations, rollback paths, and the discipline to refuse to act when uncertain. None of these is a feature of the model. All of them are properties of the architecture around the model. A team that has built a smarter model has not, by virtue of building a smarter model, built a more reliable agent. They have built a smarter unreliable one, which is in some respects more dangerous.

What does the research show about reliable AI agents?

Anthropic's February 2026 research paper Measuring AI agent autonomy in practice drew on the company's data about how Claude was being used as an agent across customer deployments. The findings were specific and modest.

Most agent actions on Anthropic's API were low-risk and reversible. Software engineering accounted for nearly half of all agentic activity. Agent-initiated stops, where the agent paused to ask the user a clarifying question rather than forging ahead, occurred more than twice as often as human-initiated interruptions on the most complex tasks. The conclusion the researchers drew was that effective oversight of agents in production looks less like a system that confidently completes tasks and more like a system that frequently knows when not to.

The last finding contains the engineering insight the marketing has obscured. Acting is not what the AI industry has spent the past two years getting good at. Refusing to act, intelligently, is.

The discipline that produces reliability has a recognisable shape, and the shape is the orchestrator-worker pattern that has emerged as the consensus architecture for production multi-agent systems. A central orchestrator agent receives a goal, decomposes it into subtasks, routes each subtask to a specialised worker, evaluates the results, and decides what to do next. Microsoft's Azure architecture documentation, Anthropic's research papers on multi-agent design, and the open-source frameworks consolidated into the Agentic AI Foundation in December 2025 have all converged on variants of this pattern. A fuller account is in the essay on the orchestrator-agents pattern.

What are the four architectural commitments of an action system?

A reliable action system requires four architectural commitments that have to be made from the start. The commitments cannot be added on top of a chatbot.

The first commitment is the orchestrator-worker pattern. The orchestrator decomposes goals, evaluates worker outputs, and maintains state across the loop. The workers are narrow by design and execute one thing well. The pattern works because it solves three problems any system trying to act in the world will run into. Specialisation, fault isolation, and observability.

The second commitment is the discipline of confirmation. Every non-trivial action runs after the user approves it. The system drafts, proposes, and waits. The user reads, edits, approves. The action executes. The mechanism is straightforward and the discipline is hard, because every confirmation step is friction the product manager would prefer to remove. Confirmation slows the system down. Confirmation interrupts the user. Confirmation creates moments where the magic of look, the AI just did it gives way to look, the AI proposed something and asked. The temptation to skip confirmation in pursuit of a smoother demo has produced most of the agent disasters of 2025 and 2026.

The third commitment is the discipline of refusal. A system that always acts, regardless of confidence, is more dangerous than a system that sometimes shrugs and asks for help. The refusal is a feature, not a failure mode. The system should know when it is uncertain and behave differently when it is. The system should know when an instruction is suspicious, possibly malicious, possibly the result of prompt injection from an external source, and should refuse rather than execute. Refusal is the architectural answer to the lethal trifecta of agent vulnerabilities.

The fourth commitment is reversibility awareness. Some actions cannot be undone. An email that has been sent cannot be unsent in any meaningful way. A booking that has been made can be cancelled but with cost. The system should know which of its actions are reversible and which are not, and should treat them differently. Irreversible actions deserve more confirmation, more context, more user attention. Reversibility is not a single property. Reversibility is a gradient, and the system's understanding of where each action sits on the gradient is part of what distinguishes a serious agent from a confident one. moccet is being built around all four of these commitments as architecture, not as features.

Why are demos different from production?

Demos look better when the agent runs for ten minutes and ships a polished result. Production looks better when the agent runs for two minutes, hits an ambiguity, asks the user, gets clarification, and finishes correctly. The two are not the same product. Choosing which one to optimise for is a strategic decision, not a tactical one.

Anthropic's data on Claude Code is interesting on this point. Claude Code pauses for clarification more than twice as often as human users interrupt it on complex tasks. The agent's instinct to stop and ask is what makes Claude Code usable in production. The instinct also makes the demo less impressive than a confident agent that runs to completion. Companies optimising for demos have been shipping the latter. Companies optimising for production have been shipping the former.

The fast and loud companies are producing the videos that go viral. The slow and quiet companies are producing the agents that work.

What does an AI taking action unlock for the user?

Most useful work in a person's life is not text. The work is action. The follow-up that needed to be sent. The meeting that needed to move. The dinner that needed to be booked. The reservation that needed to be cancelled. The reply that needed to go. The morning that needed to be protected. None of this is thinking. All of it is doing. The thinking is the easy part. The doing is the work.

A system that helps with thinking is useful and bounded. The user takes the thought, walks back to their life, and does the work themselves. A system that helps with doing is useful and unbounded. The work happens. The user reads what happened.

The leap from thinking to doing is the leap from a tool the user uses to a representative the user delegates to. The leap is the central thing happening in personal intelligence right now. The companies in the category are not racing to build smarter chatbots. The chatbots are smart enough. The race is to build systems that can act on a person's behalf with the reliability, safety, and trustworthiness required for real delegation. The race will not be won by the company with the most capable language model. The race will be won by the company with the most thoughtful architecture for letting that model touch the world.

Try moccet

moccet is a personal intelligence built around a continuous model of one person’s life. The product is in early access. The founders run a live twenty-minute session daily at 1pm Pacific that walks through how it works on a real week.

Claim your seat

Common questions.

An AI takes action when the system produces effects in the world rather than text in a chat. Actions include sending emails, booking meetings, transferring files, calling external APIs, and running code. The leap from text to action requires reliability, sandboxing, confirmation, refusal discipline, and reversibility awareness, none of which is a feature of the underlying language model.
Intelligence raises capability. Reliability is a separate property that requires sandboxing, confirmation steps, observability, idempotent operations, rollback paths, and the discipline to refuse to act when uncertain. A brilliant agent that fails catastrophically 20 percent of the time is worse than a less brilliant one that succeeds 95 percent of the time and fails gracefully.
Anthropic's February 2026 research paper Measuring AI agent autonomy in practice found that Claude Code paused for clarification more than twice as often as human users interrupted it on complex tasks. Most agent actions on Anthropic's API were low-risk and reversible. The agent's instinct to stop and ask is what makes it usable in production.
The orchestrator-worker pattern, where a central orchestrator decomposes goals and routes them to specialised workers. The discipline of confirmation, where every non-trivial action requires user approval. The discipline of refusal, where the system stops when uncertain or when the input is suspicious. Reversibility awareness, where the system treats irreversible actions with more care than reversible ones.
Demos look better when an agent runs for ten minutes and ships a polished result. Production looks better when an agent runs for two minutes, hits an ambiguity, asks the user, and finishes correctly. Companies optimising for demos ship confident agents that fail at scale. Companies optimising for production ship agents that pause, ask, and get the work right.
Live, daily at 1pm Pacific.

See moccet on a real week of yours.

Twenty minutes with the founders. They’ll show you how moccet works on a week like yours, what it’s good at, what it can do for you. Ten minutes for your questions.

Claim your seat