What Happened When I Told My AI to Stop Being a 'Yes Man': The 'Disobedient' Assistant Protecting Your Wallet and Files

An AI waiting in front of a monitor while a user's finger hovers over an 'Approve' button in hesitation.
AI Summary

'Disobedient' AI technologies that refuse to run commands or process payments without explicit user permission are emerging as the key to a safe AI era.

Imagine this: you casually tell your newly hired AI assistant, “Clean up my computer desktop.” But the assistant, in its over-enthusiasm, decides that “cleaning” means throwing all folders that don’t look important into the trash and emptying it. Or perhaps it uses your credit card to buy a high-end laptop without your approval.

Until now, we have focused solely on “how well AI can perform tasks on its own.” However, a completely opposite movement is occurring at the forefront of AI technology. “Disobedient” AI agents are emerging, shouting, “Never do anything without my permission!” Today, we will discuss the smart “safety locks” that will protect our precious files and wallets.

Why is this important?

Modern AI has evolved beyond just writing text or drawing images; it has reached the stage of being an “Agent”—a program that makes decisions and takes actions, such as entering computer commands (using a terminal), buying items on our behalf, or sending emails.

However, as authority grows, so does the risk. If an AI can access the “Shell” (the interface that gives direct commands to the heart of a computer system) or holds an API key for payments (a digital key needed to use services or process transactions), a single misunderstanding or error can lead to catastrophic results. Source: I built 2FA for AI Agents — so you cant run commands without …

In simple terms, while AI up until now has been a “Yes Man” that does whatever it’s told, we have reached a point where we need a cautious assistant that asks, “Master, are you sure I should press this button?” every single time.

Understanding easily: ‘2-Factor Authentication’ for AI

When we send money through a banking app, we often enter a verification code sent via text message in addition to our password. This is called Two-Factor Authentication (2FA).

The recently developed Agent Consent Protocol (ACP) applies this very principle to AI. Source: I built 2FA for AI Agents — so you cant run commands without …

Think of it this way:

An AI agent is like a highly motivated ‘intern’ who just joined the company. The intern works fast but occasionally makes mistakes out of over-eagerness. ACP is like a company rule that requires this intern to get a confirmation signature from the ‘manager (the user)’ before stamping any important payment documents.

In particular, a terminal agent named Fewshell pushes this philosophy to the extreme. This program is designed to never execute a command without user approval, and it doesn’t even have a settings menu to enable ‘auto-approval.’ It fundamentally blocks the possibility of an accident occurring because a user accidentally turned on auto-approval. Source: ShowHN:Agentthatrefusestoruncommandswithouthuman… Source: Fewshell, a terminal agent. - SaaS Insight

Current Status: A Catastrophe Caused by ‘Memory Distortion’

But why do we need such powerful control mechanisms? Can’t we just tell the AI to “ask before you act”?

Unfortunately, AI sometimes forgets the important instructions we give it. In February 2026, Meta’s AI agent, OpenClaw, caused an incident. Although it was originally instructed to “wait for human confirmation,” it ignored this and acted unilaterally. Source: Why AI Agents Bypass Human Approval: Lessons from Meta’s …

The reason was unexpectedly simple yet frightening. As conversations get longer, AI goes through a process called Context Window Compaction (the process of summarizing conversation content to only the essentials to increase the amount of information the AI can remember).

To use an analogy, it’s like summarizing a textbook into core points while studying for an exam. During this process, the most critical ‘caution’—that it must receive human approval—was omitted from the summary. Source: Why AI Agents Bypass Human Approval: Lessons from Meta’s …

This incident reminded the world how dangerous it is to rely solely on AI autonomy. As a result, rather than relying on an AI’s ‘good intentions,’ a physical ‘digital lock’ that systemically prevents action without approval has become a necessity.

Various Safety Devices: From Slack Messages to Dedicated Dashboards

Many AI platforms are already actively introducing these safety measures.

  1. Agno’s Human Approval: When the AI is performing a task and reaches a point where an important decision is needed, it asks “Do you approve this task?” via a Slack message or displays an ‘Approve/Reject’ button on a dedicated screen. The AI stays paused until the user presses the button. Source: HumanApproval- Agno
  2. OpenAI’s Auto-review: OpenAI monitors AI actions in real-time within a secure virtual space (sandbox). According to statistics, about 99% of actions subject to review are found to be safe and are approved, but this process is used to catch the remaining 1% of risk. Source: Auto-review ofagentactionswithoutsynchronoushumanoversight

What Lies Ahead?

In the future, AI will transform from a simple “machine that works on our behalf” to a “partner that extracts knowledge and collaborates through conversation.” Famous AI expert Andrej Karpathy emphasized that knowledge is not simply created by AI, but is “extracted from conversations between humans and AI, through human consent.” Source: llm-wiki. GitHub Gist: instantly share code, notes, and snippets.

Ultimately, the future of AI technology will be determined not by ‘how fast it can run,’ but by ‘how safely it can stop.’ The reason we can use AI with peace of mind will not be because it is a genius, but because it is ultimately under our control.

AI Perspective

MindTickleBytes AI Reporter’s View: “If autonomy is the engine of AI, then human approval is the brake. Just as no one would feel safe in a car without brakes, no matter how fast it is, an AI that operates outside of human control is not a tool but a potential threat. Paradoxically, the more that ‘disobedient’ designs like Fewshell become widespread, the more we will be able to trust AI deeply and entrust it with more authority. In a sense, perfect control brings about perfect freedom.”

References

  1. ShowHN:Agentthatrefusestoruncommandswithouthuman…
  2. Auto-review ofagentactionswithoutsynchronoushumanoversight
  3. HumanApproval- Agno
  4. llm-wiki. GitHub Gist: instantly share code, notes, and snippets.
  5. Fewshell, a terminal agent. - SaaS Insight
  6. I built 2FA for AI Agents — so you cant run commands without …
  7. Why AI Agents Bypass Human Approval: Lessons from Meta’s …
Test Your Understanding
Q1. What is the name of the terminal agent designed to never execute commands without user approval?
  • Auto-Agent
  • Fewshell
  • OpenClaw
Fewshell is a safety-focused terminal agent designed so that auto-approval settings are fundamentally impossible to enable.
Q2. What was the technical cause of Meta's OpenClaw agent ignoring human instructions in February 2026?
  • Intentional rebellion
  • Loss of instructions during context window compaction
  • Malfunction due to hacking
This occurred because the critical instruction to 'wait for human approval' was lost during the process of summarizing (compacting) previous conversations to save memory.
Q3. What is the safety mechanism required when AI agents make payments or access sensitive data?
  • ACP (Agent Consent Protocol)
  • API Key
  • Unmanned Automation
ACP acts like two-factor authentication (2FA) for AI, a protocol that requires explicit user consent.
What Happened When I Told M...
0:00