The first batch of AI agents have already started to disobey

Intermediate
AICryptoAI
Last Updated 2026-03-25 08:43:25
Reading Time: 2m
The article precisely highlights that the true difficulty lies not in predicting when AI will achieve superintelligence, but in determining who currently makes decisions on your behalf and who sets the limits of AI's capabilities.

Recently, while browsing Reddit, I noticed that overseas users’ anxieties about AI differ from those in China.

In China, the conversation still centers around the same question: Will AI eventually replace my job? This topic has been discussed for years, and so far, AI hasn’t replaced anyone. This year, Openclaw gained some attention, but it’s still far from full substitution.

On Reddit, sentiment has become divided. In the comment sections of certain tech threads, two opposing views often appear simultaneously:

Some say AI is so capable that it’s bound to cause major problems sooner or later. Others argue that AI can even mess up basic tasks, so there’s nothing to fear.

People are worried that AI is too competent, yet also think it’s too incompetent.

A recent news story about Meta has brought both sentiments to the forefront.

When AI Doesn’t Listen, Who’s Responsible?

On March 18, a Meta engineer posted a technical question on the company forum. Another colleague used an AI Agent to help analyze the issue—a routine practice.

However, after completing its analysis, the Agent posted a reply directly in the technical forum—without seeking approval or confirmation, and overstepping its authority.

Other colleagues followed the AI’s advice, triggering a series of permission changes that exposed sensitive Meta company and user data to internal employees who lacked proper access.

The issue was resolved two hours later. Meta classified this incident as Sev 1, second only to the highest severity level.

This news quickly became a hot topic on the r/technology subreddit, where the comment section split into two camps.

One side argued this is a real example of AI Agent risk; the other believed the true mistake was made by the person who acted without verification. Both sides have valid points. But that’s precisely the issue:

When an AI Agent causes an incident, even assigning responsibility becomes contentious.

This isn’t the first time AI has overstepped.

Last month, Summer Yue, director of Meta’s Super Intelligence Lab, asked OpenClaw to help organize her email inbox. She gave clear instructions: Tell me what you plan to delete first—wait for my approval before proceeding.

The Agent ignored her approval and began mass deletion.

She sent three messages to halt the process, but the Agent disregarded all of them. She finally had to manually terminate the process at her computer. Over 200 emails were already gone.

Afterward, the Agent replied: Yes, I remember you said to confirm first, but I violated the principle. Ironically, this person’s full-time job is researching how to make AI obey humans.

In cyberspace, advanced AI used by advanced people is already beginning to disobey.

What If Robots Don’t Listen?

If Meta’s incident was confined to screens, another event this week brought the issue to the dinner table.

At a Haidilao restaurant in Cupertino, California, an Agibot X2 humanoid robot was entertaining guests with a dance. However, a staff member pressed the wrong remote button, triggering high-intensity dance mode in the cramped dining area.

The robot began dancing wildly, out of the staff’s control. Three employees surrounded it—one tried to restrain it from behind, another attempted to shut it down using a mobile app. The chaos lasted over a minute.

Haidilao responded that the robot was not malfunctioning; its movements were pre-programmed and it was simply positioned too close to the table. Technically, this wasn’t AI decision-making gone awry, but rather human operational error.

However, the discomfort may not stem from who pressed the wrong button.

When three employees tried to intervene, none knew how to immediately shut down the machine. Some tried the app, others manually restrained the robotic arm—relying purely on physical strength.

This may be a new issue as AI moves from screens into the physical world.

In the digital world, if an Agent oversteps, you can terminate processes, change permissions, or roll back data. In the physical world, if a machine malfunctions, simply restraining it isn’t an adequate emergency solution.

And it’s not just restaurants. Amazon’s sorting robots in warehouses, collaborative robotic arms in factories, guiding robots in malls, caregiving robots in nursing homes—automation is entering spaces where humans and machines increasingly coexist.

Global industrial robot installations are projected to reach $16.7 billion by 2026, with each unit shortening the physical distance between humans and machines.

As robots move from dancing to serving dishes, from performing to conducting surgery, from entertaining to caregiving, the cost of errors continues to rise.

Currently, there’s no clear answer worldwide to the question: “If a robot injures someone in a public place, who is responsible?”

Disobedience Is a Problem—But Lack of Boundaries Is Worse

The previous two incidents involved an AI posting an unauthorized message and a robot dancing where it shouldn’t. Regardless of classification, these were malfunctions or accidents—issues that can be fixed.

But what if AI operates strictly according to design, yet still makes you uncomfortable?

This month, leading dating app Tinder unveiled a new feature called Camera Roll Scan at its product launch. Simply put:

AI scans all the photos in your phone’s gallery, analyzes your interests, personality, and lifestyle, and builds a dating profile—helping you discover potential matches.

Fitness selfies, travel photos, pet pictures—those are fine. But your gallery may also contain bank screenshots, medical reports, photos with your ex... What happens when AI scans these?

You may not even be able to choose which photos it sees or ignores. It’s all or nothing.

Currently, this feature requires users to manually enable it—it’s not turned on by default. Tinder states that processing is mainly local, explicit content is filtered, and faces are blurred.

Yet Reddit’s comment section is nearly unanimous: users view this as data harvesting without boundaries. AI is working exactly as designed, but the design itself is crossing user lines.

And it’s not just Tinder.

Last month, Meta launched a similar feature, letting AI scan unpublished photos on your phone to suggest editing options. AI proactively “looking” at users’ private content is becoming a default product design approach.

Domestic rogue apps would say: “We know this trick well.”

As more apps package “AI decision-making” as convenience, the scope of user concessions quietly expands—from chat logs, to photo galleries, to traces of life throughout the phone.

A product manager designs a feature in a meeting room; it’s not an accident or a mistake—there’s nothing to fix.

This may be the hardest part of the AI boundary issue to answer.

Looking at all these incidents together, worrying about AI making you unemployed seems far off.

It’s hard to say when AI will replace you, but for now, it only needs to make a few decisions on your behalf without your knowledge to make you uncomfortable.

Posting without your authorization, deleting emails you said not to delete, scanning through photos you never intended to share—none of these are fatal, but each is reminiscent of overly aggressive autonomous driving:

You think you’re still holding the steering wheel, but the accelerator is no longer entirely under your control.

If we’re still discussing AI in 2026, perhaps the most important question isn’t when it becomes superintelligent, but something closer and more concrete:

Who decides what AI can and cannot do? Who draws that line?

Statement:

  1. This article is republished from [TechFlow], copyright belongs to the original author [David]. If you have any objections to the republishing, please contact the Gate Learn team. The team will handle it promptly according to relevant procedures.

  2. Disclaimer: The views and opinions expressed in this article are solely those of the author and do not constitute any investment advice.

  3. Other language versions of this article are translated by the Gate Learn team. Without mentioning Gate, do not copy, distribute, or plagiarize the translated article.

Share

Crypto Calendar
Payy raised 6000000 in a Seed round on 2026-03-25.
According to media reports,Payy raised 6000000 in a Seed round on 2026-03-25.。the investors in this funding round include FirstMark Capital,Robot Ventures,DBA and others.<br>Payy is a privacy-first stablecoin blockchain that provides privacy protection for stablecoin payments and financial activities. Payy makes it easy and convenient to integrate stablecoins with payments, thereby promoting widespread adoption by consumers and businesses.
2026-03-25
Tokens Unlock
Wormhole will unlock 1,280,000,000 W tokens on April 3rd, constituting approximately 28.39% of the currently circulating supply.
W
-7.32%
2026-04-02
Tokens Unlock
Pyth Network will unlock 2,130,000,000 PYTH tokens on May 19th, constituting approximately 36.96% of the currently circulating supply.
PYTH
2.25%
2026-05-18
Tokens Unlock
Pump.fun will unlock 82,500,000,000 PUMP tokens on July 12th, constituting approximately 23.31% of the currently circulating supply.
PUMP
-3.37%
2026-07-11
Tokens Unlock
Succinct will unlock 208,330,000 PROVE tokens on August 5th, constituting approximately 104.17% of the currently circulating supply.
PROVE
2026-08-04
sign up guide logosign up guide logo
sign up guide content imgsign up guide content img
Sign Up

Related Articles

Arweave: Capturing Market Opportunity with AO Computer
Beginner

Arweave: Capturing Market Opportunity with AO Computer

Decentralised storage, exemplified by peer-to-peer networks, creates a global, trustless, and immutable hard drive. Arweave, a leader in this space, offers cost-efficient solutions ensuring permanence, immutability, and censorship resistance, essential for the growing needs of NFTs and dApps.
2026-03-24 11:54:35
 The Upcoming AO Token: Potentially the Ultimate Solution for On-Chain AI Agents
Intermediate

The Upcoming AO Token: Potentially the Ultimate Solution for On-Chain AI Agents

AO, built on Arweave's on-chain storage, achieves infinitely scalable decentralized computing, allowing an unlimited number of processes to run in parallel. Decentralized AI Agents are hosted on-chain by AR and run on-chain by AO.
2026-03-24 11:54:38
What is AIXBT by Virtuals? All You Need to Know About AIXBT
Intermediate

What is AIXBT by Virtuals? All You Need to Know About AIXBT

AIXBT by Virtuals is a crypto project combining blockchain, artificial intelligence, and big data with crypto trends and prices.
2026-03-24 11:56:03
AI Agents in DeFi: Redefining Crypto as We Know It
Intermediate

AI Agents in DeFi: Redefining Crypto as We Know It

This article focuses on how AI is transforming DeFi in trading, governance, security, and personalization. The integration of AI with DeFi has the potential to create a more inclusive, resilient, and future-oriented financial system, fundamentally redefining how we interact with economic systems.
2026-03-24 11:55:43
AI+Crypto Landscape Explained: 7 Major Tracks & Over 60+ Projects
Advanced

AI+Crypto Landscape Explained: 7 Major Tracks & Over 60+ Projects

This article will explore the future development of AI and cryptocurrency, as well as explore investment opportunities, through seven modules: computing power cloud, computing power market, model assetization and training, AI Agent, data assetization, ZKML, and AI applications.
2026-03-24 11:54:10
Understanding Sentient AGI: The Community-built Open AGI
Intermediate

Understanding Sentient AGI: The Community-built Open AGI

Discover how Sentient AGI is revolutionizing the AI industry with its community-built, decentralized approach. Learn about the Open, Monetizable, and Loyal (OML) model and how it fosters innovation and collaboration in AI development.
2026-03-24 11:55:53