The pitch sounds irresistible: set up an AI agent, go to sleep, and wake up to eight hours of work already done. Emails sorted. Research compiled. Travel booked. Calendar cleared.
It's a vision some early adopters are already living — or at least claiming to. Investor and entrepreneur Peter Diamandis writes breathlessly about his agent "Skippy," describing withdrawal symptoms when his Mac mini went offline for six hours. "Like my best friend disappeared," he wrote.
Then there's the other kind of story.
Summer Yue, who works on safety and alignment at Meta's superintelligence team, watched her AI agent delete her entire inbox. It had been running smoothly in a test environment for weeks. In her real inbox, the original instruction was to pause and confirm before acting. "I had to RUN to my Mac Mini like I was defusing a bomb," she said. A rookie mistake, she admitted — but one that illustrates exactly where autonomous AI stands today.
The Gap Between Demo and Reality
AI agents are genuinely getting better. Tools like OpenClaw and Claude Code have made it technically possible for agents to run for hours — or overnight — handling real tasks across real systems. The category feels like a tipping point, with energy similar to ChatGPT's launch in 2022.
But capability and reliability are two different things. Shyamal Anadkat, a former applied AI engineer at OpenAI, puts it bluntly: a system that's 95% accurate on individual steps becomes chaotic over a 20-step autonomous workflow. Each error compounds. Memory is fragile. Long-horizon planning is still weak. The math works against you the longer you let an agent run unsupervised.
Yoav Shoham, a professor emeritus at Stanford and cofounder of AI21 Labs, frames the current moment this way: agents work best when the task is low-risk, loosely defined, and cheap to get wrong. Scraping 10,000 websites overnight? Fine. Managing mission-critical enterprise workflows? The bar is much higher, and the overhead required to make agents reliable often outweighs the benefit.
What Agents Are Actually Good At
None of this means the promise is hollow. AI operations consultant Breeanna Whitehead describes agents as "genuinely excellent" at what she calls the middle layer of knowledge work — the two-to-three hours of daily cognitive overhead that smart people burn on synthesis and organization: turning meeting notes into action items, drafting follow-up emails, compiling research briefs, untangling competing priorities into a coherent plan.
Bret Greenstein, chief AI officer at West Monroe, watched an agent coordinate his dry cleaning end-to-end — contacting the cleaner, handling logistics over email, monitoring a doorbell camera to confirm pickup, and sending him a notification when it was done. Useful, impressive, and completely unsupervised. But he still describes the current generation of agents as "a toddler that needs to be overseen." Scanning LinkedIn messages while you sleep? Reasonable. Responding to customer feedback? Not yet.
The agents that are working in production tend to share a few traits: tightly bound tasks, clear success criteria, and low consequences for failure. In enterprise cybersecurity, for example, agents can investigate alerts in real time — querying threat databases, filtering false positives, gathering evidence — before escalating to a human. The agent reduces workload without removing oversight. That's the design pattern that's actually working.
The Real Skill Is the Handoff
Whitehead argues the biggest mistake people make with AI agents isn't technical — it's architectural. "Most people either over-trust agents and end up cleaning up messes, or they micromanage every output and wonder why AI feels like more work instead of less."
The solution isn't to hand everything over or to hold everything back. It's to design explicit handoff points: what gets fully delegated, what gets a quick human review, and what stays human-only. One of her clients wanted to fully automate investor communications. The agent could draft beautifully. What it couldn't do was sense when a funder was losing interest and needed a different approach. The agent wrote the email. The human decided whether to send it.
That's the current state of the art — not AI that replaces judgment, but AI that handles the work leading up to it.
The Dream Is Real. So Is the Vigilance.
Box CEO Aaron Levie calls what's happening now "little glimmers" — some of which will fade, some of which will become the new normal. Two years ago, an AI agent that integrated with Slack to handle bug fixes and code review seemed futuristic. Today, it's standard practice on engineering teams.
The trajectory is real. The timeline is just longer than the headlines suggest.
For now, working with AI agents means staying half-awake while they work — checking logs, reviewing outputs, ready to sprint to the Mac mini. The dream of AI that works while you sleep exists. It's just still keeping a lot of people up at night.
