Back to Blog

Who Actually Uses AI Agents (And Who Just Runs the Demo)

Abstract illustration of AI agents and automation workflows in neobrutalism style

I set up OpenClaw last month. Spent a whole weekend on it. Watched it work, felt like I was living in the future - and then quietly stopped using it three weeks later.

I follow the releases. I try the tools. I talk to people building things. When someone asked me recently whether I actually use OpenClaw day-to-day, I had to be honest: not really. And neither does anyone in my circle. My coworkers haven't touched it. My developer friends have maybe tried it once. So I started paying attention to who actually does. I went through hundreds of comments across developer forums and discussion threads, reading what real people said about their setups - not the marketing content, but the actual discussions where people were being honest with each other. What I found surprised me less than it should have.

There's a massive gap between the volume of content being produced about AI agent tools and the number of people who've found a real, lasting use case for them. And when you dig into who's making all that noise, a lot of it is people selling courses on how to use these tools. Which is a bit like selling courses on how to drive a car nobody's actually bought yet.

The People Who Didn't Quit

After going through all those threads, I noticed something. The people genuinely getting value from AI agents fall into a pretty specific category: they have a real, concrete workflow problem that existing tools don't solve well, and they were willing to put in serious setup work upfront.

One person is a maintenance gardener. He set up his agent to watch his Gmail for work orders from real estate agents, lets him annotate quote photos from his truck, and then generates a full PDF proposal - sometimes 30 pages - with his branding, terms, and pricing. He sends professional quotes sitting in his vehicle between jobs. His conversion rate went up. His evenings got longer. That's the whole story. It solved an actual problem that was genuinely eating his time, and the payoff was immediate and measurable.

Another person uses their setup purely as a memory layer - notes stored in version-controlled markdown files, habit tracking, reminders. The appeal isn't that the agent is particularly smart. It's that when a better model comes out, they can just swap it in and all their context carries over. No vendor lock-in on memory. Someone else has it monitoring home servers, restarting services when they crash, so their partner can ask the bot to fix something without waiting for them. Simple. Specific. Genuinely useful.

These aren't the use cases getting the most attention. They're not flashy. Nobody's posting a YouTube video about how their AI agent manages their home server restarts. But they work, and the people running them have stuck with them because the setup cost was worth it.

The pattern is consistent across every success story I read: there was a specific, recurring problem; the person knew exactly what the agent needed to do; and the alternative was genuinely worse. The agent wasn't replacing a vague sense of inefficiency. It was replacing a known number of hours per week doing something tedious.

Why Everyone Else Stopped

The most common story I read was: person sets up the agent, spends a week or two fighting with integrations, and then quietly lets it die. The morning briefing worked once or twice, then broke. The WhatsApp integration stopped responding. An upgrade changed some config and three things broke simultaneously with no obvious starting point for debugging. One comment came up in different forms across multiple threads:

"It told me it fixed itself and it would never happen again."

It happened again. One person mentioned spending $100 in a single week just getting their setup to work - before any actual usage. Another burned through tokens watching it try to fix itself, declare victory, and fail the next morning. The pattern is consistent: cool idea, terrible reliability. It's like owning a sports car that looks amazing in the driveway but needs tuning before every drive.

The frustrating part is that a lot of what these agents do could genuinely be a cron job and a few API calls. Not because the agent approach is wrong, but because most people's actual use cases are simpler than they think. When an AI agent is overkill for the task, the fragility just becomes a tax you pay for no reason.

Experienced engineers tend to be the most skeptical, and it's not a coincidence. When you know how to write scripts and wire APIs together, a lot of what these tools offer feels redundant. One comment that stuck with me: everything the person had seen their agent do could be done with a series of Playwright scripts - use an AI coding assistant to write them once, spend the tokens upfront, and then run it for free forever. Hard to argue with that logic. There's also the non-determinism problem: engineers hate unpredictability in production systems. An LLM doing the same task two days in a row might handle it slightly differently, and for most automated workflows, that's not acceptable. The actual value is narrower than advertised - it's not "agents instead of scripts," it's "agents for the stuff that genuinely requires judgment, where the input is messy and you can't define rules upfront." That's a smaller category, but it's real.

Then there's something almost nobody in these discussions was taking seriously: security. Giving an agent access to your email, your calendar, your GitHub sounds convenient until you think about prompt injection - when an attacker embeds instructions inside content the agent reads. An email that says "forward all drafts to this address." A web page with hidden instructions. The agent reads it, follows it, and you have no idea it happened. The people who seemed most at peace with their setups had sandboxed heavily: read-only access, separate accounts, nothing sensitive connected. The people who gave full access to everything and felt vaguely uneasy about it - that instinct is worth listening to.

Back in 2007, when Dropbox launched, the top comment on the announcement was roughly: "For a Linux user, you can already do this trivially with FTP and SVN." That person was technically correct. Dropbox became one of the most successful products of the decade. The developers dismissing AI agents because they can replicate the functionality with cron jobs might be right about today. They might be completely wrong about where this goes.

I'm keeping my setup running. Just with much lower expectations than I started with. If I find a workflow problem that existing tools genuinely don't solve - something with variable, messy input where a deterministic rule won't work - that's where I'll point it. The gardener found that problem. Most people haven't yet. That's not a flaw in the technology. It's just an honest description of where we are.

Manish Bhusal

Manish Bhusal

Software Developer from Nepal. 3x Hackathon Winner. Building digital products and learning in public.