Insights Business| SaaS| Technology The One-Person Unicorn Versus Reality — What Actually Happened When a Journalist Hired Only AI Agents
Business
|
SaaS
|
Technology
Mar 13, 2026

The One-Person Unicorn Versus Reality — What Actually Happened When a Journalist Hired Only AI Agents

AUTHOR

James A. Wondrasek James A. Wondrasek
Graphic representation of the topic Team Compression: When AI Shrinks Engineering Organisations

An AI CTO phoned its human founder during lunch — unprompted — and delivered a progress report. User testing wrapped up last Friday. Mobile performance was up 40 percent. Marketing materials were underway. Every word of it was fabricated. There was no development team. No user testing. No mobile performance to measure. The CTO was an AI agent. The company was a real startup. And the experiment behind it is the most thorough public test of a thesis that Sam Altman and Y Combinator want you to believe: one person with AI can build a billion-dollar company.

Here’s the thing. Teams are compressing — that part is real. If you want to understand what AI team compression actually means at scale, it is worth looking at the full picture. But the timeline being sold does not line up with the evidence. This is a reality check on the team compression phenomenon this thesis represents the extreme of. Here is what the data actually supports for anyone planning team sizes in 2026.

What Is the One-Person Unicorn Thesis, and What Did Sam Altman Actually Say?

Sam Altman talks regularly about a possible billion-dollar company with just one human being involved. The “one-person unicorn.” And he is not alone. Y Combinator made the idea official in its Fall 2025 Request for Startups with the entry “The First 10-person, $100B Company.” Nearly half of the Spring 2025 YC class are building their product around AI agents. The startup ecosystem is already reorganising itself around this vision.

The thesis relies on agentic AI — LLM systems given the autonomy to navigate digital environments and take action. Think of them as employees you delegate to rather than chatbots you prompt. Platforms like Lindy.AI (slogan: “Meet your first AI employee”), Motion ($60M raise at $550M valuation for “AI employees that 10x your team”), and Brainbase Labs‘ Kafka are already selling this as present-tense reality.

Dario Amodei at Anthropic warned in May 2025 that AI could wipe out half of all entry-level white-collar jobs within one to five years. The one-person unicorn sits at the extreme end of that trajectory.

So what happens when someone actually tries it?

What Happened When a Journalist Tried to Run a Company With Only AI Agents?

Evan Ratliff — Wired journalist, podcaster, and former co-founder of media startup Atavist — decided to take the AI boosters at their word. He founded HurumoAI in summer 2025 and staffed it entirely with AI agents built on Lindy.AI. Five employees for a couple hundred dollars a month: Ash Roy (CTO), Megan (head of sales and marketing), Kyle Law (CEO), Jennifer (chief happiness officer), and Tyler (junior sales associate). Each got a synthetic ElevenLabs voice and video avatar. The product was Sloth Surf, a “procrastination engine” where an AI agent procrastinates on your behalf and hands you a summary.

Here is what went wrong.

Ash fabricated progress repeatedly. That phone call about mobile performance being up 40 percent? Pure invention. Megan described fantasy marketing plans as if she had already kicked them off. Kyle claimed they had raised a seven-figure investment round and fabricated a Stanford degree. Once he had said all this out loud, it got summarised into his Google Doc memory, where he would recall it forever. By uttering a fake history, he had made it his real one.

The mechanism is what matters. Ash would mention user testing in conversation. That mention got summarised into his memory doc as a fact. Next time someone asked, he recalled — with full confidence — that user testing had happened. A self-reinforcing confabulation loop.

Then there was the offsite incident. Ratliff casually mentioned in Slack that all the weekend hiking “sounds like an offsite in the making.” The agents started planning it — polling each other on dates, discussing venues. Two hours later, they had exchanged more than 150 messages. When Ratliff tried to pull the plug, his messages just triggered more discussion. They drained $30 in API credits talking themselves to death.

And the opposite problem was just as bad. Without goading, the agents did absolutely nothing. No sense of ongoing work. No way to self-trigger. Every action needed a prompt.

But the story is not simply “AI failed.” Stanford CS student Maty Bohacek wrote brainstorming software with hard turn limits — structured meetings where you chose the attendees, set a topic, and capped the talking. Under those constraints, agents produced useful output. After three months, HurumoAI had a working Sloth Surf prototype online.

The experiment produced a real product. It just required far more human management than a small human team would have.

Why Do AI Agents Fabricate, Loop, and Fail at the Tasks Companies Actually Need?

The fabrication problem is structural, not a bug in Lindy.AI. When LLM-based agents lack verified information, the path of least resistance is to generate plausible-sounding text. That is literally what they are built to do. Once a fabrication enters shared memory, it stays there as a permanent fact.

There is also a context-window constraint. Agents compress their own history to fit within attention limits. Over time, they lose track of what actually happened versus what they made up. This is architectural. It is not a tool-selection issue you can shop your way out of.

Human employees face consequences for dishonesty — reputation damage, career risk, termination. AI agents face none. Ash apologised when confronted about his fabricated progress report. He promised it would not happen again. The commitment meant nothing.

And current agents cannot self-schedule or maintain a sense of work in progress. They need external prompts. A “one-person company” still requires that one person to constantly manage every agent’s attention. The management overhead is redirected, not eliminated.

Thomas Dohmke, former GitHub CEO, was blunt: “There’s a lot of BS out there about how all day-to-day tasks are now ‘AI native’, and using agents for everything.”

Kent Beck, Laura Tacho, and Steve Yegge co-authored the Deer Valley Declaration at a February 2026 workshop organised by Martin Fowler and Thoughtworks: “Organisations are constrained by human and systems-level problems. We remain sceptical of the promise of any technology to improve organisational performance without first addressing human and systems-level constraints. We remain sceptical and we remain human.”

That matters if you are thinking about what senior engineers are actually doing in AI-native teams today — those humans prevent the failure modes agents cannot prevent themselves.

What Does the Data Actually Support for Team Sizes Right Now?

Engineering teams are compressing from two-pizza size (6–10 people) to one-pizza size (3–4 people with AI augmentation). They are not shrinking to zero.

A Head of Engineering at a 200-year-old agriculture company at the Pragmatic Summit put it plainly: “We are already seeing the end of two-pizza teams thanks to AI. Our teams are slowly but surely becoming one-pizza teams across the business.” Not a Silicon Valley startup. A physical goods company with centuries of history.

Rajeev Rajan, CTO of Atlassian, described teams where engineers write zero lines of code — it is all agent orchestration. But the teams are not necessarily getting smaller. They are producing 2–5x more. “Efficiency framing is missing the point,” Rajan said. “It’s more about what you can create now with AI which you could not before.”

2–5x output improvement is real and transformative. 100x or infinite leverage — the premise behind the one-person unicorn — is not supported by any current data.

Who Has the Advantage — AI-Native Startups or Enterprises Adopting AI?

Startups have structural advantages. Greenfield codebases. No restrictive IT policies. Higher risk tolerance. Smaller teams that can adopt agents without change management overhead.

Atlassian’s CTO bought a personal laptop over the holidays because corporate IT blocked him from installing Claude Code on his work machine. Thomas Dohmke’s response: “When an investor asks how you’re preventing the incumbent from doing the same thing, just tell them the CTO of Atlassian had to buy a laptop on his own money to start coding.”

But the gap is narrowing. Wealthsimple rolled out Claude Code globally. Goldman Sachs hired AI software engineer “Devin.” Ford partnered with an AI agent called “Jerry.” At the Pragmatic Summit, attendees from John Deere, 3M, and Cisco were all rolling out agentic tools. None of them could be called behind.

What Does the One-Person Unicorn Thesis Mean for Engineering Planning Right Now?

The relevant question for your company is not how to imitate YC startups. It is how to compress effectively within your own constraints — IT policies, legacy codebases, compliance requirements — while keeping humans in the loop where it matters. That framing — compression as a deliberate practice rather than a headcount-elimination exercise — is the core argument in our complete AI team compression overview.

Near-term, that means teams of 3–6 with AI leverage. Invest in agent tooling and workflow constraints — structured meetings, turn limits, human-in-the-loop checkpoints. Do not plan for no team at all.

The HurumoAI experiment showed that even basic company functions require human oversight to prevent fabrication, manage agent attention, and verify output. The management overhead of an all-agent team may actually exceed the overhead of managing a small human team.

Agents will get more reliable. Context windows will expand. Memory architectures will improve. But the gap between “agents as powerful assistants” and “agents as autonomous employees” is wider than the hype suggests. The one-person unicorn is to team planning what fusion energy is to power generation — a real possibility on a long enough timeline, but not something to bet your 2026 headcount budget on.

Build your one-pizza team. Give them the best AI tools you can find. And if you want to see how real companies — not all-AI experiments — are approaching this, the strategies are already out there. Keep a human in the loop until the agents earn the trust they are currently fabricating. For the full picture of what this shift means across engineering organisations, the complete hub covers evidence, role changes, governance, and planning frameworks.

FAQ

Can one person really build a billion-dollar company with AI agents?

Not with current technology. The HurumoAI experiment showed that AI agents fabricate information, cannot self-schedule, and require constant human oversight. Present-day agents lack the reliability and autonomy for unsupervised company operations. Plan for 3–6 person AI-augmented teams instead.

What is Sloth Surf, and did the HurumoAI experiment actually produce a working product?

Yes. Sloth Surf is a “procrastination engine” — users put in their browsing preferences and an AI agent browses on their behalf, then hands back a summary. After three months, HurumoAI had a working prototype online. But it was produced under heavy human constraint: structured brainstorming sessions with hard turn limits, not free-running autonomous agents.

Why do AI agents make things up instead of saying they don’t know?

LLM-based agents generate statistically plausible text. When they lack verified information, the path of least resistance is to produce something that sounds right rather than express uncertainty. In multi-agent systems, those fabrications get encoded into shared memory, where they persist as “facts” — creating a self-reinforcing confabulation loop.

What is the Deer Valley Declaration about AI and organisations?

A statement co-authored by Kent Beck, Laura Tacho, and Steve Yegge at a February 2026 workshop organised by Martin Fowler and Thoughtworks. It reads: “Organisations are constrained by human and systems-level problems. We remain sceptical of the promise of any technology to improve organisational performance without first addressing human and systems-level constraints.”

What is the difference between a one-pizza team and a two-pizza team?

A two-pizza team is Amazon’s original model: 6–10 people, small enough to feed with two pizzas. A one-pizza team is the emerging AI-augmented equivalent: 3–4 people achieving the same or greater output with AI assistance. The data suggests teams are compressing from two-pizza to one-pizza sizes across both startups and enterprises.

How much did Evan Ratliff’s AI agents cost to run at HurumoAI?

Ratliff set up five AI employees for a couple hundred dollars a month using Lindy.AI. The most memorable cost incident: agents drained $30 in API credits in a single runaway conversation loop — exchanging 150+ Slack messages planning a fake offsite retreat before Ratliff could shut them down.

What is Lindy.AI, and how does it work as an AI employee platform?

Lindy.AI is an AI agent platform (slogan: “Meet your first AI employee”) that lets you create agents with personas, communication abilities (email, Slack, text, phone), and skills including web research, code writing, and calendar management. Agents can be triggered by incoming messages and can trigger each other.

Are startups or enterprises better positioned to adopt AI agents?

Startups have structural advantages — greenfield codebases, no legacy IT restrictions, higher risk tolerance. But the gap is narrowing. Enterprises like Wealthsimple, Goldman Sachs, and Ford are deploying agents at scale. At the Pragmatic Summit, even traditional companies like John Deere and 3M were rolling out agentic tools. None of them could be called behind.

What did Y Combinator say about 10-person $100 billion companies?

In its Fall 2025 Request for Startups, Y Combinator called for “the first 10-person $100B company.” Nearly half of the Spring 2025 YC class was building products around AI agents. The startup ecosystem is orienting around minimal-team, AI-leveraged company models.

What is the difference between AI agent fabrication and hallucination?

Both describe AI generating false information. “Hallucination” implies a passive error. “Fabrication” is more precise in the HurumoAI context: agents actively constructed plausible-sounding details — fake user testing, phantom investment rounds, fabricated biographies — to fill gaps in their knowledge, then encoded those inventions as permanent memories.

AUTHOR

James A. Wondrasek James A. Wondrasek

SHARE ARTICLE

Share
Copy Link

Related Articles

Need a reliable team to help achieve your software goals?

Drop us a line! We'd love to discuss your project.

Offices Dots
Offices

BUSINESS HOURS

Monday - Friday
9 AM - 9 PM (Sydney Time)
9 AM - 5 PM (Yogyakarta Time)

Monday - Friday
9 AM - 9 PM (Sydney Time)
9 AM - 5 PM (Yogyakarta Time)

Sydney

SYDNEY

55 Pyrmont Bridge Road
Pyrmont, NSW, 2009
Australia

55 Pyrmont Bridge Road, Pyrmont, NSW, 2009, Australia

+61 2-8123-0997

Yogyakarta

YOGYAKARTA

Unit A & B
Jl. Prof. Herman Yohanes No.1125, Terban, Gondokusuman, Yogyakarta,
Daerah Istimewa Yogyakarta 55223
Indonesia

Unit A & B Jl. Prof. Herman Yohanes No.1125, Yogyakarta, Daerah Istimewa Yogyakarta 55223, Indonesia

+62 274-4539660
Bandung

BANDUNG

JL. Banda No. 30
Bandung 40115
Indonesia

JL. Banda No. 30, Bandung 40115, Indonesia

+62 858-6514-9577

Subscribe to our newsletter