Skip to main content
MindStudio
Pricing
Blog About
My Workspace

Sam Altman's Most Honest Tweet: Why the CEO of OpenAI Can't Stop Working Since Building AGI Tools

Altman tweeted that someone switched to polyphasic sleep to maximize Codex usage — and called it the most honest thing he'd ever said. Here's what it reveals.

MindStudio Team RSS
Sam Altman's Most Honest Tweet: Why the CEO of OpenAI Can't Stop Working Since Building AGI Tools

The CEO of OpenAI Cannot Stop Working

Sam Altman tweeted two quotes side by side last week, and the contrast is worth sitting with. The first: “Post-AGI, no one is going to work and the economy is going to collapse.” The second: “I’m switching to polyphasic sleep because GPT-5.5 and Codex is so good that I can’t afford to be sleeping for such long stretches and miss out on working.”

He didn’t editorialize. He didn’t need to.

If you build AI tools for a living, that juxtaposition should stop you cold. The person running the organization most responsible for the “no one will work” narrative is personally rearranging his sleep schedule to get more hours with his own product. That’s not a PR move. That’s revealed preference, and revealed preference is the most honest signal you’ll ever get.

Cheyen Jiao put it cleanly: “Polyphasic sleep to maximize Codex usage is the most honest thing Sam has ever tweeted. Forget the AGI philosophy for a second. The revealed preference is that the CEO of the company building these tools literally cannot stop using them because the output per hour is too valuable to waste on sleeping.”

That’s the thing worth unpacking. Not the tweet itself — the mechanism behind it.

Why “AI Saves Time” Was Always the Wrong Frame

Everyone else built a construction worker.
We built the contractor.

🦺
CODING AGENT
Types the code you tell it to.
One file at a time.
🧠
CONTRACTOR · REMY
Runs the entire build.
UI, API, database, deploy.

For the first two years of the generative AI era, the dominant story was time savings. You’d run a survey, ask people what AI did for them, and “saves me time” would top the list. The implicit model was: AI does some work, you do less work, you clock out earlier.

That model was wrong, and we’re now accumulating enough evidence to say so clearly.

Aaron Levy, CEO of Box, wrote this week: “Sorry to anyone who thought AI would mean we’d work less, at least for now. AI makes it easy to explore more than you did before, and so you start doing far more as a result.”

Shaunu Matthew, responding to that post, described logging 6am to 10pm days consistently — not because of deadlines, but because the work kept opening up. His specific observation: it’s hard to step away when you think you just need to point an agent at a detailed spec and let it run, but you keep surfacing the next three to five things to work on while the current task is in flight.

Abdul Khadir found out about Paperclip — an open-source orchestration layer positioning itself as the infrastructure for zero-human companies — at 1am and skipped sleep entirely because of what it was unlocking for his business.

Brian Johnson, who has built his entire public identity around optimized health habits, broke his screen-off rule, turned down socializing, and fell behind on other work because of Claude. His summary: “AI is preposterous. As close to magic as I’ve experienced.”

These aren’t isolated anecdotes. They’re a pattern. And the pattern has a name.

The Infinite Backlog

There’s a concept in economics called the lump of labor fallacy — the mistaken belief that there’s a fixed amount of work in the world, and if a machine does some of it, a human loses a job. The reason it’s a fallacy is that work isn’t finite.

In any organization operating in an expanding market, there is always more to do. There are always projects that would get done if you had the time and resources. Features that would ship. Markets that would get analyzed. Content that would get produced. Partnerships that would get explored. The job of leadership is to select from this infinite backlog and translate a tiny slice of it into a roadmap. The job of individual contributors is to execute that slice.

The infinite backlog has always existed. What’s changed is that agents make it visible and immediately actionable.

When AI was an assistant — autocomplete, drafting help, summarization — it compressed time. You got more done in the same hours. But Friday afternoon still arrived. You could still look at the week and feel like you’d done enough.

Agents break that. Agents aren’t you being more productive. They’re you replicating yourself. Multiple instances, running in parallel, working while you sleep. Everything you might do could theoretically be happening right now, simultaneously. The infinite backlog stops being a theoretical future constraint and becomes a contemporary one — a neverending list of immediate unmet opportunities.

Cursor
ChatGPT
Figma
Linear
GitHub
Vercel
Supabase
remy.msagent.ai

Seven tools to build an app. Or just Remy.

Editor, preview, AI agents, deploy — all in one tab. Nothing to install.

This is why Sam Altman is considering polyphasic sleep. It’s not that Codex is addictive in the way social media is addictive. It’s that the output per hour is genuinely high enough that sleeping feels like leaving value on the table. When your tools can compound work while you’re unconscious, the calculus on sleep changes.

The Constraints That Remain

Here’s where the picture gets more complicated, and where the real implications for builders live.

The infinite backlog is real. The parallelism is real. But the constraints haven’t disappeared — they’ve shifted.

Tang Yan identified one of the most important ones: judgment drain. Managing a fleet of agents doesn’t drain you through typing. It drains you through decisions. More context switching, more verification, more choices per hour. The result is that instead of 8-10 normal productive hours, you might get four or five extremely intense hours before your cognitive capacity is genuinely depleted. Some of his friends, he notes, are already burnt out — they don’t say it out loud, but the signs are there.

The agent can keep working 24/7. The human still has a hard limit.

Beyond judgment, there are at least four other constraint categories that don’t disappear in a world of theoretically infinite agents:

Planning. Which tasks do you start, in what sequence, and when? Agents don’t self-prioritize across your full strategic context. You do.

Coordination. When you have multiple agents running in parallel, how do you ensure they’re working toward a coherent end rather than generating contradictory outputs or duplicating effort?

Evaluation. Are you trusting that every agent output is correct? Someone — probably you — has to check the work, identify failures, and route corrections back into the system.

Cost. Token costs are real. Compute is finite. The economics of token-based pricing mean that “run everything in parallel forever” is not actually a strategy available to most organizations. You will still have to choose.

And then there’s what you might call the absorption constraint: whoever is supposed to receive the output of all this agent work — a market, a customer, a reader, a user — has a finite capacity to consume it. You can generate more than they can absorb. That’s a real ceiling.

What This Means If You’re Building Right Now

The Sam Altman tweet isn’t really about sleep schedules. It’s about the fact that the most sophisticated AI user in the world — someone with access to every model, every tool, every internal capability — is personally experiencing the same phenomenon that Shaunu Matthew and Abdul Khadir and Brian Johnson are experiencing.

The tools are good enough that time is now the bottleneck. Not model quality. Not access. Time.

That reframe has concrete implications for how you should be building.

If you’re building AI applications for other people, the question is no longer “does this save time?” The question is “does this help people navigate their infinite backlog without burning out?” Those are very different design problems. The first is about efficiency. The second is about prioritization, pacing, and judgment support.

REMY IS NOT
  • a coding agent
  • no-code
  • vibe coding
  • a faster Cursor
IT IS
a general contractor for software

The one that tells the coding agents what to build.

Aaron Levy is already operationalizing this. He’s actively hiring for what he’s calling agent engineering roles — a job category that didn’t exist 12 months ago. His description of the role is specific: an internal FTE whose job is to wire up internal systems and get agents working with them effectively. The person connects Box, Salesforce, Workday. They codify workflows. They work embedded with business teams. And Levy expects this to spawn a second adjacent role: something like agent product management for internal processes, sitting on the business side rather than the technical side.

The infrastructure question for these agent engineering roles is real. Platforms like MindStudio handle a meaningful chunk of the orchestration complexity — 200+ models, 1,000+ integrations, a visual builder for chaining agents and workflows — which matters when the agent engineer’s job is wiring systems together rather than writing orchestration code from scratch.

The new roles don’t stop at agent engineering. If you follow the logic of the infinite backlog meeting its new constraints, you start to see the shape of other positions: eval engineers who build quality gates rather than assuming every agent output is trustworthy; context librarians who curate what agents know and manage the permissioning that determines what they can access; coordination architects who keep parallel workstreams legible to each other and to management.

None of these are speculative in the way “AI will create new jobs” is usually speculative. They’re responses to specific, observable constraints that already exist.

The Spec Is the New Source of Truth

One implication that’s easy to miss: when agents are doing the execution, the quality of your specification becomes the primary determinant of output quality. The person who can write a precise, complete spec — one that carries intent clearly enough for an agent to act on it without constant correction — has a skill that compounds.

This is why tools that treat the spec as the actual source of truth are worth paying attention to. Remy, MindStudio’s spec-driven app compiler, takes this seriously: you write your application as annotated markdown, and Remy compiles it into a complete TypeScript backend, SQLite database, frontend, auth, and deployment. The spec isn’t a starting point that gets abandoned when the code diverges — it’s the source of truth that the code is derived from. That’s a different relationship between intent and implementation than most builders are used to.

The broader pattern matters: as agents take on more execution, the human contribution shifts toward specification, evaluation, and judgment. The people who figure out how to write better specs faster will have a significant advantage.

The Honest Signal in the Polyphasic Sleep Tweet

Here’s the opinion this post is allowed: the “AI will eliminate work” narrative and the “AI will create new jobs” narrative are both missing the more interesting thing that’s actually happening.

Work isn’t being eliminated. The frontier of what’s possible is expanding faster than individuals and organizations can pursue it. The constraint isn’t capability anymore — it’s the human capacity to direct, evaluate, and coordinate the capability that already exists.

Sam Altman switching to polyphasic sleep isn’t a sign of workaholism or poor boundaries. It’s a data point about where the bottleneck has moved. When the CEO of OpenAI finds that time is his binding constraint for using his own tools, that tells you something about the state of the technology that no benchmark or press release can.

Remy is new. The platform isn't.

Remy
Product Manager Agent
THE PLATFORM
200+ models 1,000+ integrations Managed DB Auth Payments Deploy
BUILT BY MINDSTUDIO
Shipping agent infrastructure since 2021

Remy is the latest expression of years of platform work. Not a hastily wrapped LLM.

The people who will do best in this environment aren’t the ones who work the most hours. They’re the ones who get good at the new scarce resources: judgment about what matters, skill at evaluating agent output, and the discipline to design sustainable rhythms rather than just rewarding whoever stays up latest.

Andrej Karpathy’s work on building structured knowledge bases that agents can actually query points at one piece of this — the context that agents operate on matters as much as the agents themselves. Better context, better outputs, less correction overhead.

The AutoResearch loop pattern is another piece: agents that can run experiments, measure results, and surface improvements autonomously reduce the judgment load on the human in the loop, because the evaluation is partially automated.

These aren’t just interesting technical patterns. They’re early answers to the coordination and evaluation constraints that make the infinite backlog feel overwhelming rather than exciting.

The backlog is real. The parallelism is real. The constraints are real. The question is whether you’re designing your work — and the tools you build — around all three of those facts simultaneously, or just the first two.

Sam Altman is losing sleep over it. That’s probably the right response.

Presented by MindStudio

No spam. Unsubscribe anytime.