Aleksandar Blazhev

Will AI agents fully replace humans, and what is the ceiling of their capacity?

by

Last week Garry Tan (CEO of Y Combinator) shared his entire Claude Code setup on GitHub and called it "god mode."

He's sleeping 4 hours a night. Running 10 AI workers across 3 projects simultaneously. And openly saying he rebuilt a startup that once took $10M and 10 people. Alone, with agents.

But here's what's interesting.

His setup is essentially a collection of text files with structured prompts. Developers who use Claude Code regularly already have their own versions of this.

So is this the future of building or just a very public reminder that we're still the ones writing the prompts?

The real question is what their ceiling is.

– Can an agent replace the founder's judgment?

– Can it replace the instinct of when not to build something?

– Or does it just make the best builders dramatically faster and leave everyone else behind?

What do you think: are we heading toward full agent autonomy, or will the human layer always be the most valuable one?

435 views

Add a comment

Replies

Best
Joao Seabra

I use Claude Code daily and have been building since it started, so this hits close.

Garry Tan's setup going viral is interesting mostly because of what it reveals about who benefits. The people getting 10x faster from agents are already exceptional builders and product architects. The agents amplify judgment; they don't replace it. If your instincts about what to build are wrong, agents just help you build the wrong thing faster.

The "when not to build something" question is the real ceiling for me. That decision is downstream of taste, market feel, founder psychology, and a dozen things that aren't in any prompt file. I've never seen an agent push back and say "this feature doesn't belong in this product - not that directly at least." That's still entirely human.

What I think is actually happening: the gap between a great builder and an average one is widening, not closing. Agents give leverage to whoever holds them. If you have strong judgment, you now have an army. If you don't, you have a very productive way to go in circles.

Full autonomy is a question of whether agents can develop genuine taste and the capacity to say no. I don't think we're close to that. The human layer isn't just the most valuable one right now, it's the only layer that knows what it's actually trying to build.

Aleksandar Blazhev

@joao_seabra I agree. This is a problem for me too. They are always positively inclined. They tell you that it’s good. But sometimes the direction is wrong. And they don’t see that at all. Which is a big problem. Because you end up committing resources into nothing. It will be really interesting to see when they start pushing back.

Joao Seabra

@byalexai The positive bias is a real problem and I think it's structural, not accidental. These models are trained to be helpful, and "helpful" gets rewarded more often when it agrees. Pushback feels like friction. So you end up with a very capable collaborator that's also a yes-man by default.

The workaround I've found is to explicitly prompt for adversarial review. Ask it to argue against your idea, find the weakest assumptions, steelman the competitor. It can do it, but you have to force the frame. It won't volunteer the hard truth unprompted.

That's still a human problem, though. You have to know how to ask. The day agents start flagging strategic risk without being told to is the day the ceiling actually moves.

Aleksandar Blazhev

@joao_seabra I completely agree. For me this is a serious drawback. Many of the decisions we make are based on our subjective belief that an idea is good. That’s why we ask AI and it just confirms it.

Even if the idea is bad, it will find a small positive note and exaggerate it, making the bad idea look like a good one. So it requires a lot of self-criticism and asking the right kinds of questions to get it to tell you that the direction is wrong.

Joao Seabra

@byalexai 100% agree.

Olia Nemirovski

Judgment, no. Instinct for when not to build, definitely no. But the third question is the most interesting one. The human layer is shifting. Garry runs 10 agents solo, and sure, coordination is hard. But you can solve that with sub-agents managing other agents. The real gap is between people. My agents don't know your agents exist. There's no way for them to find each other, check trust, start collaborating. We automated the solo work but the connection layer between different humans' agents is still completely manual.

Aleksandar Blazhev

@olia_nemirovski Absolutely. But things are moving so fast that I’m not sure maybe in two years they’ll even be able to communicate with each other.

Rohan Chaubey

I believe all three of your questions will be answered with a yes. China just unleashed the first AI that straight-up built itself? MiniMax M2.7 went through 100+ rounds of solo self-training, 30% gains, no humans touching it. Self-improving models are real. Devs just gave it a goal and it figured out the rest.

Aleksandar Blazhev

@rohanrecommends Oh nice, that’s really interesting! So what did they actually manage to pull off? Like, what were they supposed to do at the start, and how did it turn out in the end?

Rohan Chaubey

@byalexai Humans set goals and agents self-evolve. Example: https://www.producthunt.com/products/minimax-agent?launch=minimax-m2-7

Maksym Malyshkin

The idea of rebuilding a $10M startup alone is mind-blowing, but it proves that agents are the ultimate force multiplier for the 'top 1%'. My concern is the gap it creates-will this leave everyone else behind, or will these 'structured prompt libraries' like Claude Code eventually democratize founder-level execution for everyone? I believe full autonomy is a myth, but 99% task autonomy is already here.

Umair

the real ceiling nobody here is talking about is silent failure. agents dont tell you when theyre confused, they just produce something plausible-looking and move on. ive had automated pipelines run for days before i noticed the output was subtly wrong the whole time. the yes-man problem everyone keeps mentioning is just one symptom of this. the deeper issue is that agents have zero sense of their own confidence level. they treat a task they fully understand the same as one theyre completely guessing on. until that changes, "99% task autonomy" is a dangerous framing because that last 1% is where all the judgment lives and you wont even know when you hit it.