fmerian

What's the best AI model for coding?

by

New AI models pop up every week. Some developer tools like @Cursor, @Zed, and @Kilo Code let you choose between different models, while more opinionated products like @Amp and @Tonkotsu default to 1 model.

Curious what the community recommends for coding tasks? Any preferences?

3K views

Add a comment

Replies

Best
Asad Iqbal
Opus 4.5 and at times codex to review CC
Nikita Ivanov

From my experience Opus is really good at generating code for a specific framework or service, but for everyday tasks like refactoring, writing helpers, or SQL queries, GPT-5.2 ends up being faster and cheaper

Overall you pick the model for the task, not the other way around

Siarhei
@fmerian Thanks for the upvote! 🙌 For coding: Claude 3.5/Opus still crushes complex logic, but new ones like Gemini CLI gaining speed. What’s your daily driver in 2026? (I’m fine-tuning voice models for HireXHub — always open to recs!)
John Brush

Nate B. Jones has an interesting piece on his channel where he compares models to different machining tools: Codex being like a CNC machine and Claude like a craftsman with a lathe:

https://youtu.be/hDpjMJw3flk?t=229

His point is that we are starting to see fundamental differences in development philosophies and corresponding tooling. Personally, I think he's on point with this.

As a developer/architect with years of experience in building systems, I'm more in the CNC machine / specification camp. But there is a case to be made for just jumping in and starting without much of a plan (but there's also a case to be made for spending more time figuring out what you're actually trying to do before you code).

Anyhoo, my two cents...

Alon Hamudot

I really liked working with Gemini 3 on all things related to UX/UI but when it comes to logic Opus 4.5 is the king.

This is how I created @NMTV

Husam

We’ve been using CC Opus 4.5 and it’s been solid for how we work. It really helps when we’re dealing with heavier thinking like product logic, edge cases, or those moments where you’re trying to understand what breaks if one decision changes. It’s not the fastest, but it saves us from making expensive mistakes.

For quick iteration or speed stuff, I understand why people lean toward faster models. For us, we don’t fight it. If the task is mechanical, tests, small refactors, glue code, we switch to something faster. Using Opus for that is a waste.

Brad Shannon

Opus 4.5 is the king, but expensive. GPT 5.2 is probably the 2nd best on this list. Sonnet 4.5 is great still, but getting old.

fmerian

Sonnet 4.5 is great still, but getting old.

Sonnet 4.5 launched 4 months ago, and yet, you're so right - it's getting old

Mykola Kondratiuk

Claude (Sonnet 3.5/3.6) for speed, Opus for complex architecture decisions. I've shipped 4 products as a weekend vibe coder over the last 3 months using almost exclusively Claude Code.

The thing nobody mentions: the model matters less than your prompting discipline. Clear specs, small scoped tasks, reviewing every output. I write maybe 30% of the code myself but review 100%. The security gap is real though - AI-generated code loves to skip input validation and auth edge cases.

For pure speed: Sonnet. For getting things right the first time on tricky stuff: Opus. GPT is fine but Claude just "gets" code structure better in my experience.

Jonathan Song

Been experimenting with different models for a conversational AI project I'm building. Here's what I've learned:

Context management is everything. The "best" model really depends on your use case:

- Complex refactoring? Opus 4.5 hands down. Worth the cost when you need deep reasoning.

- Quick iterations/prototyping? Sonnet 4.5 hits the sweet spot - fast enough to stay in flow, smart enough to handle most tasks.

- Frontend/UI work? Gemini Pro surprised me with speed and quality.

The real game-changer isn't just the model though - it's how you structure your prompts and manage context. I've found that keeping a clean git history and feeding the model focused diffs (not entire repos) makes a huge difference regardless of which model you use.

Also learned the hard way: don't let context bloat past 40-50%. Quality drops fast after that.

fmerian

this 👆

Abdul Rehman

Still leaning toward Sonnet for coding — feels the most reliable for complex logic so far 👍