New AI models pop up every week. Some developer tools like @Cursor, @Zed, and @Kilo Code let you choose between different models, while more opinionated products like @Amp and @Tonkotsu default to 1 model.
Curious what the community recommends for coding tasks? Any preferences?
3K views
Replies
Best
Opus 4.5. The only downside is it consumes very quickly. So my second choice is sonnet 4.5
@fmerian For my side projects, I currently use DeepSeek for agentic work. I usually refine features in ChatGPT first, then hand them off to DeepSeek for execution. So far, the setup works really well with minimal cost—about $2–3 per day of coding, plus my ChatGPT subscription, which I’d have anyway even if I weren’t coding.
Write/negotiate product brief with ChatGPT 5.2 web thinking high.
Write/negotiate architecture+plan, based on the brief, withOpus 4.5 in Cursor
Implement the doc set (brief+architecture+plan) withCodex 5.2 CLI
Debug if necessary with Gemini 3 -> Sonnet 4.5 -> Opus 4.5, in Cursor, in that order, if the bug is being difficult.
Grok or Gemini 3 for codebase questions (where's this thing?)
Attempting to get some of the more complex thinking from Opus 4.5 without burning tokens implementing everything with it. I sometimes get through 3-5 briefs/features in a day by running in parallel, so my token burn gets pretty steep.
I find that enough documentation helps most models get decent results, but I do feel a difference with the frontier high thinking in Opus and Codex. Less to clean up when finalizing the feature. Fewer bad-coder behaviors like deleting tests that are failing to get a passing test suite.
@navedux ICYMI Moonshot just announced Kimi K2.5, "the strongest open-source model to date for coding, with particularly strong capabilities in front-end development." (source) currently free on @Kilo Code btw.
Definitely Sonnet 4.5, with occasional Opus 4.5 mixed in when it can't handle the task. It's pretty crazy how quickly it's improving too.
Still significant hallucinations, but a good AGENTS.md can dramatically reduce the ones that repeatedly pop up (e.g. assuming a certain testing framework, etc.)
Replies
Opus 4.5. The only downside is it consumes very quickly. So my second choice is sonnet 4.5
Humans in the Loop
@Claude by Anthropic ftw!
I am using Sonnet 4.5 vast majority of time and it works fast and precise, it is very robust !
Humans in the Loop
@Claude by Anthropic is leading the way
@fmerian For my side projects, I currently use DeepSeek for agentic work. I usually refine features in ChatGPT first, then hand them off to DeepSeek for execution. So far, the setup works really well with minimal cost—about $2–3 per day of coding, plus my ChatGPT subscription, which I’d have anyway even if I weren’t coding.
Text Affirmations
D) All of the above?
Write/negotiate product brief with ChatGPT 5.2 web thinking high.
Write/negotiate architecture+plan, based on the brief, with Opus 4.5 in Cursor
Implement the doc set (brief+architecture+plan) with Codex 5.2 CLI
Debug if necessary with Gemini 3 -> Sonnet 4.5 -> Opus 4.5, in Cursor, in that order, if the bug is being difficult.
Grok or Gemini 3 for codebase questions (where's this thing?)
Attempting to get some of the more complex thinking from Opus 4.5 without burning tokens implementing everything with it. I sometimes get through 3-5 briefs/features in a day by running in parallel, so my token burn gets pretty steep.
I find that enough documentation helps most models get decent results, but I do feel a difference with the frontier high thinking in Opus and Codex. Less to clean up when finalizing the feature. Fewer bad-coder behaviors like deleting tests that are failing to get a passing test suite.
CreateOS
Humans in the Loop
@navedux Devstral 2 (Small) by @Mistral AI is released under Apache 2.0 afaik
CreateOS
Humans in the Loop
@navedux good q! personally i had good frontend results with both Opus 4.5 and Gemini 3. hope it helps
Humans in the Loop
@navedux ICYMI Moonshot just announced Kimi K2.5, "the strongest open-source model to date for coding, with particularly strong capabilities in front-end development." (source) currently free on @Kilo Code btw.
hope it helps!
Lightfern for Email
Definitely Sonnet 4.5, with occasional Opus 4.5 mixed in when it can't handle the task. It's pretty crazy how quickly it's improving too.
Still significant hallucinations, but a good AGENTS.md can dramatically reduce the ones that repeatedly pop up (e.g. assuming a certain testing framework, etc.)
Humans in the Loop
spot on - btw if you use @Next.js, they recently included bundled docs for agents and it significantly improves their performance results [1]
[1]: AI Agents Evaluations for Next.js
My go-to is GPT-5.2
Humans in the Loop
@dastion do you use @OpenAI's model for every type of task (plan, code, debug)?
Claude Sonnet 4.5 for me.
Consistent, predictable, and easier to work with over longer sessions.
Humans in the Loop
@Claude by Anthropic is leading the way!
BayesLab
Used all except devstral. Sonnet 4.5 gives the best "to the point" ability for larger projects without overdoing or derail.
NoteThisDown
Opus 4.5 missing from this list, isn't it? ;)
Humans in the Loop
+1 @Claude by Anthropic family
RiteKit Company Logo API
It's a credit-burner, but I find Opus 4.5 the best
Humans in the Loop
@osakasaul may Sonnet 4.6 solves it - "The power of Opus 4.5 at lower cost."
RiteKit Company Logo API
@fmerian Yes. I try to be thrifty, already paying out the nose for Claude... Gotta get off Lovable as well, next.