DEV Community

Cover image for 🤖 OpenAI o3 vs. Gemini 2.5 vs. OpenAI o4-Mini on Coding 🤔

🤖 OpenAI o3 vs. Gemini 2.5 vs. OpenAI o4-Mini on Coding 🤔

Shrijal Acharya on April 22, 2025

TL;DR If you want to skip to the conclusion, here’s a quick summary of the findings comparing OpenAI o3, Gemini 2.5, and OpenAI o4-Mini:...
Collapse
 
acoliver profile image
Andrew C. Oliver

Interesting. I've had better luck with o4-mini and o3 for coding. Granted my work is very iterative and human in the loop. Gemini tends to make a mess of things for me after round 1. o4-mini is my goto but o3 is for debugging. I still use Claude a lot for the tradeoff of speed.

My expectations are lower though. Step one anything that runs and is roughly a game is pretty good. I'd expect to iterate more with more clarifying prompts. The fully autonomous thing seems like a pipe dream for now for anything beyond "implement this API"

Collapse
 
shekharrr profile image
Shekhar Rajput

Nicely said. I agree with you. o3 seems to be more better with debugging stuff.

Collapse
 
arindam_1729 profile image
Arindam Majumder

NIce Comparison, Shrijal!

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Appreciate you, Arindam 🫡

Collapse
 
bh4skar profile image
Bhaskar Prajapati

Im loving all the hype around these AI models nowadays. Gemini 2.5 seems to be the one to go for.

Collapse
 
n1ckn1ne911 profile image
Nick Nine

It's the worst. Always hallucinates on a bigger codebase.

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Does it? Haven't really used it on a bigger codebase. What's your typical usecase?

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Not always. As said, choose the models based on your use case.

Collapse
 
arindam_1729 profile image
Arindam Majumder

Nice Comparison, Shrijal!

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Appreciate you, Arindam 🫡

Collapse
 
arya_joy_a628d0a39511bd03 profile image
Arya Joy

OpenAI O3 handles complex code well, while Gemini penumbra ffxiv 2.5 is strong in logic and structure.
O4-Mini is fast and efficient great for everyday coding tasks on the go

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Well said!

Collapse
 
n1ckn1ne911 profile image
Nick Nine

Love the zero-shot concept for testing. This really tests raw performance. But I don't think Gemini is a good choice for bigger codebase. Maybe works for such smaller projects. Go for o1, that's more than enough for most of the use case. Jack of all.

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Glad you liked it. I've heard good things about this model o1, and it definitely sounds great. I haven't tried it, though…

Collapse
 
nevodavid profile image
Nevo David Composio

Honestly, I get super annoyed when I can't just pick one model for everything - feels like it's always tradeoffs. Makes me double check every AI response now, not gonna lie.

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Yeah, same for me, though. Thank you for checking out, Nevo! :)

Collapse
 
zako_mako_9a4826822204c78 profile image
Zako Mako

I WANT TO JOIN

Collapse
 
shricodev profile image
Shrijal Acharya Composio

What exactly? 👀

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Share your thoughts on the comments! 👇

Collapse
 
larastewart_engdev profile image
Lara Stewart - DevOps Cloud Engineer

Love this detailed comparison, Shrijal! 🤍

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Thank you, Lara! I appreciate you taking the time to check out the blog post. 😊

Collapse
 
satriaar profile image
Satria Akbar Rizki

Wow amazing but deepseek why include for compare response?

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Thank you, Satria, but where did I mention DeepSeek exactly?

Collapse
 
sawata_2_shayne profile image
Shayne Villarin

Am I the only one sticking with codellama still? It just gets my job done. Never had to look back.

Collapse
 
shricodev profile image
Shrijal Acharya Composio

I'm surprised. How do you use it? 👀 Maybe locally with Ollama?

Collapse
 
sawata_2_shayne profile image
Shayne Villarin

lol yeah, local setup. For vscode, I use an extension. Here you can find it: marketplace.visualstudio.com/items...

Thread Thread
 
shricodev profile image
Shrijal Acharya Composio

🙌

Collapse
 
aavash_parajuli_72 profile image
Aavash Parajuli

Love this.

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Thanks!

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Thank you! I agree that the recent new smaller models are really great. I love the Gemini 1.5 flash and also the recent pretty small and solid model, Gemma 3 27B. Have you tried it out?

Collapse
 
ricorizz profile image
Rico Rizz

Impressive!

Collapse
 
shricodev profile image
Shrijal Acharya Composio

Thank you, @ricorizz ✌️