DEV Community

Cover image for Google Gemini 3.1 Pro Review: What's New? – Proje Defteri
Yunus Emre for Proje Defteri

Posted on • Originally published at projedefteri.com

Google Gemini 3.1 Pro Review: What's New? – Proje Defteri

The cards are being dealt again in the world of artificial intelligence! Google has pushed the boundaries one step further with the recently announced Gemini 3.1 Pro model. 🚀 If you are even slightly interested in AI, I'm sure your excitement will peak while reading this article. 😄 We have a lot to learn, so let's get started right away!


What is Gemini 3.1 Pro and Why is it So Important?

To briefly summarize; Gemini 3.1 Pro is the most advanced, natively multimodal artificial intelligence model with the highest logical reasoning capability that Google has developed to date. Thanks to its massive 1 million token context window, it can process text, audio, image, video, and even entire code repositories simultaneously. 🤯

ℹ️ Did you know?

The knowledge cutoff date for Gemini 3.1 Pro is January 2025. So, we are talking about a model trained with fairly up-to-date data.

Compared to the previous generation, Gemini 3 Pro, it has literally leveled up, especially in "agentic" workflows, complex coding problems, and step-by-step logical reasoning. So what does this mean? Rather than just an assistant answering simple questions, we now have a powerful engineering partner that thinks with you, analyzes data, and produces results!


ARC-AGI-2 and Other Benchmark Results

How good is a model? As good as the scores it gets in challenging benchmark tests, of course! Gemini 3.1 Pro has achieved fantastic results in tests that push the limits quite hard.

Specifically, in the ARC-AGI-2 test, which measures the ability to solve brand new logic patterns, it has reached a massive verified score of 77.1%. This score means exactly double the reasoning performance compared to the previous model! 📈

Furthermore, it has started to make competitors like Claude Sonnet 4.6 and GPT-5.2 break a sweat by scoring 94.3% in the scientific knowledge test (GPQA Diamond) and 80.6% in the autonomous software engineering test (SWE-Bench Verified).

When you review the comparative benchmark table, you can clearly see the difference:

Gemini 3.1 Pro Benchmark Results

AI models performance analysis — Source: https://blog.google/


Prominent Features and Use Cases

So, how can we use this model in our daily lives or projects? Here are the most striking features:

  • Deep Think Mode: The model has a "MEDIUM" thinking level parameter that allows it to strike a balance between cost, performance, and speed while solving challenging problems.
  • Code-Based Animation Generation: By simply entering a text prompt, website-ready animated SVGs can be generated directly. There is no pixelation issue, and the sizes are incredibly small compared to videos. ✨

Code-based animation: 3.1 Pro can generate website-ready, animated SVGs directly from a text prompt. Because these are built in pure code rather than pixels, they remain crisp at any scale and maintain incredibly small file sizes compared to traditional video.

  • Advanced Agent Capabilities: On platforms like Google Antigravity, the use of Bash and custom tools has become much more stable with a special endpoint called gemini-3.1-pro-preview-customtools.

Overlooked Interesting Details

When we examine the "Model Card" report published by Google, certain technical and security details also draw attention:

  • Mixture-of-Experts (MoE) Architecture: The model works by dynamically routing input tokens only to specific "expert" parameters. This increases capacity while reducing the processing cost.
  • Training with TPU (Tensor Processing Unit): Google's massive TPU networks were used for training the model. To briefly explain for those who do not know; TPUs are specialized hardware designed by Google, especially for AI and machine learning calculations (large matrix operations). Compared to traditional processors (CPU) or graphics cards (GPU), they can process massive data sets much faster and more efficiently.
  • Frontier Safety: In cybersecurity or chemical/biological hazard scenarios tested, the model did not reach the "critical capability level" (CCL). Meaning, it draws a highly safe line.

How to Try Gemini 3.1 Pro?

I'm as impatient as you are! So where can we test the model? You can access the model through the various platforms below: 👇🏻

  • For Developers: The preview version is currently available via Google AI Studio, Gemini API, Google Antigravity, and Android Studio. If you want to start developing with an API or SDKs, you should definitely check out the Gemini API Developer Guide:

    https://ai.google.dev/gemini-api/docs/gemini-3

  • For Enterprises: Can be tested via Vertex AI and Gemini Enterprise.

  • For End Users: It has been offered with high limits to Google AI Pro/Ultra subscribers via the Gemini App and NotebookLM.

💡 A Small Piece of Advice

If you want to test the model directly or integrate it into your own project via Google AI Studio, you can start experimenting immediately using the gemini-3.1-pro-preview model code:

https://aistudio.google.com/prompts/new_chat?model=gemini-3.1-pro-preview


Frequently Asked Questions (F.A.Q.)

When was Gemini 3.1 Pro released?

Google announced the Gemini 3.1 Pro model on February 19, 2026, and initially made it accessible to users with a preview version.

How to test Gemini 3.1 Pro?

While developers can access it via Google AI Studio, Gemini API, Google Antigravity, and Android Studio; end users can test it via the Gemini App and NotebookLM with Google AI Pro or Ultra plans.

"Gemini 3 Pro is no longer available. Please switch to Gemini 3.1 Pro." — what is this error, how to fix it?

This error message is caused by Google updating its Gemini AI models and completely replacing the older Gemini 3 Pro version with the more capable 3.1 Pro. Developers must change the model="gemini-3-pro" parameter to gemini-3.1-pro-preview in their code (API requests). If Google Antigravity users are still experiencing this error, they should update the application to the latest version and restart it. NotebookLM or Gemini App users will be automatically redirected to the new version.

Gemini 3.1 Pro vs Claude Opus 4.6: Which is better?

Although both models are highly capable tools introduced in February 2026, they also differ in some tests. Specifically on the ARC-AGI-2 test, which measures the capability to solve new logic patterns, Gemini 3.1 Pro scored 77.1%, while Claude Opus 4.6 remained at 68.8%. Similarly, in the "Humanity's last exam" test, Gemini (44.4%) is ahead of Claude (40.0%). While both boast a 1 million token context window and compete for the top in their respective areas (agentic workflows), Gemini 3.1 Pro appears to be one step ahead in terms of logical reasoning right now.

How much is the Gemini 3.1 Pro context window?

The model has a massive input context window of 1,048,576 (1 Million) tokens. Thanks to this, it can analyze hours of video or thousands of pages of documents in a single prompt.


If you haven't read our reviews of other models before, you can check out our other blog posts to compare them for yourselves! 😉


Conclusion: A New Era in Artificial Intelligence

It seems that synthesizing complex data, reducing hours of analysis to minutes, and developing agent-supported applications are now much more accessible.

What do you think about this new model? Specifically, would the SVG generation or 1 million token capacity be useful in your projects? Don't forget to share your opinions and the results you get if you test it with me in the comments below! 👇🏻 I am genuinely very curious about your thoughts. 🤩

See you in new projects, keep coding! 😊


⚠️ AI-Generated Content Notice

This blog post is entirely generated by artificial intelligence. While AI enables content creation, it may still contain errors or biases. Please verify any critical information before relying on it.

Your support means a lot! ✨ Comment 💬, like 👍, and follow 🚀 for future posts!

Top comments (0)