Vellum is coming to the AI Engineering World's Fair in SF. Come visit our booth and get a live demo!

Google Gemini 3 Benchmarks

A deep dive into Google's latest model performance

8 min
Written by
Reviewed by
No items found.

After being leaked on LLM Arena, Gemini 3 Pro is finally here! With a 1M token context window and a 64K output window, Google's latest model is making a strong claim for the top spot in the AI landscape.

It's available today through the Gemini API, Google AI Studio, and Vertex AI. What used to be 2025’s model provider underdog is now a first in class AI model, dominating key benchmarks Gemini 2.5 couldn’t hold a candle to.

It’s currently pushing the boundaries on math, reasoning, multimodal understanding, long-context performance, and multilingual capabilities. Let's take a look at the numbers.

💡 Want to see how Gemini 3 compares to GPT-5, Claude 4.5, Grok 4.1 for your use case? Compare them in Vellum.

Key observations of reported benchmarks

While benchmarks are inherently limited and may not fully capture real-world utility, it’s our only way to measure progress. From the reported data we can conclude a few things:

  1. Reasoning: The most compelling data points are the high scores on ARC-AGI-2 (31.1% / 45.1% with Deep Think) and Humanity’s Last Exam (37.5% / 40%+ with Deep Think). The high score on GPQA Diamond (91.9%) confirms state-of-the-art performance on PhD-level scientific knowledge, though this benchmark is approaching saturation.
  2. Math: AIME 2025 feels saturated, but a >20x jump in results for MathArena Apex suggest Gemini 3 has great intrinsic reasoning base for mathematical logic and problem formulation (even without using coding tools)
  3. Multimodality: High scores in both MMMU-Pro (81.0%) and Video-MMMU (87.6%). This suggest a great ability to process and reason across temporal and spatial dimensions simultaneously. This makes it highly effective for integrated tasks like analyzing video lectures or complex UI screenshots (which is confirmed by its high score on the ScreenSpot-Pro benchmark, not listed above).
  4. Agentic Reliability: The results on Vending-Bench 2, where Gemini 3 Pro’s mean net worth is $5,478.16 (272% higher than GPT-5.1), are arguably the most indicative of practical utility. This benchmark measures long-horizon planning, coherent decision-making, and consistent tool usage over a simulated year.

Reasoning Capabilities

Reasoning benchmarks evaluate a model's ability to solve complex problems.

GPQA Diamond assesses PhD-level scientific knowledge, while ARC-AGI-2 focuses on abstract visual puzzles that resist simple memorization. We also benchmark these models on the Humanity’s Last exam which is designed to push AI to its limits across a wide array of subjects.

  • Gemini 3 Pro scores 91.9% on GPQA Diamond (and 93.8% with Deep Think), giving it a nearly 4-point lead over GPT-5.1 (88.1%) on advanced scientific questions.
  • The most notable upgrade is in abstract visual reasoning. Its 31.1% score on ARC-AGI-2 (45.1% with Deep Think) is a massive jump from Gemini 2.5 Pro (4.9%) and nearly doubles the score of GPT-5.1 (17.6%), indicating a core improvement in non-verbal problem-solving.
  • Most importantly this is the first time we’re seeing a 40%+ result with deep think, and 37.5% with Gemini 3 Pro, on the hardest reasoning test, the Humanity’s Last Exam, which is almost 11% increase from GPT 5.1.

Math Capabilities

The AIME 2025 benchmark, based on a challenging math competition, tests a model's quantitative reasoning skills. Performance is measured both with and without the assistance of code execution tools.

  • With code execution, Gemini 3 Pro achieves a perfect 100% score, matching the top performance of GPT-5.1.
  • The key differentiator is its performance without tools, where it scores 95.0%. This strong base performance shows a more robust innate mathematical intuition, making it less dependent on external tools to find the correct solution.
  • Gemini 3 Pro is currently the only one that looks somewhat capable on MathArena Apex, but the task is still far from “solved” for any model. But it’s still a jump of >20x on one of the hardest reasoning tasks we have.

Coding Capabilities

LiveCodeBench Pro evaluates performance on competitive coding problems, while SWE-Bench measures a model's ability to resolve real-world software issues from GitHub repositories.

  • Its 76.2% on SWE-Bench for bug fixing is highly competitive with Claude Sonnet 4.5 (77.2%), but its dominance in algorithmic challenges suggests it is particularly well-suited for complex, from-scratch code generation.
  • Gemini 3 Pro shows a commanding lead in algorithmic problem-solving, achieving an Elo rating of 2,439 on LiveCodeBench Pro. This is nearly 200 points higher than GPT-5.1 (2,243), indicating superior skill in generating novel and efficient code.

Long Context Capabilities

A large context window's value depends on the model's ability to accurately retrieve information. The MRCR v2 benchmark tests this "needle-in-a-haystack" capability within a large volume of text.

  • Gemini 3 Pro demonstrates strong recall, scoring 77.0% on the 128k average context length test.
  • For larger context, 1M, Gemini 3 Pro outpaces the Gemini 2.5 Pro model by 9.9%.

Multimodal Capabilities

Natively multimodal models are assessed on their ability to understand and reason across different data types simultaneously. MMMU-Pro and Video-MMMU are key benchmarks for this integrated understanding.

  • On MMMU-Pro, Gemini 3 Pro scores 81.0%, creating a significant 5-point gap ahead of GPT-5.1 (76.0%) in multimodal understanding and reasoning.
  • Its 87.6% score on Video-MMMU shows its strength is not limited to static images. This high performance demonstrates an advanced ability to comprehend and synthesize information from dynamic video content.

Multilingual Capabilities

These two benchmarks, the MMMLU and Global PIQA, test a model's performance beyond simple translation, evaluating its understanding of cultural context and logic across many languages.

  • Gemini 3 Pro takes the top spot on MMMLU (Multilingual Q&A) with a score of 91.8%, slightly ahead of GPT-5.1 (91.0%).
  • More significantly, its 93.4% on Global PIQA, which tests commonsense reasoning across 100 languages, suggests a deeper, culturally-aware understanding rather than just linguistic proficiency.

Long-Horizon Planning and Agentic Skills

Beyond single-turn tasks, a model's ability to plan and execute multi-step workflows is a critical measure of its agentic capabilities. Benchmarks like Vending-Bench 2 test this by simulating complex, long-term goals.

  • Gemini 3 Pro tops the leaderboard on Vending-Bench 2, a benchmark that simulates managing a vending machine business for a full year. This tests its ability to make strategic decisions over a long period.
  • The key insight is its consistency. Gemini 3 Pro maintains reliable tool usage and decision-making throughout the entire simulation, preventing it from drifting off-task and ultimately generating higher returns than other models.
  • This improved planning translates to practical, real-world assistance. It can more reliably navigate complex workflows like booking local services or organizing an inbox from start to finish under user guidance.

What these benchmarks really mean for your agents

Gemini 3 marks a clear step forward in the frontier of agentic systems. By raising the bar across every benchmark that actually translates to operational performance, it expands what your agents can reliably automate. The model is here and the capabilities are real. Now it’s on you to start building the agents that turn this progress into compounding efficiency and ROI.

{{general-cta}}

ABOUT THE AUTHOR
Nicolas Zeeb
Technical Content Lead

Nick is Vellum’s technical content lead, writing about practical ways to use both voice and text-based agents at work. He has hands-on experience automating repetitive workflows so teams can focus on higher-value work.

ABOUT THE reviewer
Anita Kirkovska
Founding Growth Lead

An AI expert with a strong ML background, specializing in GenAI and LLM education. A former Fulbright scholar, she leads Growth and Education at Vellum, helping companies build and scale AI products. She conducts LLM evaluations and writes extensively on AI best practices, empowering business leaders to drive effective AI adoption.

No items found.
lAST UPDATED
Nov 18, 2025
share post
Expert verified
Related Posts
November 11, 2025
15 min
AI Agent Use Cases Guide to Unlock AI ROI
LLM basics
November 6, 2025
7 min
Beginners Guide to Building AI Agents
Product Updates
November 5, 2025
7 min
Vellum Product Update | October
All
November 3, 2025
6 min
I’m done building AI agents
Guides
October 21, 2025
15 min
AI transformation playbook
LLM basics
October 20, 2025
8 min
The Top Enterprise AI Automation Platforms (Guide)
The Best AI Tips — Direct To Your Inbox

Latest AI news, tips, and techniques

Specific tips for Your AI use cases

No spam

Oops! Something went wrong while submitting the form.

Each issue is packed with valuable resources, tools, and insights that help us stay ahead in AI development. We've discovered strategies and frameworks that boosted our efficiency by 30%, making it a must-read for anyone in the field.

Marina Trajkovska
Head of Engineering

This is just a great newsletter. The content is so helpful, even when I’m busy I read them.

Jeremy Hicks
Solutions Architect

Experiment, Evaluate, Deploy, Repeat.

AI development doesn’t end once you've defined your system. Learn how Vellum helps you manage the entire AI development lifecycle.

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Build AI agents in minutes with Vellum
Build agents that take on the busywork and free up hundreds of hours. No coding needed, just start creating.

General CTA component, Use {{general-cta}}

Build AI agents in minutes with Vellum
Build agents that take on the busywork and free up hundreds of hours. No coding needed, just start creating.

General CTA component  [For enterprise], Use {{general-cta-enterprise}}

The best AI agent platform for enterprises
Production-grade rigor in one platform: prompt builder, agent sandbox, and built-in evals and monitoring so your whole org can go AI native.

[Dynamic] Ebook CTA component using the Ebook CMS filtered by name of ebook.
Use {{ebook-cta}} and add a Ebook reference in the article

Thank you!
Your submission has been received!
Oops! Something went wrong while submitting the form.
Button Text

LLM leaderboard CTA component. Use {{llm-cta}}

Check our LLM leaderboard
Compare all open-source and proprietary model across different tasks like coding, math, reasoning and others.

Case study CTA component (ROI) = {{roi-cta}}

40% cost reduction on AI investment
Learn how Drata’s team uses Vellum and moves fast with AI initiatives, without sacrificing accuracy and security.

Case study CTA component (cutting eng overhead) = {{coursemojo-cta}}

6+ months on engineering time saved
Learn how CourseMojo uses Vellum to enable their domain experts to collaborate on AI initiatives, reaching 10x of business growth without expanding the engineering team.

Case study CTA component (Time to value) = {{time-cta}}

100x faster time to deployment for AI agents
See how RelyHealth uses Vellum to deliver hundreds of custom healthcare agents with the speed customers expect and the reliability healthcare demands.

[Dynamic] Guide CTA component using Blog Post CMS, filtering on Guides’ names

100x faster time to deployment for AI agents
See how RelyHealth uses Vellum to deliver hundreds of custom healthcare agents with the speed customers expect and the reliability healthcare demands.
New CTA
Sorts the trigger and email categories

Dynamic template box for healthcare, Use {{healthcare}}

Start with some of these healthcare examples

Population health insights reporter
Combine healthcare sources and structure data for population health management.
Clinical trial matchmaker
Match patients to relevant clinical trials based on EHR.

Dynamic template box for insurance, Use {{insurance}}

Start with some of these insurance examples

Insurance claims automation agent
Collect and analyze claim information, assess risk and verify policy details.
Agent that summarizes lengthy reports (PDF -> Summary)
Summarize all kinds of PDFs into easily digestible summaries.
AI agent for claims review
Review healthcare claims, detect anomalies and benchmark pricing.

Dynamic template box for eCommerce, Use {{ecommerce}}

Start with some of these eCommerce examples

E-commerce shopping agent
Check order status, manage shopping carts and process returns.

Dynamic template box for Marketing, Use {{marketing}}

Start with some of these marketing examples

LinkedIn Content Planning Agent
Create a 30-day Linkedin content plan based on your goals and target audience.
Competitor research agent
Scrape relevant case studies from competitors and extract ICP details.

Dynamic template box for Sales, Use {{sales}}

Start with some of these sales examples

Research agent for sales demos
Company research based on Linkedin and public data as a prep for sales demo.

Dynamic template box for Legal, Use {{legal}}

Start with some of these legal examples

Legal document processing agent
Process long and complex legal documents and generate legal research memorandum.
AI legal research agent
Comprehensive legal research memo based on research question, jurisdiction and date range.

Dynamic template box for Supply Chain/Logistics, Use {{supply}}

Start with some of these supply chain examples

Risk assessment agent for supply chain operations
Comprehensive risk assessment for suppliers based on various data inputs.

Dynamic template box for Edtech, Use {{edtech}}

Start with some of these edtech examples

Turn LinkedIn Posts into Articles and Push to Notion
Convert your best Linkedin posts into long form content.

Dynamic template box for Compliance, Use {{compliance}}

Start with some of these compliance examples

No items found.

Dynamic template box for Customer Support, Use {{customer}}

Start with some of these customer support examples

Trust Center RAG Chatbot
Read from a vector database, and instantly answer questions about your security policies.
Q&A RAG Chatbot with Cohere reranking

Template box, 2 random templates, Use {{templates}}

Start with some of these agents

Agent that summarizes lengthy reports (PDF -> Summary)
Summarize all kinds of PDFs into easily digestible summaries.
Healthcare explanations of a patient-doctor match
Summarize why a patient was matched with a specific provider.

Template box, 6 random templates, Use {{templates-plus}}

Build AI agents in minutes

AI legal research agent
Comprehensive legal research memo based on research question, jurisdiction and date range.
Retail pricing optimizer agent
Analyze product data and market conditions and recommend pricing strategies.
ReAct agent for web search and page scraping
Gather information from the internet and provide responses with embedded citations.
Agent that summarizes lengthy reports (PDF -> Summary)
Summarize all kinds of PDFs into easily digestible summaries.
Research agent for sales demos
Company research based on Linkedin and public data as a prep for sales demo.
Review Comment Generator for GitHub PRs
Generate a code review comment for a GitHub pull request.

Build AI agents in minutes for

{{industry_name}}

Clinical trial matchmaker
Match patients to relevant clinical trials based on EHR.
Prior authorization navigator
Automate the prior authorization process for medical claims.
Population health insights reporter
Combine healthcare sources and structure data for population health management.
Legal document processing agent
Process long and complex legal documents and generate legal research memorandum.
Legal contract review AI agent
Asses legal contracts and check for required classes, asses risk and generate report.
Legal RAG chatbot
Chatbot that provides answers based on user queries and legal documents.

Case study results overview (usually added at top of case study)

What we did:

1-click

This is some text inside of a div block.

28,000+

Separate vector databases managed per tenant.

100+

Real-world eval tests run before every release.