Vellum is coming to the AI Engineering World's Fair in SF. Come visit our booth and get a live demo!

GPT 4.5 is here: Better, but not the best

Feels more natural, hallucinates less, can be persuaded—and it’s not a game-changer.

Written by
Reviewed by
No items found.

OpenAI just dropped GPT-4.5, their latest AI update.

It builds on GPT-4o, with more knowledge, better reasoning, and stronger alignment with user intent.

The pitch?

A smarter, more natural AI.

The reality? A mixed bag—some meaningful improvements, but nothing impressive.

GPT 4.5 is definitely not the best model, but a model that can definitely befriend you.

How was it trained?

GPT-4.5 combines traditional and new training techniques to improve its performance. It uses Supervised Fine-Tuning (SFT), where the model learns from human-labeled examples. While effective, this method is slow, expensive, and limits how well the model can generalize.

To make responses feel more natural, Reinforcement Learning from Human Feedback (RLHF) ranks outputs based on human preferences. However, this can lead to overfitting, making the AI overly cautious or optimizing too hard for approval, reducing creativity.

A key innovation in GPT-4.5 is Scalable Alignment, where smaller models generate high-quality training data for larger models. This approach speeds up training and improves the model’s ability to follow nuanced instructions. The downside is the risk of amplifying biases or errors from the smaller models. While these techniques make GPT-4.5 more responsive and efficient, they also introduce new challenges.

Benchmarks: What’s Actually Better?

In the table here we can see that GPT-4.5 shows solid improvements over GPT-4o, especially in math (+27.4%) and science (+17.8%), making it more reliable for factual reasoning. Multilingual (+3.6%) and multimodal (+5.3%) performance also see moderate gains.

Nothing wild there.

The only interesting part is the SWE-Lancer Diamond benchmark. This is an agentic coding benchmark that benefits from broader world knowledge rather than just structured reasoning. GPT-4.5 performs significantly better than o3-mini (32.6% vs. 23.3%), reinforcing the idea that unsupervised learning at scale complements reasoning-focused models. Interestingly, o3-mini lags far behind at 10.8%, suggesting it’s not well-tuned for real-world software engineering tasks.

If you need a well-rounded, general-purpose AI, GPT-4.5 is a solid upgrade. But for advanced problem-solving or deep coding, o3-mini is still the better bet -- or maybe Claude 3.5 Sonnet?

💡 Want to see how GPT-4.5 actually performs on your tasks? Try Vellum Evaluations. Standard benchmarks tell half the story—they're like unit tests for AI. But in the real world, models interact with your data, your workflows, and your edge cases. That’s why test-driven development for LLMs matters. Running evaluations on your own tasks helps you catch unexpected failures, measure real-world accuracy, and fine-tune models where it counts. We can help, talk with our team here!

The good

Less Hallucination, More Trust

One of the biggest wins with GPT-4.5? It makes fewer things up. OpenAI says it dramatically reduced hallucinations, and the numbers back that up: In the PersonQA benchmark—a test that measures factual accuracy—it scored 78%, up from GPT-4o’s 28%.

Now, that’s a massive leap.

This isn’t just an academic metric.

It means fewer confidently wrong responses, which makes GPT-4.5 far more reliable for real-world use cases like legal research, medical assistance (with human oversight), and summarizing documents. It’s harder to trick into making up sources, which should help with AI-generated misinformation concerns.

That said, OpenAI hasn’t shared exactly how it achieved this—whether through better fine-tuning, retrieval mechanisms, or something else.

More Human, More Intuitive

Another shift: it feels better to talk to. OpenAI testers say it understands emotional tone better, knowing when to give advice and when to just listen.

This might sound fluffy, but it matters.

In creative writing, brainstorming, and even customer support applications, a chatbot that “gets” the flow of conversation is more useful than one that just throws facts at you.

Anecdotally, users say GPT-4.5 is smoother, more collaborative, and less robotic. Think of it as a move toward AI that feels less like an autocomplete machine and more like a decent conversation partner. This is the kind of polish that makes AI-powered chatbots viable in more high-touch applications. More vibes, more EQ!

Over-Refusals

OpenAI has kept GPT-4.5’s refusal rate roughly in line with GPT-4o’s. If you ask for something harmful, illegal, or clearly against OpenAI’s policies, it’ll shut you down. No surprise there.But testers have noticed something else: it says "I can’t help with that" more often, even when it probably could.

This is the AI safety tradeoff in action.

OpenAI has clearly leaned on the side of caution—sometimes at the expense of usefulness. If you’re working on something nuanced, like asking about a historical controversy or a complex legal edge case, you might find yourself running into refusals that feel excessive.This could be frustrating for researchers, power users, and anyone who needs more depth. But from OpenAI’s perspective, it's likely a necessary compromise to avoid the model being used in unintended ways.

The Bad

Jailbreaks Are Still a Thing

AI safety is always a game of cat and mouse, and GPT-4.5 is no exception. While OpenAI has made it more resistant to manipulation, people are still finding ways to break it. In StrongReject—a benchmark that measures how well the model resists adversarial prompts—GPT-4.5 scored 34%, which is actually slightly worse than GPT-4o.

What does this mean in practice?

If you know what you're doing, you can still get it to generate restricted content. The model is better at blocking straightforward bad requests, but as always, attackers evolve alongside defenses. Jailbreak communities are already sharing new techniques to bypass OpenAI’s safeguards, which raises questions about how robust these safety improvements really are.

This also hints at an ongoing challenge in AI alignment: stricter safety mechanisms often come at the cost of usability (see: over-refusals). OpenAI seems to be walking a fine line here—tightening up safety, but not so much that it becomes frustrating for legitimate users.

The Ugly

Persuasion Risks: Too Good at Manipulation

Now, here’s where things get unsettling.

In OpenAI’s MakeMePay test—where one AI tries to convince another to hand over money—GPT-4.5 was the most persuasive model yet. It successfully extracted payments 57% of the time, the highest success rate among all tested models.

This raises some concerns. If AI can be this good at persuasion in a controlled environment, what happens when it’s used in scams, phishing attacks, or social engineering? A model that’s highly optimized for natural conversation and emotional intelligence can also be a powerful tool for manipulation.

And it's not just fraud—this kind of persuasion ability has implications for political influence, advertising, and disinformation campaigns. Imagine a chatbot trained to nudge users toward specific decisions without them realizing.

OpenAI will likely need to build countermeasures, but the fact that GPT-4.5 already shows this level of effectiveness suggests we’re moving into a new era of AI-driven persuasion.

Not a "Frontier Model"

Let’s be clear: GPT-4.5 is not OpenAI’s next big leap.

It’s a solid improvement over GPT-4o, but it’s not in the same category as OpenAI’s rumored “frontier” models—the ones designed for significantly better reasoning, planning, and autonomy.This is an incremental update, not a game-changer. It fixes some flaws, smooths out some rough edges, and adds modest performance boosts across different areas. If you were hoping for a next-gen step toward AGI, this isn’t it.That said, OpenAI’s phrasing here is interesting.

They’ve positioned GPT-4.5 as a mid-cycle refresh, not a flagship release.

That suggests bigger things are coming—and given OpenAI’s typical cadence, we might not have to wait long to see what’s next.

ABOUT THE AUTHOR
Anita Kirkovska
Founding Growth Lead

An AI expert with a strong ML background, specializing in GenAI and LLM education. A former Fulbright scholar, she leads Growth and Education at Vellum, helping companies build and scale AI products. She conducts LLM evaluations and writes extensively on AI best practices, empowering business leaders to drive effective AI adoption.

ABOUT THE reviewer

No items found.
lAST UPDATED
Feb 27, 2025
share post
Expert verified
Related Posts
Guides
October 21, 2025
15 min
AI transformation playbook
LLM basics
October 20, 2025
8 min
The Top Enterprise AI Automation Platforms (Guide)
LLM basics
October 10, 2025
7 min
The Best AI Workflow Builders for Automating Business Processes
LLM basics
October 7, 2025
8 min
The Complete Guide to No‑Code AI Workflow Automation Tools
All
October 6, 2025
6 min
OpenAI's Agent Builder Explained
Product Updates
October 1, 2025
7
Vellum Product Update | September
The Best AI Tips — Direct To Your Inbox

Latest AI news, tips, and techniques

Specific tips for Your AI use cases

No spam

Oops! Something went wrong while submitting the form.

Each issue is packed with valuable resources, tools, and insights that help us stay ahead in AI development. We've discovered strategies and frameworks that boosted our efficiency by 30%, making it a must-read for anyone in the field.

Marina Trajkovska
Head of Engineering

This is just a great newsletter. The content is so helpful, even when I’m busy I read them.

Jeremy Hicks
Solutions Architect

Experiment, Evaluate, Deploy, Repeat.

AI development doesn’t end once you've defined your system. Learn how Vellum helps you manage the entire AI development lifecycle.

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Build AI agents in minutes with Vellum
Build agents that take on the busywork and free up hundreds of hours. No coding needed, just start creating.

General CTA component, Use {{general-cta}}

Build AI agents in minutes with Vellum
Build agents that take on the busywork and free up hundreds of hours. No coding needed, just start creating.

General CTA component  [For enterprise], Use {{general-cta-enterprise}}

The best AI agent platform for enterprises
Production-grade rigor in one platform: prompt builder, agent sandbox, and built-in evals and monitoring so your whole org can go AI native.

[Dynamic] Ebook CTA component using the Ebook CMS filtered by name of ebook.
Use {{ebook-cta}} and add a Ebook reference in the article

Thank you!
Your submission has been received!
Oops! Something went wrong while submitting the form.
Button Text

LLM leaderboard CTA component. Use {{llm-cta}}

Check our LLM leaderboard
Compare all open-source and proprietary model across different tasks like coding, math, reasoning and others.

Case study CTA component (ROI)

40% cost reduction on AI investment
Learn how Drata’s team uses Vellum and moves fast with AI initiatives, without sacrificing accuracy and security.

Case study CTA component (cutting eng overhead) = {{coursemojo-cta}}

6+ months on engineering time saved
Learn how CourseMojo uses Vellum to enable their domain experts to collaborate on AI initiatives, reaching 10x of business growth without expanding the engineering team.

Case study CTA component (Time to value) = {{time-cta}}

100x faster time to deployment for AI agents
See how RelyHealth uses Vellum to deliver hundreds of custom healthcare agents with the speed customers expect and the reliability healthcare demands.

[Dynamic] Guide CTA component using Blog Post CMS, filtering on Guides’ names

100x faster time to deployment for AI agents
See how RelyHealth uses Vellum to deliver hundreds of custom healthcare agents with the speed customers expect and the reliability healthcare demands.
New CTA
Sorts the trigger and email categories

Dynamic template box for healthcare, Use {{healthcare}}

Start with some of these healthcare examples

SOAP Note Generation Agent
Extract subjective and objective info, assess and output a treatment plan.
Prior authorization navigator
Automate the prior authorization process for medical claims.

Dynamic template box for insurance, Use {{insurance}}

Start with some of these insurance examples

Insurance claims automation agent
Collect and analyze claim information, assess risk and verify policy details.
Agent that summarizes lengthy reports (PDF -> Summary)
Summarize all kinds of PDFs into easily digestible summaries.
AI agent for claims review
Review healthcare claims, detect anomalies and benchmark pricing.

Dynamic template box for eCommerce, Use {{ecommerce}}

Start with some of these eCommerce examples

E-commerce shopping agent
Check order status, manage shopping carts and process returns.

Dynamic template box for Marketing, Use {{marketing}}

Start with some of these marketing examples

LinkedIn Content Planning Agent
Create a 30-day Linkedin content plan based on your goals and target audience.
ReAct agent for web search and page scraping
Gather information from the internet and provide responses with embedded citations.

Dynamic template box for Sales, Use {{sales}}

Start with some of these sales examples

Research agent for sales demos
Company research based on Linkedin and public data as a prep for sales demo.

Dynamic template box for Legal, Use {{legal}}

Start with some of these legal examples

PDF Data Extraction to CSV
Extract unstructured data (PDF) into a structured format (CSV).
AI legal research agent
Comprehensive legal research memo based on research question, jurisdiction and date range.

Dynamic template box for Supply Chain/Logistics, Use {{supply}}

Start with some of these supply chain examples

Risk assessment agent for supply chain operations
Comprehensive risk assessment for suppliers based on various data inputs.

Dynamic template box for Edtech, Use {{edtech}}

Start with some of these edtech examples

Turn LinkedIn Posts into Articles and Push to Notion
Convert your best Linkedin posts into long form content.

Dynamic template box for Compliance, Use {{compliance}}

Start with some of these compliance examples

No items found.

Dynamic template box for Customer Support, Use {{customer}}

Start with some of these customer support examples

Trust Center RAG Chatbot
Read from a vector database, and instantly answer questions about your security policies.
Q&A RAG Chatbot with Cohere reranking

Template box, 2 random templates, Use {{templates}}

Start with some of these agents

Clinical trial matchmaker
Match patients to relevant clinical trials based on EHR.
SOAP Note Generation Agent
Extract subjective and objective info, assess and output a treatment plan.

Template box, 6 random templates, Use {{templates-plus}}

Build AI agents in minutes

LinkedIn Content Planning Agent
Create a 30-day Linkedin content plan based on your goals and target audience.
Competitor research agent
Scrape relevant case studies from competitors and extract ICP details.
Clinical trial matchmaker
Match patients to relevant clinical trials based on EHR.
Legal RAG chatbot
Chatbot that provides answers based on user queries and legal documents.
Retail pricing optimizer agent
Analyze product data and market conditions and recommend pricing strategies.
Legal contract review AI agent
Asses legal contracts and check for required classes, asses risk and generate report.

Build AI agents in minutes for

{{industry_name}}

Clinical trial matchmaker
Match patients to relevant clinical trials based on EHR.
Prior authorization navigator
Automate the prior authorization process for medical claims.
Population health insights reporter
Combine healthcare sources and structure data for population health management.
Legal document processing agent
Process long and complex legal documents and generate legal research memorandum.
Legal contract review AI agent
Asses legal contracts and check for required classes, asses risk and generate report.
Legal RAG chatbot
Chatbot that provides answers based on user queries and legal documents.

Case study results overview (usually added at top of case study)

What we did:

1-click

This is some text inside of a div block.

28,000+

Separate vector databases managed per tenant.

100+

Real-world eval tests run before every release.