Vigyata.AI
Is this your channel?

GPT-5.4 Is Here (Worse Than Sonnet 4.6?)

3.1K views· 30 likes· 15:49· Mar 5, 2026

🛍️ Products Mentioned (4)

Want to learn how to use AI? Join our skool group, we also have a free weekly call: https://www.skool.com/data-and-ai GPT-5.4 just dropped and OpenAI is calling it their most capable model yet — built for "professional work." But does it actually deliver? In this video, I break down the GPT-5.4 press release, benchmarks, pricing, and then put it head-to-head against Claude Sonnet 4.6 with real-world tests: niche knowledge questions, spreadsheet data cleanup, and more. The results might surprise you. We compare API pricing across GPT-5.4, GPT-5.4 Pro, Claude Sonnet 4.6, Claude Opus, and Gemini 3.1 Pro — so you know exactly what you're paying for. If you're choosing a default model for your workflows and automations, this one's for you. TIMESTAMPS 0:00 - GPT-5.4 Is Here 0:23 - Press Release Breakdown 1:00 - Benchmarks Overview 1:21 - Knowledge Work & Spreadsheet Claims 1:55 - Computer Use & Vision 2:34 - Availability & Pricing Breakdown 3:33 - GPT-5.4 vs Claude vs Gemini Pricing Comparison 4:32 - Test 1: Niche Knowledge — Most Valuable Baseball Cards 5:45 - Test 1 Results: Sonnet vs GPT-5.4 6:33 - Test 2: Niche Knowledge — Oldest Presidential Card Sets 7:40 - GPT-5.4 Hallucinations vs Sonnet Accuracy 9:00 - Test 3: Spreadsheet Data Cleanup 10:10 - Sonnet Spreadsheet Results 10:50 - GPT-5.4 Spreadsheet Results 11:20 - Final Verdict & Takeaways 🚀 Hire me for Data Work: https://ryanandmattdatascience.com/data-freelancing/ 👨‍💻 Mentorships: https://ryanandmattdatascience.com/mentorship/ 📧 Email: ryannolandata@gmail.com 🌐 Website & Blog: https://ryanandmattdatascience.com/ OTHER SOCIALS: Ryan’s LinkedIn: https://www.linkedin.com/in/ryan-p-nolan/ Matt’s LinkedIn: https://www.linkedin.com/in/matt-payne-ceo/ Twitter/X: https://x.com/RyanMattDS *This is an affiliate program. We receive a small portion of the final sale at no extra cost to you.

About This Video

GPT-5.4 is here, and on paper it’s OpenAI’s most direct challenge to Claude yet—“designed for professional work,” better benchmarks, and a big push around agents, computer use, and cleaner knowledge-work outputs. In this video, I walk through the press release, what actually changed versus GPT-5.2, and the pricing that matters if you’re picking a default model for automations. The headline: the context window stays the same (1M), and the big surprise is how expensive output tokens get—especially if you accidentally pick GPT-5.4 Pro. Then I stop reading marketing and run real tests. I put GPT-5.4 head-to-head with Claude Sonnet 4.6 on niche knowledge questions (high-end baseball cards and old presidential card sets) and a spreadsheet cleanup task with messy emails, inconsistent dates, and broken fields. In my tests, Sonnet 4.6 was consistently more accurate on the niche research and did a cleaner job on the spreadsheet—plus it produced an issue log and an easy-to-download cleaned file. GPT-5.4, honestly, hallucinated hard on the presidential set question and didn’t fully deliver on the spreadsheet cleanup. If you’re choosing a model for workflows, this is why you test with your edge cases, not benchmarks.

Frequently Asked Questions

🎬 More from Ryan & Matt Data Science