Connect with us

News

OpenAI Slashes ChatGPT o3 API Prices by 80% Without Hurting Performance

Published

on

<p data-start&equals;"250" data-end&equals;"421">OpenAI is making waves again&comma; but this time it’s not a new model or feature&period; It’s a major price drop—one that could shift how developers build and scale AI apps overnight&period;<&sol;p>&NewLine;<p data-start&equals;"423" data-end&equals;"676">On Wednesday&comma; OpenAI announced it’s reducing prices for its most capable reasoning model&comma; o3&comma; by a staggering 80&percnt;&period; And here&&num;8217&semi;s the kicker&colon; the model’s performance hasn&&num;8217&semi;t changed one bit&period; That’s not just a marketing line&period; Independent testers backed it up&period;<&sol;p>&NewLine;<h2 data-start&equals;"678" data-end&equals;"713">What’s Actually Changed With o3&quest;<&sol;h2>&NewLine;<p data-start&equals;"715" data-end&equals;"854">The price cut is massive&period; We&&num;8217&semi;re talking about a reduction from &dollar;10 to &dollar;2 per million input tokens&comma; and &dollar;40 to &dollar;8 per million output tokens&period;<&sol;p>&NewLine;<p data-start&equals;"856" data-end&equals;"1002">That alone is enough to turn heads in the dev community&period; But it’s not just the numbers that matter—it’s what didn’t change that’s more impressive&period;<&sol;p>&NewLine;<p data-start&equals;"1004" data-end&equals;"1290">The o3 model you get through the API today is the exact same one you got before the price drop&period; OpenAI clarified this point clearly on X&comma; saying they simply &OpenCurlyDoubleQuote;optimized &lbrack;their&rsqb; inference stack&period;” Basically&comma; they found a more efficient way to run the same model—nothing more&comma; nothing less&period;<&sol;p>&NewLine;<p data-start&equals;"1004" data-end&equals;"1290"><a href&equals;"https&colon;&sol;&sol;www&period;theibulletin&period;com&sol;wp-content&sol;uploads&sol;2025&sol;06&sol;OpenAI-o3-model-pricing-2025-chart&period;jpg"><img class&equals;"aligncenter size-full wp-image-57661" src&equals;"https&colon;&sol;&sol;www&period;theibulletin&period;com&sol;wp-content&sol;uploads&sol;2025&sol;06&sol;OpenAI-o3-model-pricing-2025-chart&period;jpg" alt&equals;"OpenAI o3 model pricing 2025 chart" width&equals;"1486" height&equals;"817" &sol;><&sol;a><&sol;p>&NewLine;<h2 data-start&equals;"1292" data-end&equals;"1334">Performance&quest; Still Rock Solid&comma; Says ARC<&sol;h2>&NewLine;<p data-start&equals;"1336" data-end&equals;"1466">There were plenty of raised eyebrows after the announcement&period; A price drop that steep often smells like a downgrade&period; Not this time&period;<&sol;p>&NewLine;<p data-start&equals;"1468" data-end&equals;"1600">ARC Prize&comma; a benchmark group that independently tests AI models&comma; confirmed the performance of the o3-2025-04-16 model hasn’t budged&period;<&sol;p>&NewLine;<p data-start&equals;"1602" data-end&equals;"1762">Just one sentence from their statement summed it all up&colon;<br data-start&equals;"1658" data-end&equals;"1661" &sol;>&OpenCurlyDoubleQuote;We compared the retest results with the original results and observed no difference in performance&period;”<&sol;p>&NewLine;<p data-start&equals;"1764" data-end&equals;"1782">Let’s pause there&period;<&sol;p>&NewLine;<p data-start&equals;"1784" data-end&equals;"1876">That&&num;8217&semi;s a critical point&period; No performance hit&period; No sneaky model swap&period; Just better backend work&period;<&sol;p>&NewLine;<h2 data-start&equals;"1878" data-end&equals;"1921">Why This Actually Matters for Developers<&sol;h2>&NewLine;<p data-start&equals;"1923" data-end&equals;"2022">For devs building with OpenAI’s tools&comma; pricing isn’t just a budgeting line item—it’s make or break&period;<&sol;p>&NewLine;<p data-start&equals;"2024" data-end&equals;"2062">Here’s what this means for developers&colon;<&sol;p>&NewLine;<ul data-start&equals;"2064" data-end&equals;"2193">&NewLine;<li data-start&equals;"2064" data-end&equals;"2105">&NewLine;<p data-start&equals;"2066" data-end&equals;"2105">Input now costs &dollar;2 per million tokens<&sol;p>&NewLine;<&sol;li>&NewLine;<li data-start&equals;"2106" data-end&equals;"2148">&NewLine;<p data-start&equals;"2108" data-end&equals;"2148">Output now costs &dollar;8 per million tokens<&sol;p>&NewLine;<&sol;li>&NewLine;<li data-start&equals;"2149" data-end&equals;"2193">&NewLine;<p data-start&equals;"2151" data-end&equals;"2193">Same model&comma; same accuracy&comma; lower burn rate<&sol;p>&NewLine;<&sol;li>&NewLine;<&sol;ul>&NewLine;<p data-start&equals;"2195" data-end&equals;"2387">Apps like Cursor and Windsurf&comma; which are built directly on the API&comma; instantly become more cost-effective to run&period; That trickles down into cheaper tools for users—or better margins for startups&period;<&sol;p>&NewLine;<p data-start&equals;"2389" data-end&equals;"2496">And if you’re an indie developer&quest; This move could bring enterprise-grade AI into your weekend side project&period;<&sol;p>&NewLine;<h2 data-start&equals;"2498" data-end&equals;"2522">A Look at the Numbers<&sol;h2>&NewLine;<p data-start&equals;"2524" data-end&equals;"2610">Let’s put it in perspective with a table&period; Here&&num;8217&semi;s how the old and new pricing stack up&colon;<&sol;p>&NewLine;<div class&equals;"&lowbar;tableContainer&lowbar;16hzy&lowbar;1">&NewLine;<div class&equals;"&lowbar;tableWrapper&lowbar;16hzy&lowbar;14 group flex w-fit flex-col-reverse" tabindex&equals;"-1">&NewLine;<table class&equals;"w-fit min-w-&lpar;--thread-content-width&rpar;" data-start&equals;"2612" data-end&equals;"2940">&NewLine;<thead data-start&equals;"2612" data-end&equals;"2691">&NewLine;<tr data-start&equals;"2612" data-end&equals;"2691">&NewLine;<th data-start&equals;"2612" data-end&equals;"2625" data-col-size&equals;"sm">Token Type<&sol;th>&NewLine;<th data-start&equals;"2625" data-end&equals;"2651" data-col-size&equals;"sm">Old Price &lpar;per million&rpar;<&sol;th>&NewLine;<th data-start&equals;"2651" data-end&equals;"2677" data-col-size&equals;"sm">New Price &lpar;per million&rpar;<&sol;th>&NewLine;<th data-start&equals;"2677" data-end&equals;"2691" data-col-size&equals;"sm">&percnt; Decrease<&sol;th>&NewLine;<&sol;tr>&NewLine;<&sol;thead>&NewLine;<tbody data-start&equals;"2775" data-end&equals;"2940">&NewLine;<tr data-start&equals;"2775" data-end&equals;"2857">&NewLine;<td data-start&equals;"2775" data-end&equals;"2788" data-col-size&equals;"sm">Input<&sol;td>&NewLine;<td data-start&equals;"2788" data-end&equals;"2815" data-col-size&equals;"sm">&dollar;10<&sol;td>&NewLine;<td data-start&equals;"2815" data-end&equals;"2842" data-col-size&equals;"sm">&dollar;2<&sol;td>&NewLine;<td data-start&equals;"2842" data-end&equals;"2857" data-col-size&equals;"sm">80&percnt;<&sol;td>&NewLine;<&sol;tr>&NewLine;<tr data-start&equals;"2858" data-end&equals;"2940">&NewLine;<td data-start&equals;"2858" data-end&equals;"2871" data-col-size&equals;"sm">Output<&sol;td>&NewLine;<td data-start&equals;"2871" data-end&equals;"2898" data-col-size&equals;"sm">&dollar;40<&sol;td>&NewLine;<td data-start&equals;"2898" data-end&equals;"2925" data-col-size&equals;"sm">&dollar;8<&sol;td>&NewLine;<td data-start&equals;"2925" data-end&equals;"2940" data-col-size&equals;"sm">80&percnt;<&sol;td>&NewLine;<&sol;tr>&NewLine;<&sol;tbody>&NewLine;<&sol;table>&NewLine;<div class&equals;"sticky end-&lpar;--thread-content-margin&rpar; h-0 self-end select-none">&NewLine;<div class&equals;"absolute end-0 flex items-end"><&sol;div>&NewLine;<&sol;div>&NewLine;<&sol;div>&NewLine;<&sol;div>&NewLine;<p data-start&equals;"2942" data-end&equals;"3015">That’s not a small dip&period; That’s OpenAI essentially opening the floodgates&period;<&sol;p>&NewLine;<h2 data-start&equals;"3017" data-end&equals;"3055">The &OpenCurlyDoubleQuote;o3-pro” Surprise That Followed<&sol;h2>&NewLine;<p data-start&equals;"3057" data-end&equals;"3150">While all eyes were on the price cut&comma; OpenAI quietly added something else to the API&colon; o3-pro&period;<&sol;p>&NewLine;<p data-start&equals;"3152" data-end&equals;"3362">This new variant of the model is built for users who want even better output quality&period; It uses more compute&comma; which likely means it’ll cost more—but also means it can give stronger responses in complex scenarios&period;<&sol;p>&NewLine;<p data-start&equals;"3364" data-end&equals;"3523">The timing wasn’t random either&period; OpenAI knows that not every user wants to save money&period; Some want better answers&comma; even at a higher price&period; o3-pro fills that gap&period;<&sol;p>&NewLine;<h2 data-start&equals;"3606" data-end&equals;"3654">API Users Win Big&comma; Regular Users… Not So Much<&sol;h2>&NewLine;<p data-start&equals;"3656" data-end&equals;"3773">If you’re using ChatGPT through the regular app&comma; this change doesn’t directly affect you&period; Prices there haven’t moved&period;<&sol;p>&NewLine;<p data-start&equals;"3775" data-end&equals;"3921">But under the hood&comma; the ripple effect is real&period; Lower API costs mean third-party tools powered by ChatGPT might become faster&comma; smarter&comma; or cheaper&period;<&sol;p>&NewLine;<p data-start&equals;"3923" data-end&equals;"4010">You might not see it on your bill&comma; but you might feel it in the apps you use every day&period;<&sol;p>&NewLine;

Leela Sehgal is an Indian author who works at ketion.com. She writes short and meaningful articles on various topics, such as culture, politics, health, and more. She is also a feminist who explores the issues of identity and empowerment in her works. She is a talented and versatile writer who delivers quality and diverse content to her readers.

Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

TRENDING