← Back to all posts

Prompt Optimization Just Crushed Frontier Models (And Made Them 90x Cheaper)

AI/MLSeptember 24, 202511 min read
Prompt Optimization Just Crushed Frontier Models (And Made Them 90x Cheaper)

TL;DR: Databricks took an open-source model (gpt-oss-120b), threw automated prompt optimization at it (GEPA), and suddenly it's beating Claude Opus 4.1 by ~3% while being 90x cheaper to run. Same trick lifted Claude models by 6-7%. This isn't incremental improvement - this is the entire quality-cost curve shifting left.

so here's something that shouldn't be possible but apparently is: you can take a mid-tier open-source LLM, optimize its prompts automatically, and suddenly it's outperforming the most expensive frontier models on the market

not by a little. by enough that it actually matters

and here's the kicker - it costs 90x less to serve

the setup: everyone's chasing better models

the default playbook in AI right now is basically: throw more compute at bigger models, pray they get smarter, charge customers more because "it's frontier AI." rinse, repeat

but Databricks just said "what if we stop training new models and just make the prompts better?"

they built this thing called GEPA (it's a prompt optimization technique that came out of Databricks and UC Berkeley research). and it doesn't change the model at all - it just figures out better ways to ask the model questions

sounds simple. sounds almost dumb. but the results are wild

the numbers that made me do a double-take

they took gpt-oss-120b - which is basically an open-source model that most people weren't even talking about - and ran GEPA on it

the result?

  • beat Claude Sonnet 4 by ~3%
  • beat Claude Opus 4.1 by ~3%
  • costs roughly 20x less than Sonnet 4 to serve
  • costs roughly 90x less than Opus 4.1 to serve

that last point is the one that breaks everything. you're not just getting comparable performance. you're getting BETTER performance at a fraction of the cost

Real talk: This is the kind of shift that makes CFOs stop caring which model is "best" and start asking "why the hell are we paying 90x more for worse results?"

what is GEPA actually doing?

ok so GEPA isn't magic. it's just really good at search

instead of you sitting there crafting the perfect prompt through trial and error (which is what most people do), GEPA runs an iterative search guided by evaluation feedback

basically:

  • starts with a baseline prompt
  • tries variations systematically
  • measures which ones actually work better on your task
  • keeps iterating until it finds something that performs

it's like if you took the "prompt engineering" meme seriously and actually automated the entire grind of testing thousands of prompt variations

the crazy part? it works on BOTH open-source and frontier models. they used the same technique on Claude Sonnet 4 and Claude Opus 4.1 and lifted them by 6-7% too

the benchmark: information extraction at scale

they tested this on something actually hard - information extraction from real enterprise documents

not toy problems. not "summarize this paragraph." we're talking:

  • documents over 100 pages long
  • extraction schemas with 70+ fields
  • hierarchical nested structures
  • domain-specific jargon across finance, legal, commerce, healthcare

this is the kind of stuff where enterprises actually pay money because it's painful to do manually and hard to automate

they built a benchmark called IE Bench specifically for this. and when they ran the tests, the results were pretty clear:

Key finding: GEPA-optimized gpt-oss-120b doesn't just compete with frontier models - it beats them. On real tasks. With real evaluation metrics. While costing a fraction to run.

the lifetime cost analysis (where economics breaks)

here's where it gets interesting for anyone who's actually deploying AI at scale

they calculated the total cost of:

  • running the optimization process (one-time cost)
  • serving 100k requests in production (ongoing cost)

at 100k requests, the lifetime cost breakdown looks like:

  • gpt-oss-120b with GEPA: cheapest by far, both optimization and serving
  • Claude Sonnet 4 with GEPA: similar total cost to GPT 4.1 with SFT
  • Claude Opus 4.1 with GEPA: most expensive due to high serving costs

the brutal truth: at 100k requests, serving costs dominate everything. the one-time optimization overhead gets amortized so fast it basically disappears

by 10M requests? optimization costs aren't even visible on the chart anymore

prompt optimization vs supervised fine-tuning

so naturally people are gonna ask: how does this compare to just fine-tuning the model?

they tested that too. ran supervised fine-tuning (SFT) on GPT 4.1 and compared it to GEPA optimization

results:

  • GEPA delivers performance on par with or better than SFT
  • GEPA reduces serving costs by 20% compared to SFT
  • both techniques can work TOGETHER for even better results

that last point is important. this isn't GEPA vs fine-tuning. it's GEPA + fine-tuning if you want maximum performance

but if you're choosing one? GEPA gives you better quality-cost tradeoff. especially when you factor in:

  • no need to maintain fine-tuned model versions
  • can apply to any base model instantly
  • doesn't require labeled training data
  • optimization happens in hours, not days

why this matters more than another benchmark

look, we get new AI benchmarks every week. most of them are noise

but this one matters because it shifts the entire conversation

before: "we need to use Claude Opus because it's the best, even though it's expensive"

after: "why are we paying 90x more for worse performance when we could optimize prompts on an open model?"

this isn't about squeezing 2% more accuracy out of a model. it's about fundamentally changing the economics of deploying AI in production

The shift: Prompt optimization isn't just an efficiency hack. It's moving the entire quality-cost Pareto frontier. You're getting both better performance AND lower costs. That's not supposed to be possible.

what this means for teams shipping AI

if you're building AI products right now, this changes your playbook:

  • stop defaulting to frontier models: you might be overpaying for worse results
  • invest in prompt optimization infrastructure: the ROI is massive and compounds over time
  • measure quality AND cost together: neither matters alone anymore
  • consider open-source models seriously: with proper optimization, they can beat proprietary models
  • think about lifetime costs: one-time optimization overhead is nothing compared to ongoing serving costs

and if you're at a company currently burning money on expensive model APIs, you probably just got handed a way to cut costs by 20-90x without sacrificing quality

that's the kind of win that gets noticed. fast

the bigger picture: AI economics just shifted

here's what really matters about this work:

for years, the AI industry has operated on the assumption that better performance = bigger models = higher costs. if you wanted SOTA results, you paid frontier model prices. end of story

this research breaks that assumption

you can now get better results at lower costs by optimizing how you use existing models. the leverage isn't in training bigger models - it's in using the models we already have more intelligently

and that's a fundamentally different game

Bottom line: The AI cost crisis everyone's been worried about? It might not be inevitable. We might have just been using models inefficiently this whole time. Better prompts > bigger models. Economics just got interesting again.

References

Note: This research comes from Databricks Mosaic Research in collaboration with UC Berkeley. The techniques (GEPA, TAO, RLVR, ALHF) are now available in Databricks Agent Bricks. The IE Bench evaluation suite spans finance, legal, commerce, and healthcare domains with real enterprise complexity.