Tech10
Back to blog

When Fine-Tuning Actually Beats RAG (And the Cases It's Wasted On)

When Fine Tuning Actually Beats Rag
AI StrategyMar 19, 20269 min readDoreid Haddad
Listen to this article0:00

Most of what you'll read about fine-tuning is either a vendor pitch or a warning. The truth is boring and specific: fine-tuning beats RAG in a short list of cases, loses in most others, and gets picked for the wrong reasons more than any other technique in the AI stack.

The cases where it actually wins are worth knowing cold, because the teams that need them really do need them. The cases where it gets wasted are worth knowing even better, because that's where most of the $40,000 mistakes come from.

The honest comparison, one more time

RAG gives the model new knowledge by looking things up at query time. Fine-tuning changes how the model behaves by retraining it on your examples. This is explored in more depth in RAG vs fine-tuning: which one solves your problem.

Fine-tuning doesn't teach the model facts reliably. It teaches the model how to respond. The 2024 Microsoft Research paper "Fine-Tuning or Retrieval?" put this on paper: fine-tuning alone added about six percentage points on agriculture QA accuracy, but the gains came from pattern matching on the training distribution, not from stored knowledge. If you want the model to know something, use RAG. If you want the model to behave a specific way, fine-tune.

Almost every fine-tuning failure traces back to that single confusion.

Win case one: structured output that has to be exact

The clearest fine-tuning win is in structured output tasks where a downstream system will fail if the format drifts.

A logistics company processing 40,000 shipping manifests a day needs every output to be parseable JSON with exact field names and types. With Claude Opus 4.6 and a careful prompt, you can get to 97% format compliance. That sounds great until you realize 3% of 40,000 is 1,200 broken records a day, each of which needs human intervention. At roughly 2 minutes per intervention, that's 40 hours of human time daily.

A fine-tuned model on 2,000 input-output pairs from the last six months can get that number to 99.7% or higher. That's about 120 broken records a day, down to 4 hours of human work. The math on that is obvious the first time you see it.

This is the cleanest fine-tuning win and the one most people ignore. Structured output compliance is boring. It's also what keeps pipelines alive.

Win case two: domain jargon the base model mangles

Base models learn from a broad corpus. They're fluent in common language and weaker at domain-specific usage, especially when the domain uses common words in unusual ways.

Examples I've watched this matter:

  • In pharmaceutical research, "safety signal" means a statistically unusual pattern in adverse event reports. The base model will interpret it as "a safety-related alert," which is wrong enough to break downstream analysis.
  • In insurance, "loss reserve" is an accounting liability, not a negative number you're sad about. Base models without fine-tuning get this wrong about a quarter of the time in my experience.
  • In legal drafting, "shall" has a specific prescriptive meaning that courts have written entire opinions about. A model trained on general internet text uses it as an archaic synonym for "will."

In these cases, adding more examples to a prompt helps, but only up to the prompt size limit. Fine-tuning on 500 to 2,000 correctly annotated domain examples solves the vocabulary problem permanently and cheaply at query time.

The trap version: teams fine-tune on domain documents thinking it will teach the model facts in that domain. It won't, reliably. Fine-tuning teaches vocabulary and style. For the facts, you still need RAG.

Win case three: style and tone at scale

A global consumer brand ships 12,000 emails a month across 16 markets. Each market has a tone profile, a compliance-approved vocabulary list, and a specific formatting convention. A frontier model with a tight system prompt gets you to about 85% on-brand, measured by blind human review.

Fine-tuning on 800 approved historical emails per market pushes that number to 94% or higher. The prompt also gets shorter, because the model has the style baked in. Cost per email drops. Quality improves.

This only works if "good" can be defined from examples. If the style guide is "use your judgment," fine-tuning can't help you, because you can't produce the training set. This is where most style projects die: the team thinks they have a style, but they don't, and fine-tuning surfaces that fast.

The trap version: teams fine-tune for "voice" without an eval set. They ship a model that the CMO loves on their three example prompts and the rest of the company hates. An eval set of 50 diverse scenarios, scored blind, is the minimum acceptable bar.

Win case four: latency and cost at very high volume

Fine-tuned smaller models are faster and cheaper than large frontier models running big prompts. This only pays off at volume.

Consider a product classification pipeline. A merchant has 3 million SKUs and wants each one classified into the correct GPC (Global Product Classification) taxonomy. With GPT-5 and a prompt with 10 few-shot examples, each classification costs about $0.003 and takes about 1.2 seconds. That's $9,000 and 1,000 hours of wall time for the batch.

With a fine-tuned smaller model, each classification costs about $0.0004 and takes 300 milliseconds. That's $1,200 and 250 hours. Over a 3-million-SKU catalog, the difference pays for the fine-tuning project in the first batch.

The break-even volume for this kind of shrink depends on the specific model and task, but the shape is consistent: above about 100,000 calls a day, smaller fine-tuned models start to crush the frontier-model-with-big-prompt option. We walk through the full numbers in the real cost of RAG vs fine-tuning.

Win case five: behavior consistency under adversarial inputs

The last win case is the one you don't think about until you get burned by it.

Frontier models are trained to be helpful, which means they want to answer every query. In regulated environments, this is a liability. A customer service model for a pharmacy chain cannot be helpful when asked about off-label drug use. A banking chatbot cannot be helpful when asked for investment advice. The prompt-engineering fix is a long list of "do not answer questions about X," which works until a user asks a slightly creative version and the model complies.

Fine-tuning on rejection examples (300 to 800 carefully constructed refusals) builds the "no" behavior into the weights. The model is less helpful in general (by design) and more consistently refuses specific categories. This is what legal and compliance teams actually want.

The RAG version of this doesn't exist. You can't RAG your way into reliable refusal behavior.

Waste case one: teaching the model new facts

This is the most common waste I see. A team has a 500-page internal policy manual. They fine-tune the base model on the full text, thinking it will "teach the model their policies." It doesn't work.

Fine-tuning on raw documents doesn't reliably install facts in the model's weights. The model sees the text, adjusts parameters toward similar-sounding text, and doesn't actually memorize the policies. When you ask "what's our refund policy?", you get something that sounds like a refund policy but may or may not be yours.

The correct tool for this is RAG. Every time. Fine-tuning on your documents is $8,000 wasted, every time.

Waste case two: "we need fine-tuning for our industry"

Teams in legal, healthcare, and finance sometimes arrive at fine-tuning by assumption. "We're in a regulated industry, so we need a fine-tuned model." This is wrong in both directions.

You don't need fine-tuning to be in a regulated industry. You need retrieval with audit trails, access controls on your data, and clear evaluation. A well-built RAG system meets most regulatory requirements better than a fine-tuned model, because you can show exactly what source the answer came from.

And you can't solve regulatory problems with fine-tuning. If your AI outputs need to be auditable to the document level, fine-tuning makes that harder, not easier. The provenance of a fine-tuned answer is "the model's weights," which is not an answer your regulator will accept.

Waste case three: the "tone" project that's really a prompt problem

Every consulting practice sees this one. A team wants their chatbot to "sound more like us." They get a quote for a fine-tuning project. They don't first try writing a tight system prompt with three worked examples.

Run the prompt experiment first. Write a 400-word system prompt that describes the voice, lists three prohibited patterns, and gives two full worked examples. Put that on a frontier model. Evaluate 20 scenarios blind. If the quality clears the bar, you're done. Ship it.

About 70% of "tone" projects I've encountered can be solved with a careful prompt and an eval set. The other 30% are genuinely fine-tuning cases. That's a 10-to-1 ratio of "skip fine-tuning" to "fine-tune" for tone work.

What a fine-tuning decision should actually look like

If you've honestly cleared the traps, the decision process is:

1. Confirm the problem is behavior, not knowledge. If the model doesn't know your data, RAG is the answer, not fine-tuning.

2. Confirm prompt engineering didn't solve it. Write the best prompt you can. Include 2 to 3 worked examples. Evaluate on 20 scenarios. If quality clears 90%, stop.

3. Confirm you have the data. You need 500 to 5,000 clean, consistent input-output pairs depending on task complexity. Quality beats quantity here. 200 great examples outperform 2,000 sloppy ones. Anthropic's recent fine-tuning case studies show this over and over.

4. Confirm the volume or accuracy requirement justifies the cost. Budget $20,000 to $40,000 all-in for a first fine-tuning project. Divide by the cost savings or quality improvement you expect. If it doesn't pay back in 12 months, it's not worth starting.

5. Plan for the retraining cycle. Base models ship new versions every 3 to 6 months. You will retrain. Factor that into the ongoing cost.

If all five check out, fine-tune. If any don't, don't.

What the benchmarks hide

Vendor benchmarks are optimistic. Real-world performance on your specific data is usually 10 to 20 points below the benchmark number. I've written elsewhere that vendors sell benchmarks, not business outcomes, and this is a case where the pattern really does matter.

A fine-tuned model that benchmarks at 95% on its own eval set will probably run 80% to 85% on your data, at least at first. Plan for that. Budget eval iterations. Don't ship the first model to production.

The decision, compressed

Fine-tune when the problem is behavior, the requirement is exact, the data exists, and the volume or accuracy ask justifies the cost. Don't fine-tune to inject facts, to sound regulated, to feel serious, or to skip prompt engineering.

Most teams I meet don't need fine-tuning. The ones that do, really do. Most teams don't need fine-tuning unpacks the pattern more directly.

Frequently Asked Questions

When should I fine-tune instead of RAG?

Fine-tune when the problem is behavior (format, tone, vocabulary, structured output), not knowledge. Use RAG when the problem is missing or changing information. If both apply, build RAG first and add fine-tuning only if a specific gap justifies it.

Can fine-tuning teach the model facts?

Not reliably. Fine-tuning teaches the model patterns and behavior. Facts are better installed via RAG, where the model retrieves the document at query time and can cite it. Teams who fine-tune to add facts usually end up with a model that sounds right but isn't.

How many examples do I need to fine-tune?

200 to 500 examples work for narrow classification tasks. Structured extraction needs 500 to 2,000. Style and generation tasks benefit from 1,000 to 5,000. Enterprise production baselines run 10,000 to 50,000. Quality of labeling matters more than count.

Do I still need RAG if I fine-tune?

Often yes. Fine-tuning bakes in behavior; RAG provides current knowledge. High-performance systems combine both: fine-tune for style and format, RAG for the facts. Most teams should build RAG first and validate it before adding fine-tuning on top.

Sources
Doreid Haddad
Written byDoreid Haddad

Founder, Tech10

Doreid Haddad is the founder of Tech10. He has spent over a decade designing AI systems, marketing automation, and digital transformation strategies for global enterprise companies. His work focuses on building systems that actually work in production, not just in demos. Based in Rome.

Read more about Doreid

Keep reading