Early draft — longer version landing soon.
Clients frequently ask us to fine-tune a model. Our default answer is: probably not yet. Fine-tuning is expensive to set up, fragile to maintain, and almost always solvable with a well-designed retrieval layer plus better prompts.
The decision tree we actually run
- Is the problem “the model doesn’t know our stuff”? That’s a retrieval problem. Start with RAG.
- Is the problem “the model’s tone is wrong”? That’s a prompt-engineering problem. Iterate on the system prompt and few-shot examples.
- Is the problem “the model is bad at our specific task format”? Now we’re talking. Fine-tuning might help — but only after you’ve exhausted 1 and 2.
Fine-tuning buys you consistency on a narrow task. It doesn’t buy you knowledge of your documents, and it doesn’t buy you a better personality.
We’ve shipped exactly one fine-tune in the last year. We’ve shipped many RAGs.