latentbrief
← Back to concepts

Concept

Fine-tuning

Taking a general-purpose AI model and giving it additional training on a specific subject, so it becomes noticeably better at that particular domain.

A general-purpose AI model is a bit like a highly educated generalist - someone who has read widely across many subjects and can hold a reasonable conversation about almost anything. Fine-tuning is like sending that generalist to do a deep specialisation. They spend months immersed in one field - reading nothing but medical research papers, or legal contracts, or customer support transcripts - and come out much more capable in that specific area.

The way this works technically is that after the initial large-scale training (which costs enormous amounts of money and takes months), you run a second, smaller training process using a focused dataset relevant to your use case. This second training adjusts the model to perform better on your specific type of content. The model does not forget everything it learned before - it layers new expertise on top of existing capabilities.

Fine-tuning is particularly valuable when you need the AI to adopt a specific style, follow particular rules consistently, or handle a type of content that is not well represented in general internet text. A model fine-tuned on your company's customer emails will respond in your brand voice more naturally than a general model ever could, just by prompting it differently.

The cost and complexity of fine-tuning has dropped significantly in recent years. Newer techniques allow you to adjust only a small part of the model's parameters rather than retraining the whole thing, which means companies with modest budgets can now build fine-tuned models that would have required enormous resources just a few years ago.

Fine-tuning and RAG are often used together. RAG gives the model access to specific documents at the moment of answering. Fine-tuning teaches the model to behave differently - to adopt a certain style, use specific terminology correctly, or follow a particular workflow - regardless of what documents it is reading.

Analogy

A generalist consultant who spends six months embedded with a law firm. They do not forget everything they knew before. Instead, they layer real legal expertise on top of their existing skills. After the experience, they are noticeably better at legal work than they were before - even when working on cases that were not part of their training.

Real-world example

GitHub Copilot - the AI coding assistant built into many developers' tools - is built on a base AI model that was then fine-tuned on vast amounts of code. The fine-tuning is what makes it code-specific: it understands programming syntax, common patterns, and developer conventions far better than a general text model would.

Why it matters

Fine-tuning is how companies build AI products that cannot easily be copied. A model fine-tuned on years of proprietary customer conversations, internal documents, or specialised data produces better results in that domain than any generic model - and that advantage is hard for competitors to replicate without access to the same data.

In the news

Related concepts