Concept
Fine-tuning
Taking a general-purpose AI model and giving it additional training on a specific subject, so it becomes noticeably better at that particular domain.
A general-purpose AI model is a bit like a highly educated generalist - someone who has read widely across many subjects and can hold a reasonable conversation about almost anything. Fine-tuning is like sending that generalist to do a deep specialisation. They spend months immersed in one field - reading nothing but medical research papers, or legal contracts, or customer support transcripts - and come out much more capable in that specific area.
The way this works technically is that after the initial large-scale training (which costs enormous amounts of money and takes months), you run a second, smaller training process using a focused dataset relevant to your use case. This second training adjusts the model to perform better on your specific type of content. The model does not forget everything it learned before - it layers new expertise on top of existing capabilities.
Fine-tuning is particularly valuable when you need the AI to adopt a specific style, follow particular rules consistently, or handle a type of content that is not well represented in general internet text. A model fine-tuned on your company's customer emails will respond in your brand voice more naturally than a general model ever could, just by prompting it differently.
The cost and complexity of fine-tuning has dropped significantly in recent years. Newer techniques allow you to adjust only a small part of the model's parameters rather than retraining the whole thing, which means companies with modest budgets can now build fine-tuned models that would have required enormous resources just a few years ago.
Fine-tuning and RAG are often used together. RAG gives the model access to specific documents at the moment of answering. Fine-tuning teaches the model to behave differently - to adopt a certain style, use specific terminology correctly, or follow a particular workflow - regardless of what documents it is reading.
Analogy
A generalist consultant who spends six months embedded with a law firm. They do not forget everything they knew before. Instead, they layer real legal expertise on top of their existing skills. After the experience, they are noticeably better at legal work than they were before - even when working on cases that were not part of their training.
Real-world example
GitHub Copilot - the AI coding assistant built into many developers' tools - is built on a base AI model that was then fine-tuned on vast amounts of code. The fine-tuning is what makes it code-specific: it understands programming syntax, common patterns, and developer conventions far better than a general text model would.
Why it matters
Fine-tuning is how companies build AI products that cannot easily be copied. A model fine-tuned on years of proprietary customer conversations, internal documents, or specialised data produces better results in that domain than any generic model - and that advantage is hard for competitors to replicate without access to the same data.
In the news
Unlocking Private Data for AI Without Sharing
arXiv CS.LG · 2d ago
Secure AI Fine-Tuning Workflow Unveiled
AWS ML Blog · 3d ago
AI Fine-Tuning Method Boosts Generalization Without Sacrificing Performance
arXiv CS.LG · 4d ago
AWS Rolls Out Tool to Track EU AI Act Compliance for LLMs
Amazon Web Services (AWS) · 4d ago
AI Advances Push Boundaries of Reinforcement Learning
LessWrong, arXiv CS.AI · 4d ago
Related concepts