Fine Tuning

Posted in
Definitions

Definition

Fine-tuning in the context of Large Language Models (LLMs) refers to the process of adapting a pre-trained model to perform specialised tasks or excel in specific domains by training it further on a smaller, task-oriented dataset. 

This technique updates the model’s parameters to align its outputs with targeted requirements, bridging the gap between general-purpose capabilities and domain-specific expertise.

The process typically involves structured steps like dataset preparation, model initialisation, and iterative training to optimise performance for the target use case.

Key aspects of fine-tuning include:

Adaptation

Starting with a model trained on vast general data, then refining it using curated datasets (e.g., medical reports, legal documents) to handle niche terminology and context.

Efficiency

Leveraging existing model knowledge rather than training from scratch, reducing computational costs compared to full pre-training.

Methods

Ranges from full fine-tuning (updating all parameters) to parameter-efficient approaches like LoRA (Low-Rank Adaptation), which modifies only subsets of weights.

Applications

Enables customisation for tasks like medical report generation, legal analysis, and controlled conversational tones while maintaining compliance with data regulations.