Cookie Consent
Hi, this website uses essential cookies to ensure its proper operation and tracking cookies to understand how you interact with it. The latter will be set only after consent.
Read our Privacy Policy
VOCABULARY

Model Fine-Tuning

Model fine-tuning is a process in machine learning where a pre-trained model is adapted to a specific task. This approach is often used in deep learning, where a model is initially trained on a large dataset, such as ImageNet, and then fine-tuned on a smaller, more specific dataset.

How Model Fine-Tuning works

The fine-tuning technique starts with a pre-trained model, which has been trained on a large dataset with a vast amount of information. This pre-training step allows the model to capture general features and representations about the data, which can be a useful starting point for a specific task.

The process involves slightly adjusting the parameters of the pre-trained model according to the specific task by continuing the training on the new smaller, more specific dataset. This involves "fine-tuning" the values of the parameters to make them more suited to the new task.

This way, the model maintains most of the information from the pre-training phase and adapts it to the new task. This is especially useful when the new specific dataset is relatively small and does not have enough information to train a model from scratch.

The degree of fine-tuning can vary, from adjusting just the final layers of the model to tweaking the entire network. The choice depends on the similarity between the initial task and the specific task. If they are quite similar, adjustment may only be needed in the final layers. If they are different, a larger amount of fine-tuning might be required.

This technique is often used in transfer learning, where the knowledge gained while solving one problem is applied to a different but related problem.

Lakera LLM Security Playbook
Learn how to protect against the most common LLM vulnerabilities

Download this guide to delve into the most common LLM security risks and ways to mitigate them.

Related terms
Activate
untouchable mode.
Get started for free.

Lakera Guard protects your LLM applications from cybersecurity risks with a single line of code. Get started in minutes. Become stronger every day.

Join our Slack Community.

Several people are typing about AI/ML security. 
Come join us and 1000+ others in a chat that’s thoroughly SFW.