VOCABULARY

Regularization

Regularization in machine learning is a technique used to prevent overfitting, where a model performs well on training data but poorly on unseen data. Regularization adds a penalty to the loss function used to train the model, discouraging overly complex models and promoting simpler, more generalizable ones.

Why Regularization Matters

Regularization works by adding an extra term to the loss function - a term that increases as the model complexity increases. For example, L1 and L2 are common regularization techniques that add the sum of the absolute values (L1) or the sum of the squares (L2) of the model coefficients to the loss function. This penalizes large coefficients, leading to simpler models that are less likely to overfit on the training data.

‍

Lakera LLM Security Playbook
Learn how to protect against the most common LLM vulnerabilities

Download this guide to delve into the most common LLM security risks and ways to mitigate them.

Related terms
Activate
untouchable mode.
Get started for free.

Lakera Guard protects your LLM applications from cybersecurity risks with a single line of code. Get started in minutes. Become stronger every day.

Join our Slack Community.

Several people are typing about AI/ML security. 
Come join us and 1000+ others in a chat that’s thoroughly SFW.