VOCABULARY

Stochastic Gradient Descent (SGD)

Stochastic Gradient Descent (SGD) is an optimization algorithm used in machine learning to minimize a function by iteratively moving towards the minimum value of the function. It's an approximation of gradient descent, where updates to the model parameters are made using a subset of the data rather than the full dataset.

How SGD Works

In SGD, the model parameters are updated for each training example or a small batch of training examples. This frequent updating with a limited amount of data introduces randomness in the optimization path, which can help escape local minima and often leads to faster convergence on large datasets.

Lakera LLM Security Playbook
Learn how to protect against the most common LLM vulnerabilities

Download this guide to delve into the most common LLM security risks and ways to mitigate them.

Related terms
Activate
untouchable mode.
Get started for free.

Lakera Guard protects your LLM applications from cybersecurity risks with a single line of code. Get started in minutes. Become stronger every day.

Join our Slack Community.

Several people are typing about AI/ML security. 
Come join us and 1000+ others in a chat that’s thoroughly SFW.