Cookie Consent
Hi, this website uses essential cookies to ensure its proper operation and tracking cookies to understand how you interact with it. The latter will be set only after consent.
Read our Privacy Policy

Rectified Linear Unit (ReLU)

Rectified Linear Unit (ReLU) is a type of activation function used in neural networks, particularly in deep learning models. It is defined as the positive part of its argument:

ReLU(x)=max(0,x), effectively setting all negative values in the input to zero.

Why ReLU Matters

ReLU is favored in many neural networks due to its simplicity and efficiency. It introduces non-linearity into the model without affecting the scalability and speed of convergence during training. This function helps the network overcome issues like the vanishing gradient problem

Lakera LLM Security Playbook
Learn how to protect against the most common LLM vulnerabilities

Download this guide to delve into the most common LLM security risks and ways to mitigate them.

Related terms
untouchable mode.
Get started for free.

Lakera Guard protects your LLM applications from cybersecurity risks with a single line of code. Get started in minutes. Become stronger every day.

Join our Slack Community.

Several people are typing about AI/ML security. 
Come join us and 1000+ others in a chat that’s thoroughly SFW.