What is a ReLU layer?

A Rectified Linear Unit(ReLU) is a non-linear activation function that performs on multi-layer neural networks. (e.g., f(x) = max(0,x) where x = input value).

What happens in ReLU layer?

In this layer we remove every negative value from the filtered image and replace it with zero. This function only activates when the node input is above a certain quantity. So, when the input is below zero the output is zero. However, when the input rises above a certain threshold it has linear relationship with the dependent variable. This means that it is able to accelerate the speed of a training data set in a deep neural network that is faster than other activation functions – this is done to avoid summing up with zero.

A ReLU layer:

  • performs element-wise operations.
  • has an output that is a rectified feature map.

Examples of ReLU

x f(x)=x F(X)
-110 f(-110)=0 0
-15 f(-15)=0 0
15 f(15)=15 15
25 f(25)=25 25

New on Educative
Learn to Code
Learn any Language as a beginner
Develop a human edge in an AI powered world and learn to code with AI from our beginner friendly catalog
🏆 Leaderboard
Daily Coding Challenge
Solve a new coding challenge every day and climb the leaderboard

Free Resources