2 5 which of the following holds good for rectified linear unit rectif
Question
2.5 Which of the following holds good for rectified linear unit?
Rectified Linear Unit is computationally cheaper than both sigmoid and tanh activation
function.
Using ReLU activation function brings out model sparsity.
Since ReLU activation function is a linear region for positive input values, it could lead to
an "dying ReLU" issue.
By avoiding the activated value to be 0 along negative axis, "dying ReLU" issue can be
addressed.