Rectified Linear Unit

ReLU formula

⚠ Switch to EXCALIDRAW VIEW in the MORE OPTIONS menu of this document. ⚠

Text Elements

Link to original

Dying ReLU Problem

  • dead state of ReLU neuron most input ranges are negative and the output is 0
  • does not recover because gradient of 0 is 0
  • Causes
    • high learning rate accidently sets negative weights
    • large negative bias bias terms become negative
  • deep ReLU network will eventually die in probability as the depth goes to infinite 

Leaky ReLU

  • avoid dying ReLU non zero gradient in the entire range

Parametric ReLU