Activate deep learning neurons faster with Dynamic RELU (ep. 101)
Source: In this episode I briefly explain the concept behind activation functions in deep learning. One of the most widely used activation function is the rectified linear unit (ReLU). While there are several flavors of ReLU in the literature, in this episode I speak about a very interesting approach that keeps computational complexity low while improving performance quite consistently. This episode is supported by At pryml we let companies share confid
|
|