-
Notifications
You must be signed in to change notification settings - Fork 7
Activation_Fns
Abil N George edited this page Sep 1, 2014
·
1 revision
Activation Function is a function used to transform the activation level of a unit (neuron) into an output signal. Typically, activation functions have a "squashing" effect.
Python-DNN currently support following Activation Functions:
-
sigmoid:
Sigmoid function with equation: f(x) = 1/(1 + e^(-x)).This is an S-shaped (sigmoid) curve, with output in the range (0,1).
-
tanh:
The Hyperbolic tangent function is a sigmoid curve, like the logistic function, except that output lies in the range (-1,+1).
-
relu:
The rectifier is an activation function defined as f(x) = max(0, x)
-
cappedrelu:
It is same as ReLU except we cap the units at 6.ie, f(x) = min(max(x,0),6)