Activation Function: Deep Neural Networks
Activation Function: Deep Neural Networks
Activation Function: Deep Neural Networks
y(x) = F(x)
Deep refers to number of
hidden layers.
Biological neuron & mathematical neuron
Activation Functions
The activation function of a node defines the output of that node given an
input or set of inputs.
In Biologically inspired the activation function is usually an abstraction
representing the rate of action potential firing in the cell.
These functions should be nonlinear to encode complex patterns of the data.
The activation functions used in deep neural network are multi-state
activation,Sigmoid, Tanh,ReLU.
Multi-state activation functions[1]
The MSAFs reveal that these activation functions have potentials for altering
the parameter distribution of the DNN models, improving model
performances and reducing model sizes.
❖Identity Function
❖Step Function
❖Logistic Function(Sigmoid Function)
❖TanH
❖Arch Tan
❖Rectified Linear Unit(ReLU’s)
❖Soft Plus
❖Leaky Rectified Linear Unit(LReLU’s)
❖Parameteric rectified linear unit (PReLU)
❖Randomized leaky rectified linear unit (RReLU)
❖Exponential linear unit (ELU)
❖S-shaped rectified linear activation unit (SReLU)
❖Adaptive piecewise linear (APL)
❖SoftExponential
➢Identity Function
f(x)=x
Derivative of f(x)
f ‘(x)=1
Range(-∞ , ∞)
It is called Linear activation Function.
➢Binary Step Function:
Derivative of f(x)
0 for x ≠ 0
f ‘(x) = ? for x ≥ 0
Range{0 , 1}
➢Sigmoidal Function:
Derivative of
Range(0 , 1)
It normalizes an input real value into range between 0 and
1
➢TanH Function:
f(x) = tanh(x) =
Derivative of
Range(-1 , 1)
The tanh(x) function is a rescaled version of the sigmoid,
and its output range is [−1 , 1] instead of [0 , 1].
➢Rectified Linear Unit:
Derivative of f(x)
0 for x < 0
f ‘(x) = 1 for x ≥ 0
Range{0 , ∞}
ReLUs improve neural networks is by speeding up training. The gradient
➢Soft Plus:
Derivative of f(x)
Range(0,∞)
The softplus function can be approximated by max
function (or hard max) i.e max(0,x+N(0,1))
➢Exponential Linear Unit:
Derivative of f(x)
Range(-α , ∞)
Exponential linear unit (ELU) which speeds up learning in deep neural
networks and leads to higher classification accuracies.
Application
Deep Neural Network Activation Functions
Activation Function
1. Logistic Function
2. SoftMax Function
3. Rectifier Linear Unit(ReLU)
Logistic Function
Where,
e –> natural logarithm base (Euler’s Number)
X0 –> x – value of sigmoid’s midpoint ( -∞ to +∞)
L – > Curve’s maximum value
k – > steepness of the curve.
Logistic Function Curve
Application
• Various fields of where LOGISTIC function can be used
1. Biomathematics
2. Chemistry
3. Economics
4. GeoScience
5. Probability
6. Sciology
7. linguistic
8. Statics
9.Ecology
10. Medicine
Ecology
• Population growth
1. Logistic function Equation is common
model for Population Growth
2. Rate of Reproduction is directly
proportional to Existing Population and
amount of available resources
3. Equation
where,
P-> Population Size
r -> growth rate
t -> time
Medicine
• Where,
x -> input to a neuron
- This function also Ramp function
- Analogous to half-wave rectifier
Variants of ReLU
• Noisy ReLU
• Speech Recognition
SoftMax Function
.
SoftMax Function
Data set used Number of data used Combination of Activation Function used
function(B,U or H).
DNN with two activation function
Among 15 DNN, the DNN with U-B-H is the best with the minimum average
testing MSE (0.0351) for Dow Jones Industrial Average data.
Thus DNN using different activation functions can perform better than one
using a single function.
Trainable Activation Functions [3]
The nonlinear activation function of the ith unit in the lth layer, σi power(l) (x), can be represented in
the power series as follows:
N is the approximation degree and a power(l)i,l are
coefficients to be retrained
Proposed activation function
Retrain activation function
[2] Genetic Deep Neural Networks Using Different Activation Functions for Financial Data
Mining.(©2015 IEEE Luna M. Zhang Soft Tech Consulting, Inc. Chantilly, USA).
[3] Deep Neural Network Using Trainable Activation Functions.(©2016 IEEE Hoon Chung, Sung Joo
Lee and Jeon Gue Park Electronics and Telecommunications Research Institute, Daejeon, Korea)
Thank you!