Performance Analysis of Various Activation Functions in Neural
Performance Analysis of Various Activation Functions in Neural
Introduction
One of the most attractive properties of ANNs is the possibility to adapt their behavior to the changing characteristics of the modeled system. Last decades, many researchers have investigated a variety of methods to improve ANN performance by optimizing training methods, learn parameters, or network structure, comparably few works is done towards using activation functions.
This function is especially advantageous to use in neural networks trained by backpropagation algorithms. Because it is easy to distinguish, and this can interestingly minimize the computation capacity for training. The term sigmoid means Sshaped, and logistic form of the sigmoid maps
This function is similar to the sigmoid function. For this type of activation function described in below figure, it goes well for applications that produce output values in the range of [-1, 1].
Hyperbolic Tangent Function is similar to sigmoid function. Its range outputs between -1 and 1 as seen in below figure
RBF are typically used to build up function approximations of the form y (x) =wi g(ll x cll) i=1->N where the approximating function y(x) is represented as a sum of N radial basis functions, each associated with a different center ci, and weighted by an appropriate coefficient wi. The weights wi can be estimated using the matrix methods of linear least squares, because the approximating function is linear in the weights. Figure below shows that two unnormalized Gaussian radial basis functions in one input dimension. The basis function centers are located at c1=0.75 and c2=3.25
where ai is input coefficient, ci is the center, wi is the weight in Multi Layered Perceptron (MLP), 2w is an opening angle which can be any value in the range of [-/2, /2] and determines the different forms of the decision borders.
CONCLUSION
In this study, we have used five conventional differentiable and monotonic activation functions for the evolution of MLP architecture along with Generalized Delta rule learning. These proposed well-known and effective activation functions are Bi-polar sigmoid, Uni-polar sigmoid, Tanh, Conic Section, and Radial Bases Function (RBF). Having compared their performances, simulation results show that Tanh (hyperbolic tangent) function performs better recognition accuracy than those of the other functions. In other words, the neural network computed good results when Tanh-Tanh combination of activation functions was used for both neurons (or nodes) of hidden and output layers.