Figure 1 : Multilayer perceptron
The Multilayer Perceptron
The multilayer perceptron (MLP) is a hierarchical structure of several perceptrons, and overcomes the shortcomings of these single-layer networks.
The multilayer perceptron is an artificial neural network that learns nonlinear function mappings. The multilayer perceptron is capable of learning a rich variety of nonlineardecision surfaces.
Nonlinear functions can be represented by multilayer perceptrons with units that use nonlinear activation functions. Multiple layers of cascaded linear units still produce only linear mappings.
Differentiable Activation Functions
The training algorithm for multilayer networks requires differentiable, continuous nonlinear activation functions. Such a function is the sigmoid, or logistic function:
where s is the sum: s=S i=0d wi xi of products from the weights wi and the inputs xi.
Sometimes s is called alternatively squashing function as it maps a very large input domain to a small range of outputs.
Another nonlinear function often used in practice is the hyperbolic tangent:
o = tanh( s ) = ( es - e-s ) / (es + e-s)
No comments:
Post a Comment