Hyperconic multilayer perceptron for function approximation

2015 
In this paper, three different multilayer perceptrons are used to solve function approximation problems. We compare the performance of the traditional Multilayer Perceptron, the Hypersphere Multilayer Perceptron and our own Hyperconic Multilayer Perceptron in the approximation of six benchmark functions. The Hypersphere Multilayer Perceptron and the Hy-perconic Multilayer Perceptron architectures include neurons whose activation transfer functions produce non-linear (non-piecewise linear) boundaries. The goal of this paper is to evaluate the advantages of the neural network models, whether they need a greater or smaller number of neurons in the hidden layer, with particular interest is higher order neurons, in order to approximate continuous functions. The training of the neural network models is performed by using an evolutionary algorithm designed by the authors and which is based on spherical inversions. The non-linear nature of the reproduction operator of the evolutionary algorithm furnishes an enhanced search capability to obtain the global optimum during the training stage.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    10
    References
    1
    Citations
    NaN
    KQI
    []