Successive Adaptive Linear Neural Modeling for
Equidistant Real Roots Finding
ETCM 2018 (Oral Presentation)
- Joseph Gonzalez Yachay Tech University
- Fernando Zhapa Yachay Tech University
- Oscar Guarnizo Yachay Tech University
- Francisco Ortega-Zamorano Yachay Tech University
Abstract
This work aims to implement a model to find equidistant real roots using a Successive Adaptive Linear Neural Modeling, consisting of two models: a Self Organized Map (SOM) and an Adaptative Linear Neuron (Adaline). A SOM model has been used with a new neighborhood function and a physical distance. The SOM model can delimit the areas where a single root exists, dividing the task at hand into sub-processes and reducing the complexity. Then, we defined a neural model based on single Adaline neurons with a pocket. This model is applied sequentially to each pair of regions to find the real root values with reduced precision. Finally, several experiments were done considering CPU time, relative error, the distance between the roots, and polynomial degrees. The results show that the time complexity grows in a logarithmic way. Also, the error does not increase at a higher rate than the polynomial degree or the root distance.
Video
Self Organized Map (SOM)
The random samples from a given function (which represents a given equation) are taken as initial patterns. Then, a self-organized map will receive the patterns and classify them into regions of possible roots. The topology of the SOM is simply linear because the patterns are distributed around the x-axis. There are n + 1 neurons because there are n roots dividing a whole space, which will produce n + 1 regions (one region per neuron).
We used a neighborhood function during the training in order to update five neurons, the winning neuron and the closest neurons on both sides, two on the left and two on the right. Then, the neurons are multiplied by 1, 0.5, or 0.1 depending on the topological distance (0 winning neuron, 1 the first closest, and 2 the second closes).
Then, the training used the following equation
Adaline
After finding the regions with SOM, a dimension is added into the ouput patterns given the regions distribution. Then, an Adaptive Linear Neuron (Adaline) is launched n times (per pair of regions) to find the n respective roots. Additionally, Adaline used a pocket to store the weight of neurons with the best accuracy for improving performance.
Equidistan Roots
This modeling only works for real equidistant roots because the SOM allows us to divide the space into (quasi) equally
distributed regions.
It can also be written as
with a root constraint
Results
This model is a two-process and we check their behavior visually as bellow or in the following video.
We evaluated our method with polynomials of different degrees from 2 to 10, calculating its root approximation, CPU time, and relative error.
Related links
- For more elaborated approaches for finding roots in polynomials using neural networks, check A Neural Root Finder of Polynomials Based on Root Moments
and A constructive approach for finding arbitrary roots of polynomials by neural networks by De-Shuang Huan et al.
Citation
Powered by Jon Barron and Michaƫl Gharbi.