A multi-level abstraction model for competitive learning neural networks
2008
Competitive learning neural networks are powerful analytical tools for data clustering and topology-preserving visualization. However, they are limited in the sense of being unable to achieve more than one task on the same network. When applied to clustering tasks, every neuron unit is supposed to represent one of the inherent data clusters, while learning topology requires much more neuron units. The aim of this work is to figure out how connections between neurons -- specially those inserted by a competitive Hebbian learning -- can be exploited to construct higher abstraction levels of a topology-preserving neural network. The idea is to devote the basic level of the network to detailed description of the data while taking advantage of higher abstraction levels to facilitate quantitative analysis and structural overview of the network. The abstraction is done in two different fashions: macro and micro abstractions. In macro abstraction, the objective is to cluster the first-level neuron units into their principal groups corresponding to the clusters inherent in the data. In contrast, micro abstraction is designed to capture underlying clusters according to a given granularity degree. The abstraction-level units are also connected to reflect a simplified structure of the data distribution. At the end, the basic level of the network can be hierarchically organized with one or more abstraction levels, permitting both qualitative and quantitative analysis of the data. Simulations on many synthetic data shown the relevance of proposed model.
Keywords:
- Correction
- Source
- Cite
- Save
- Machine Reading By IdeaReader
7
References
0
Citations
NaN
KQI