logo
    Centroid Neural Network with Bhattacharyya Kernel
    0
    Citation
    0
    Reference
    20
    Related Paper
    Abstract:
    A clustering algorithm for Gaussian Probability Distribution Function (GPDF) data called Centroid Neural Network with a Bhattacharyya Kernel (BK-CNN) is proposed in this paper. The proposed BK-CNN is based on the unsupervised competitive Centroid Neural Network (CNN) and employs a kernel method for data projection. The kernel method adopted in the proposed BK-CNN is used to project data from the low dimensional input feature space into higher dimensional feature space so as the nonlinear problems associated with input space can be solved linearly in the feature space. In order to cluster the GPDF data, the Bhattacharyya kernel is used to measure the distance between two probability distributions for data projection. With the incorporation of the kernel method, the proposed BK-CNN is capable of dealing with nonlinear separation boundaries and can successfully allocate more code vector in the region that GPDF data are densely distributed. When applied to GPDF data in an image classification probleml, the experiment results show that the proposed BK-CNN algorithm gives 1.7%-4.3% improvements in average classification accuracy over other conventional algorithm such as k-means, Self-Organizing Map (SOM) and CNN algorithms with a Bhattacharyya distance, classed as Bk-Means, B-SOM, B-CNN algorithms.
    Keywords:
    Bhattacharyya distance
    Centroid
    Feature vector
    Kernel (algebra)
    Feature (linguistics)
    In this paper, two new clustering algorithms based on fuzzy c-means for data with tolerance are proposed. Kernel functions which map the data from the original space into higher dimensional feature space are introduced into the proposed algorithms. Nonlinear boundary of clusters can be easily found by using the kernel functions. First, two clustering algorithms for data with tolerance are introduced. One is based on standard method and the other is on entropy-based one. Second, two objective functions in feature space are shown corresponding to two methods, respectively. Third, Karush-Kuhn-Tucker conditions of two objective functions are considered, respectively, and these conditions are re-expressed with kernel functions as the representation of an inner product for mapping from original pattern space into higher dimensional feature space than the original one. Last, two iterative algorithms are proposed for the objective functions, respectively.
    Kernel (algebra)
    String kernel
    Feature vector
    FLAME clustering
    A robust fundamental matrix estimation algorithm based on kernel fuzzy clustering is proposed.The residual features of match points are detected by this algorithm.Using the kernel functions,the nonlinear dividable features in the original one-dimensional space can be mapped to a high-dimensional feature space,in which the clustering can be performed effectively.The match points are classified into inliers and outliers by the fuzzy clustering method.The inliers and the outliers are modeled by Gaussian function,and the judgment values of the divisibility of the two sets are defined and calculated.The iteration will continue until the judgment value converges.Experimental results of the fundamental matrix computation on both simulated and real data demonstrate the superiority of the proposed algorithm in precision and efficiency over classical RANdom SAmple Consensus(RANSAC) algorithm.
    RANSAC
    Eight-point algorithm
    Kernel (algebra)
    Citations (0)
    Similarity matrix is critical to the performance of spectral clustering. Mercer kernels have become popular largely due to its successes in applying kernel methods such as kernel PCA. A novel spectral clustering method is proposed based on local neighborhood in kernel space (SC-LNK), which assumes that each data point can be linearly reconstructed from its neighbors. The SC-LNK algorithm tries to project the data to a feature space by the Mercer kernel, and then learn a sparse matrix using linear reconstruction as the similarity graph for spectral clustering. Experiments have been performed on synthetic and real world data sets and have shown that spectral clustering based on linear reconstruction in kernel space outperforms the conventional spectral clustering and the other two algorithms, especially in real world data sets.
    Spectral Clustering
    Kernel (algebra)
    Similarity (geometry)
    Citations (8)
    In this paper, we present a scheme of similarity measure learning based on kernel optimization. Employing a data-dependent kernel model, the proposed scheme optimizes the spatial distribution of the training data in the feature space, aiming to maximize the class separability of the data in the feature space. The learned similarity measure, derived from the optimized kernel, exhibits a favorable feature to the task of pattern classification, that the spatial resolution of the embedding space is expanded around the boundary areas, and shrunk around the homogeneous areas. Experiments demonstrate that using the learned similarity measure can substantially improve the performances of the K-nearest-neighbor classifier.
    Similarity measure
    Kernel (algebra)
    Feature vector
    Similarity (geometry)
    Multiple kernel learning
    Citations (0)
    Kernel clustering algorithms have the ability to capture the non-linear structure inherent in many real world data sets and thereby, achieve better clustering performance than Euclidean distance based clustering algorithms. However, their quadratic computational complexity renders them non-scalable to large data sets. In this paper, we employ random Fourier maps, originally proposed for large scale classification, to accelerate kernel clustering. The key idea behind the use of random Fourier maps for clustering is to project the data into a low-dimensional space where the inner product of the transformed data points approximates the kernel similarity between them. An efficient linear clustering algorithm can then be applied to the points in the transformed space. We also propose an improved scheme which uses the top singular vectors of the transformed data matrix to perform clustering, and yields a better approximation of kernel clustering under appropriate conditions. Our empirical studies demonstrate that the proposed schemes can be efficiently applied to large data sets containing millions of data points, while achieving accuracy similar to that achieved by state-of-the-art kernel clustering algorithms.
    Data stream clustering
    Kernel (algebra)
    Constrained clustering
    Clustering high-dimensional data
    Citations (87)
    Anomaly detection is an important research direction in the field of data mining and industrial dataset preprocess. The paper proposed a kernel neighbor density definition with parallel computing mechanism for anomaly detection algorithm. The kernel neighbor density formula calculates the density of points in high dimensional space. In our definition, we adopt the median operation because the breakdown point of the median is the largest possible. So the definition could be a very robust estimate of the data location, and parallel computing mechanism is introduced to improve the efficiency of algorithms. We use two real datasets and three different kernel functions to evaluate the performance of algorithms. The experiment results confirm that the presented definition of kernel neighbor density improves the performance of algorithms and the Gaussian kernel function has the best effect.
    Kernel (algebra)
    Kernel density estimation
    Anomaly (physics)