Similar Manifold Learning Based on Selective Cluster Ensemble for Image Clustering
Author:
Affiliation:

Clc Number:

TP181

Fund Project:

National Key Research and Development Program of China (2018YFA070170, 2018YFA0701701); National Natural Science Foundation of China (61672364)

  • Article
  • | |
  • Metrics
  • |
  • Reference [31]
  • |
  • Related
  • | | |
  • Comments
    Abstract:

    Manifold learning is one of the most important research directions nowadays. The performance of manifold learning methods is affected by the choice of reduced dimension. When the reduced dimension is the intrinsic dimension, it is easily to handle the original data. However, intrinsic dimension estimation is still a challenge of manifold learning. In this study, a novel unsupervised method is proposed, called similar manifold learning based on selective cluster ensemble (SML-SCE), which avoids the estimation of intrinsic dimension and achieves a promising performance. SML-SCE generates representative anchors with modified balanced K-means based hierarchical K-means (MBKHK) to construct similarity matrix efficiently. Moreover, multiple similar low-dimensional embeddings in different dimensions are obtained, which are the different presentations of original data. The diversity of these similar low-dimensional embeddings is benefit to the ensemble learning. Therefore, selective cluster ensemble method is taken advantage of as the combination rule. For the clustering results obtained by K-means in similar low-dimensional embeddings, the normalized mutual information (NMI) is calculated between clusterings as weight. Finally, the low weight clusterings is discarded and a selective vote scheme is adopted based on weight to obtain the final clustering. Extensive experiments on several data sets demonstrate the validity of the proposed method.

    Reference
    [1] Keogh E, Mueen A. Curse of dimensionality. In:Encyclopedia of Machine Learning and Data Mining. 2017. 314-315.
    [2] Choi JY, Bae SH, Qiu X, Fox G. High performance dimension reduction and visualization for large high-dimensional data analysis. In:Proc. of the IEEE/ACM Int'l Conf. on Cluster, Cloud and Grid Computing. 2010. 331-340.
    [3] Liu H, Ming S, Sheng L. Infinite ensemble for image clustering. In:Proc. of the ACM SIGKDD Int'l Conf. on Knowledge Discovery and Data Mining. 2016. 1745-1754.
    [4] Lu H, Setiono R, Liu H. Effective data mining using neural networks. IEEE Trans. on Knowledge and Data Engineering, 1996,8(6):957-961.
    [5] Singh A, Ganapathysubramanian B, Singh AK. Machine learning for high-throughput stress phenotyping in plants. Trends in Plant Science, 2016,21(2):110-124.
    [6] Li FZ, Zhang L, Zhang Z. Lie Group Machine Learning. Walter de Gruyter GmbH and Co KG, 2018.
    [7] Zhao Y, You X, Yu S. Multi-view manifold learning with locality alignment. Pattern Recognition, 2018,78:154-166.
    [8] Tenenbaum JB, Silva VD, Langford JC. A global geometric framework for nonlinear dimensionality reduction. Science, 2000,290(5500):2319-2323.
    [9] Roweis ST, Saul LK. Nonlinear dimensionality reduction by locally linear embedding, Science, 2000,290(5500):2323-2326.
    [10] Belkin M, Niyogi P. Laplacian eigenmaps and spectral techniques for embedding and clustering. In:Proc. of the Int'l Conf. on Neural Information Processing Systems:Natural and Synthetic. 2002. 585-591.
    [11] He XF, Niyogi P. Locality preserving projections. Advances in Neural Information Processing Systems, 2003,16(1):186-197.
    [12] Cai D, He XF, Han J. Spectral regression:A unified subspace learning framework for content-based image retrieval. In:Proc. of the ACM Int'l Conf. on Multimedia. 2007,60:403-412.
    [13] Nie FP, Zhu W, Li XL. Unsupervised large graph embedding. In:Proc. of the 31st AAAI Conf. on Artificial Intelligence. 2017. 2422-2428.
    [14] Li YY. Curvature-aware manifold learning. Pattern Recognition, 2018,83:273-286.
    [15] Costa JA, Hero AO. Learning intrinsic dimension and intrinsic entropy of high-dimensional datasets. In:Proc. of the European Signal Processing Conf. 2004. 369-372.
    [16] Zhou ZH. When semi-supervised learning meets ensemble learning. Frontiers of Electrical and Electronic Engineering in China, 2011,6(1):6-16.
    [17] Tang W, Zhou ZH. Bagging-based selective clusterer ensemble. Ruan Jian Xue Bao/Journal of Software, 2005,16(4):496-502(in Chinese with English abstract).[doi:10.1360/jos160496]
    [18] Tesmer M, Perez CA, Zurada JM. Normalized mutual information feature selection. IEEE Trans. on Neural Networks, 2009, 20(2):189-201.
    [19] Zhu W, Nie FP, Li X. Fast spectral clustering with efficient large graph construction. In:Proc. of the IEEE Int'l Conf. on Acoustics, Speech and Signal Processing. 2017. 2492-2496.
    [20] Cox TF, Cox MA. Multidimensional Scaling. Chapman and Hall, 2000.
    [21] Liu W, He J, Chang SF. Large graph construction for scalable semi-supervised learning. In:Proc. of the Int'l Conf. on Machine Learning. 2010. 679-686.
    [22] Nie FP, Wang X, Jordan MI, Huang H. The constrained Laplacian rank algorithm for graph-based clustering. In:Proc. of the 30th AAAI Conf. on Artificial Intelligence. 2016. 1969-1976.
    [23] Jain AK, Law MHC. Data clustering:A user's Dilemma. In:Proc. of the Int'l Conf. on Pattern Recognition and Machine Intelligence. 2005,3776:1-10.
    [24] Jolliffe IT. Principal component analysis. Journal of Marketing Research, 2002,87(100):513.
    [25] Luo XH, Zhang L, Li FZ, Wang BJ. Graph embedding-based ensemble learning for image clustering. In:Proc. of the 24th Int'l Conf. on Pattern Recognition. 2018. 213-218.
    [26] Luo XH, Zhang L, Li FZ, Hu CX. Unsupervised ensemble learning based on graph embedding for image clustering. In:Proc. of the Int'l Conf. on Neural Information Processing. 2018,11303:38-47.
    [27] Alimoglu F, Alpaydin E, Denizhan Y. Combining multiple classifiers for pen-based handwritten digit recognition. In:Proc. of the 4th Int'l Conf. on Document Analysis and Recognition. 1996,2:637-640.
    [28] Nene SA, Nayar SK, Murase H. Columbia object image library (coil-20). Columbia University, 1996. http://www.cs.columbia.edu/CAVE/software/softlib/coil-20.php
    [29] Leibe B, Schiele B. Interleaving object categorization and segmentation. Cognitive Vision Systems, 2006,3948:145-161.
    附中文参考文献:
    [17] 唐伟,周志华.基于Bagging的选择性聚类集成.软件学报,2005,16(4):496-502.[doi:10.1360/jos160496]
    Related
    Cited by
Get Citation

罗晓慧,李凡长,张莉,高家俊.基于选择聚类集成的相似流形学习算法.软件学报,2020,31(4):991-1001

Copy
Share
Article Metrics
  • Abstract:
  • PDF:
  • HTML:
  • Cited by:
History
  • Received:May 29,2019
  • Revised:August 01,2019
  • Online: January 14,2020
  • Published: April 06,2020
You are the first2050444Visitors
Copyright: Institute of Software, Chinese Academy of Sciences Beijing ICP No. 05046678-4
Address:4# South Fourth Street, Zhong Guan Cun, Beijing 100190,Postal Code:100190
Phone:010-62562563 Fax:010-62562533 Email:jos@iscas.ac.cn
Technical Support:Beijing Qinyun Technology Development Co., Ltd.

Beijing Public Network Security No. 11040202500063