一类模糊聚类算法研究及其应用
详细信息    本馆镜像全文|  推荐本文 |  |   获取CNKI官网全文
摘要
聚类分析作为一个重要的工具已经广泛应用于多个领域(?)模糊聚类算法由于具有良好的聚类性能与数据表达能力,已经成为近年来研究的热点.本文对当前主要的模糊聚类算法进行了研究,针对这些算法中存在的不完善之处提出了相应的改进算法,并对一类基于核的模糊聚类算法的收敛性给出了理论上的证明.本文所作的工作归纳起来主要有以下几点:
     一、深入研究了FCM类算法,提出了一种新的基于核的模糊聚类模型(IKFCM聚类模型),并得到三种不同形式的IKFCM聚类算法-IKFCM1、IKFCM2和IKDFCM算法.IKFCM1、IKFCM2算法通过核函数将数据映射到高维的特征空间,提高了算法发现非线性可分形状聚类结构的能力.IKDFCM算法利用核化距离作为聚类的相异性测度,对噪声与野值点有着更好的鲁棒性,计算的时间空间复杂度相对较低.
     二、证明了IKFCM算法与基于核的FCM算法(KFCM)的收敛性,这是对原有非核聚类算法收敛性定理的一种推广.收敛性定理表明特征空间内的此类模糊聚类算法的收敛性与数据核矩阵的秩之间有着密切的关系,核化距离形式算法的收敛性与核函数的凸性之间存在着密切的关系.
     三、提出了带有凸包约束的(核)可能性聚类模型,并引入全局优化技术对提出的模型进行求解,较好的解决了原始算法容易陷入局部极值与鞍点的问题.提出的算法对解的可行域进行限制,克服了原始算法中易产生重合聚类的不足,并且比普通的基于优化技术的(核)可能性聚类具有更高的效率.
     四、提出了利用迭代不动点的吸引域进行聚类的想法,并引入了一种新的聚类有效性指标,得到了一种新的均值漂移聚类算法及其快速算法,算法避免了FCM类算法中人为对初始中心作出假设的不足,并实现了对大数据集的聚类.
With the development of the internet and information technology, human access more and more data every day. An important tool to deal with these data is cluster analysis. Compared with traditional methods of hard clustering, fuzzy clustering has a better ability to describe the data structure, which makes it become the mainstream of the clustering algorithms. In 1969, Ruspini first introduced the concept of fuzzy set theory into the cluster analysis fields. Since then many clustering methods based on fuzzy set theory have been proposed. The objective function based fuzzy clustering algorithms are the most popular used method in fuzzy clustering. These algorithms obtain the fuzzy partitions and clustering results of the data sets by solving a nonlinear programming problem. FCM-type algorithms (here mainly refer to the fuzzy objective function based partition clustering algorithms which are derived from FCM) are one of the most important classes of fuzzy clustering algorithms. A great deal of research has been conducted on FCM-type algorithms. In 2008, Miyamoto et al. mainly discussed the latest developments and practical applications of FCM-type algorithms in their works.
     This dissertation mainly focus on the studying of the FCM-type algorithms, and some algorithms are proposed to overcome the defect of the original algorithms. We prove the convergence theorem of a class of kernel based fuzzy clustering algorithms.
     FCM model is a probabilistic type constraint clustering model. It has been successfully applied into many areas. In the literature, there are a large number of researches have been conducted on FCM, they mainly focuse on the improvements of the model, parameter selection, study of the validity of clustering and how to obtain better solutions of the optimization problems. In these algorithms, the most two important algorithms are the possibilistic c-means clustering algorithm (PCM), which was proposed by Krishnapuram and Keller, and the possibilistic fuzzy c-means clustering algorithm (PCM), which was proposed by Pal and Bezdek.
     In FCM, the membership for a datum crossing all classes sums to 1, which makes its membership disable to reflect the real distance from the datum to the cluster center, and this is also the reason why FCM is sensitive to noise and outliers. To address this problem, PCM relaxes the constraint of FCM, and a penalty item is added on the objective function to avoid the meaningless optimal solutions. Although PCM solves the noise sensitivity problem of FCM, it also generates some new problems. Solving the optimization problem of PCM is equivalent to solve a number of independent sub-optimization problems, and each sub-optimization problem is dependent on only one center, which makes it tend to generate coincident clusters. From the experiments we found that, the global optimal solution is obtained only when all the clusters are coincident, that is to say we must try to obtain suboptimal solutions to get meaningful partitions. It obviously violates the target of designing the model. To solve the problems in FCM and PCM, Pal and Bezdek put forward the possibilistic fuzzy c-means clustering model. PFCM have the advantages of both FCM and PCM, in the mean while, it avoids the defects of both algorithms, i.e., PFCM does not generate coincident clusters and it also has a good noise robustness. There are c + 5 parameters which are set by the user in PFCM, however, how to set these parameters still has no theoretical basis, which makes PFCM have a strong dependence on the parameter settings. FCM, PCM and PFCM algorithms tend to find compact spherical clusters, they lacks the ability of indicating non-convex cluster structures of the data set. To address this issue, kernel technique is introduced into such algorithms, and kernel-based FCM, PCM and PFCM algorithm have been proposed. Such algorithms have the ability of finding the clusters with non-convex structures, but our experiments show that these algorithms also inherit the defect of the original algorithms, such as sensitivity to noise (in FCM), tending to generate coincident clusters (in PCM) and having more parameters (in PFCM).
     Considering the above problems, we put forward a kernel based improved fuzzy c-means clustering model, and three different forms algorithms (IKFCM1, IKFCM2 and IKDFCM algorithm) are obtained from this model. IKFCM1 and IKFCM2 have the ability of finding non-convex clusters, and IKDFCM has a better noise robustness and lower time complexity. The contrast experiments with the above algorithms show the clustering performance of the proposed algorithm.
     The above discussed algorithms still exist or partly exist the following problems: (1) The algorithm requires the assumption on the number of clusters and can not execute unsupervised clustering. (2) Solving the optimization problem of the clustering model is equivalent to solving a number of independent sub-optimization problems, and each sub-optimization problem is dependent on only one center, which makes it tend to generate coincident clusters. (3) These algorithms eventually boil down to the optimization problem of minimizing the objective function, and the traditional iterative method can not guarantee to converge to the global optimal solution, although it can be solved through the introduction of global optimization methods, it requires high time complexity to get precise global optimum. To solve the above problems, we put forward GKPCM (Generalized Kernel Possibilistic C-Means clustering) clustering model, which is a generalization of the conventional possibilistic and kernel based possibilistic clustering model. By limiting the feasible region of the solutions and using the global optimization techniques (here uses the simulated annealing as an example) to solve the optimization problem, GKPCM inherits the advantages of robustness to noise and avoids the problem of generating coincident clusters, and can well find the global optimal solution, and speed up the convergence speed by decreasing the search range of the global optimization techniques.
     Kernel-based fuzzy clustering is a new branch of fuzzy clustering algorithms, both of two important recent books of fuzzy clustering algorithms have sole part to introduce the kernel based fuzzy clustering algorithms. Although kernel clustering algorithm has many good properties, the convergence of such algorithms is still not studied. It is well-known that convergence is the basis for the iterative algorithms. In this dissertation, we establish the convergence of different algorithms that are derived from two kernel based fuzzy clustering model (including KFCM and IKFCM model). Take the KFCM for example, we give the following convergence theorem of KFCM2 and KDFCM:
     Theorem 0.1 (convergence of KFCM2) Let X contain at least n≥c≥2 points, fuzzy factor m > 1 , the rank of kernel matrix K is denoted by R(K) which satisfies R(K)≥c, let U~(0) be the starting point of iteration with U~(0)∈M_(fc1) , let {U~((k))}_(k=1,2,…)denote the sequence generated by the KFCM2 algorithm, then {U~((k))}_(k=1,2,…) either terminates at a point in (?) or there is a subsequence converging to a point in (?), where (?); (1)(?). (2)
     Theorem 0.2 (convergence of KDFCM ) Let X contain at least n > c > 2 different points , fuzzy factor m > 1 , K is the kernel function,θ∈R~s, f(z;θ) = K(z, z) -2K(θ, z) is a convex function with respect to z , let V~((0)) be the starting point of iteration with V~((0))∈R~s , let {U~((k)), V~((k))}_(k=1,2,…) denote the sequence generated by the KDFCM algorithm, then {U~((k)), V~((k))}_(k=1,2,…) either terminates at a point in (?) or there is a subsequence converging to a point in (?), where (?)= {(U~*, V~*)∈M_(fc1)×R~(sc) |(?); (3)(?). (4)
     The other algorithms of different forms of IKFCM and KFCM model also have the similar convergence theorems. The convergence theorem of FCM and IFCM can be obtainned from the proved theorems of KFCM and IKFCM if we set the kernel function for some special forms. Hence the convergence theorem of KFCM is a generalization of the original FCM algorithm, and the convergence theorem of IKFCM is the first proof for the IFCM algorithm. The theorems show that the convergence of IKFCM and KFCM has a close relations with the ranks of the kernel matrix and the convexity of the kernel functions, which also provide us a theoretical basis of selecting reasonable kernel functions when executing such algorithms.
     An important issue of the objective function based fuzzy clustering algorithms is that such algorithm is sensitive to the initializations. Inspired by the convergence theorem 6.2.1, we propose a new thought of partitioning the data by the domain of attraction of the fixed point. Based on this idea, a new mean shift clustering algorithm called unsupervised multi-scale fuzzy clustering (UMF) is put forward, and a cluster validity is proposed to determine the final partition of UMF. Compared to the traditional fuzzy clustering, UMF has the following characteristics: UMF is not influenced by the initializations of the cluster centers; it can indicate the data structure from different "partition scales"; it can execute unsupervised clustering for the given data set. A fast vision of UMF is also presented in order to meet the need of dealing with the large data sets. Finally, we apply UMF to improve the EM based mixture-resolving methods and the fuzzy clustering algorithms, the experimental results show the performance of UMF.
引文
[1] 丁泽进,于剑.聚类分析技术综述,机器学习及其应用(王珏,周志华,周傲英)[G].北京:清华大学出版社,2006,4:79-80.
    [2] 冯果忱.非线性方程组迭代解法[M].上海:上海科学技术出版社出版,1989.
    [3] 罗会兰.聚类集成关键技术研究[D].浙江:浙江大学计算机学院, 2007.
    [4] Siegelmann H T, Vapnik V, Hur A B, Horn D. Support vector clustering [J]. J. Mach. Learn. Res, 2001(2): 125-137.
    [5] Gunopulos D, Agrawal R, Gehrke J. Automatic subspace clustering of high dimensional data for data mining applications [J]. ACM SIGMOD Int. Conf. on Management of Data, Seattle, WA., 1998, 94-105.
    [6] Suganthan P N, Qinand A K. Kernel neural gas algorithms with application to cluster analysis [C]. ICPR -17th International Conference on Pattern Recognition, 2004 (4): 617-620.
    [7] Selim S Z, Al-Sultan K S. A global algorithm for the fuzzy clustering problem [J]. Pattern Recognition, 1993, 26(9): 1357-1361.
    [8] Fred A L N. Finding consistent clusters in data partitions [C]. Multiple Classifier Systems, Second International Workshop, MCS 2001, Cambridge, UK., 2001, 309-318.
    [9] Anderberg M. Cluster analysis for applications [M]. New York: Academic, 1973.
    [10] Esogbue A O. Optimal clustering of fuzzy data via fuzzy dynamic programming [J]. Fuzzy Sets and Systems, 1986, (18): 283-298.
    [11] Landau S, Everitt B, Leese M. Cluster analysis [M]. London: Arnold, 2001.
    [12] Miiller K R, Schokopf B, Smola A J. Nonlinear component analysis as a kernel eigenvalue problem [J]. Neural Comput, 1998, 10(5): 1299-1319.
    [13] Backer E, Jain A. A clustering performance measure based on fuzzy set decomposition [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1981, PAMI-3(1): 66-75.
    [14] Ball G, Hall D. A clustering technique for summarizing multivariate data [J]. Behav. Sci, 1967, 12: 153-155.
    [15] Baraldi A, Blonda P. A survey of fuzzy clustering algorithms for pattern recognition- Part Ⅰ andⅡ [J]. IEEE Transactions on Systems Man and Cybernetics Part B-Cybernetics, 1999, 29(6): 778-801.
    [16] Baraldi A, Schenato L. Soft-to-hard model transition in clustering: A review [R]. Tech. report, TR-99-010, 1999.
    [17] Dobkin D P, Barber C B, Huhdanpaa H. The quick algorithm for convex hulls [J]. Acm Transactions on Mathematical Software, 1996, 22(4): 469-483.
    [18] Siegelmann H T, Vapnik V, Ben-Hur A, Horn D.Support vector clustering [J]. Journal of Machine Learning Research, 2001, 2: 125-137.
    [19] Ressel P, Berg C. Christensen J P R. Harmonic analysis on semigroups [M]. New York: Springer-Verlag, 1984.
    [20] Berkhin P. survey of clustering data mining techniques [EB/OL]. [2001-4-15] http://www.accrue.com/products/rp_cluster_review.pdf.
    [21] Harris J O, Bezdek J C. Convex decompositions of fuzzy partitions [J]. J. Math. Anal.& Appl, 1979.
    [22] Bishop C. Neural networks for pattern recognition [M]. New York: Oxford Univ. Press, 1995.
    [23] Vapnik V, Cortes C. Support vector networks [J]. Mach. Learn., 1995, 20: 273-297.
    [24] Celeux G, Govaert G. A classification em algorithm for clustering and two stochastic versions [J]. Comput. Statist. Data Anal., 1992, 14: 315-332.
    [25] Zhou L, Chen N, Chen A. An incremental grid density-based clustering algorithm [J]. Journal of Software, 2002, 13(1): 1-7.
    [26] Cheng Y Z. Mean shift, mode seeking, and clustering [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1995, 17(8): 790-799.
    [27] Mulier F, Cherkassky V. Learning from data concepts, theory and methods [M]. John Wiley & Sons, 1998.
    [28] Hao P, Chiang J. A new kernel-based fuzzy clustering approach: Support vector clustering with cell growing [J]. IEEE Transactions on Fuzzy Systems, 2003, 11(4): 518-527.
    [29] Zahn C T. Graph-theoretical methods for detecting and describing gestalt clusters [J]. IEEE Transactions on Computers, 1971, 20(1), 68-86.
    [30] Prade H, Dubois D. Possibility theory [M]. New York: Plenum Press, 1988.
    [31] Fyfe C, Macdonald D. The kernel self-organising map [C], Fourth International Conference on Knowledge-Based Intelligent Engineering Systems and Allied Technologies, 2000, 1: 317-320.
    [32] Chen S C, Zhang D Q. Fuzzy clustering using kernel method [C]. The 2002 International Conference on Control and Automation, ICCA, 2002, 162-163.
    [33] Chen S C, Zhang D Q. Kernel based fuzzy and possibilistic c-means clustering [C]. Proceedings of the International Conference Artificial Neural Network, Turkey, 2003, 122-125.
    [34] Chen S C, Zhang D Q. A novel kernelized fuzzy c-means algorithm with application in medical image segmentation [J]. Artif. Intell. Med., 2004, 32(1): 37-50.
    [35] Rajesh N, Davé, Raghu Krishnapuram. Robust clustering methods: A unified view [J]. IEEE Transactions on Fuzzy Systems, 1997, 5(2): 270-293.
    [36] Valverde L, Mantaras De R L. New results in fuzzy clustering based on the concept of indistinguishability relation [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1988, 10(5): 754-757.
    [37] Zha H, Ding C, He X. Spectral min2max cut for graph partitioning and data clustering [C]. Proc. of the IEEE Intl Conf. on Data Mining, 2001, 107-114.
    [38] Fridlysand J, Dudoit. Bagging to improve the accuracy of a clustering procedure [J]. Bininformatics, 2003, 19(9), 1090-1099.
    [39] Schikuta E. Grid-clustering: An efficient hierarchical clustering method for very large data sets [C]. Proceedings of the 13th Intrnational Conference on Pattern Recognition, 1996, 101-105.
    [40] Ruspini E H. A new approach to clustering [J], Information and control, 1969, 15(1): 22-32.
    [41] Verri A, Camastra F. A novel kernel method for clustering [J]. IEEE Trans. Pattern Anal. Mach. Intell., 2005, 27(5), 801-804.
    [42] Fasulo D. An analysis of recent work on clustering algorithms [R]. Dept. Comput. Sci. Eng., Univ.Washington, Seattle,WA, 1999.
    [43] Filippone M, Camastra F, Masulli F, Rovetta S. A survey of kernel and spectral methods for clustering [J]. Pattern Recognition, 2008, 41(1): 176-190.
    [44] Fisher D. Knowledge acquisition via incremental conceptual clustering [J]. Mach. Learn., 1987, 2: 139-172.
    [45] Perkal J, Zubrzycki S, Florek K, Lukaszewicz J. Sur laliaison et la division des points d'un ensemble fini [J]. Colloquium Mathematicae, 1951, 2: 282-285.
    [46] Fraley C, Raftery A. Model-based clustering, discriminant analysis and density estimation [J]. J. Amer. Statist. Assoc, 2002, 97: 611-631.
    [47] Jain A K, Fred A. Robust data clustering [C]. Proc. of IEEE Computer Society Conf. on Computer Vision and Pattern Recognition (CVPR'03), Wisconsin, USA, 2003, 128-136.
    [48] Jain A K, Fred A L N. Data clustering using evidence accumulation [C], Proe. of the 16th Int. Conference on Pattern Recognition ICPR 2002, Quebec, Canada: IEEE Computer Society, 2002, 276-280.
    [49] Frigui H, Krishnapuram R. Clustering by competitive agglomeration [J]. Pattern Recognition, 1997, 30(7), 1109-1119.
    [50] Fritzke B. Some competitive learning methods [J/OL].[1997-12-20] http://www. neuroinfonnatik.ruhr-uni-bochum.de/ini/VDM/research/gsn/JavaPaper.
    [51] Stafylopatis A, Frossyniotis D, Likas A. A clustering method based on boosting [J]. Pattern Recognition Letters, 2004, 25(6): 641-654.
    [52] Sugeno M, Fukuyama Y. A new method of choosing the number of clusters for the fuzzy c-means method [C]. Proceedings of Fifth Fuzzy Systems Symposium, 1989, 247-250.
    [53] Floudas C A and Gounaris C E. A review of recent advances in global optimization [M]. Springer, Berlin, 2008.
    [54] Han E, Karypis G, Kumar V. Chameleon: Hierarchical clustering using dynamic modeling [J]. IEEE Computer, 1999, 32(8): 68-75.
    [55] Sokal R R, Gabriel K R. A new statistical approach to geographic variation analysis [J]. Syst.Zool., 1969, 18: 259-278.
    [56] Choudhary A, Goil S, Nagesh H. Mafia: Efficient and scalable subspace clustering for very large data sets [R]. cpdc-tr-9906-010, Center for Parallel and Distributed Computing, Department of Electrical & Computer Engineering, Northwestern University, 1999.
    [57] Toussaint G T. The relative neighborhood graph of a finite planar set [J]. Pattern Recognition, 1980, 12: 261-268.
    [58] Doring C, Timm H, Borgelt C, Kruse R. An extension to possibilistic fuzzy cluster analysis [J]. Fuzzy Sets Syst, 2004, 147: 3-16.
    [59] Kahng A B, Hagen L. New spectral met hods for ratio cut partitioning and clustering [J], IEEE Trans. Computer-Aided Design, 1992, 11(9), 1074-1085.
    [60] Halkidi M, Batistakis Y, Vazirgiannis M. Clustering validity checking methods: Part Ⅱ [J]. SIGMOD RECORD, 2002, 31(3), 19-27.
    [61] Kamber M, Han J. Data mining: Concepts and techniques [M]. San Francisco, US: Morgan Kaufmann, 2001.
    [62] Kamber M, Han J, Data mining: Concepts and techniques [M]. San Francisco: Morgan Kaufmann Publishers, 2001.
    [63] Hansen P, Ja'umard B. Cluster analysis and mathematical programming [J]. Math. Program., 1997, 79: 191-215.
    [64] Harary F. Graph theory [M]. Reading, MA: Addison-Wesley, 1969.
    [65] Bezdek J, Hathaway R, Tucker W. An improved convergence theory for the fuzzy isodata clustering algorithms [G]. Analysis of Fuzzy Information (Bezdek J C, ed.), Boca Raton: CRC Press, 1987, 3: 123-132.
    [66] He Q. A review of clustering algorithms as applied to ir, UIUCLIS-1999/6+IRG [R]. Univ. Illinois at Urbana-Champaign, 1999.
    [67] Keim D A, Hinneburg A. Optimal grid-clustring: Towards breaking the curse of dimensionality in high-dimensional clustering [C], Proceedings of the 25th VLDB Conference, 1999, 506-517.
    [68] Hoppner F, Klawonn F. A contribution to convergence theory of fuzzy c-means and derivatives [J]. IEEE Transactions on Fuzzy Systems, 2003, 11(5), 682-694.
    [69] Huang Z X. A fast clustering algorithm to cluster very large categorical data sets in data mining [C]. SIGMOD Workshop on Research Issues on Data Mining and Knowledge Discovery (SIGMOD-DMKD '97), 1997.
    [70] Jain A, Dubes R. Algorithms for clustering data [M]. Englewood Cliffs, NJ: Prentice-Hall, 1988.
    [71] Jain A K, Duin RPW, Mao JC. Statistical pattern recognition: A review [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2000, 22(1), 4-37.
    [72] Jain A K, Murty M N, Flynn P J. Data clustering: A review [J]. Acm Computing Surveys, 1999, 31(3), 264-323.
    [73] Ehrlich R, James, Bezdek C, Full W. Fcm: Fuzzy c-means algorithm [J]. Computers and Geoscience, 1984.
    [74] Bezdek J C. Cluster validity with fuzzy sets [J]. Cybernetics and Systems, 1974, 3: 58-74.
    [75] Bezdek J C. Numerical taxonomy with fuzzy sets [J]. Journal of Mathematical Biology, 1974, 1: 57-71.
    [76] Bezdek J C. Pattern recognition with fuzzy objective function algorithms [M]. Plenum Press, New York, 1981.
    [77] Bezdek J C. Partition structures: a tutorial [G], The Analysis of Fuzzy Information (Bezdek J C, ed.), CRC Press, Boca Raton, FL, 1987.
    [78] Blake C L, Merz C J. UCI repository of machine learning databases, a huge collection of artificial and real-world data sets, 1998. Available from: http://mlearn.ics.uci.edu/MLRepository.html.
    [79] Graham D B, Allinson N M. Characterizing virtual eigensignatures for general purpose face recognition [G], in: H. Wechsler, P.J. Phillips, V. Bruce, F. Fogelman-Soulie, T.S. Huang (Eds.), Face Recognition: From Theory to Applications, NATO ASI Series F, Computer and Systems Sciences, 1998, 163: 446-456.
    [80] Dunn J C. A fuzzy relative of the isodata process its use in detecting compact well-separated clusters [J]. Cybernetics and Systems, 1974, 3: 32-57.
    [81] Dunn J C. A graph theoretic analysis of pattern classification via tamura's fuzzy relation [J]. IEEE Trans. SMC, 1974, 4(3): 310-313.
    [82] Kleinberg J. An impossibility theorem for clustering [C]. Conf. Advances in Neural Information Processing Systems, 2002, 15: 463-470.
    [83] Rose K. Deterministic annealing, clustering and optimization [D], Ph.D. thesis, California Institute of Technology, 1991.
    [84] Kaufman L, Rousseeuw P. Finding groups in data: An introduction to cluster analysis [M]. Wiley, 1990.
    [85] Vecchi M P, Kirkpatrick S, Gelatt C D J. Optimization by simulated annealing [J]. Science, 1983, 220: 671-680.
    [86] Dubes R C, Klein R W. Experiments in projection and clustering by simulated annealing [J]. Pattern Recognition, 1989, 22(2): 213-220.
    [87] Kohonen T. The self-organizing map [C]. Proceedings of the IEEE 1990, 78(9): 1464- 1480.
    [88] Kohonen T. Self-organizing maps [M]. New York: Springer-Verlag, 2001.
    [89] Kolatch E. Clustering algorithms for spatial databases: A survey [J/OL].[2001-8-26]. http://citeseer.nj.nec.com/436 843.html.
    [90] Fukunaga K, Koontz W L G, Narendra P M. A branch and bound clustering algorithm [J]. IEEE Transactions on Computers, 1975, 24(9): 908-914.
    [91] Keller J M, Krishnapuram R. A possibilistic approach to clustering [J]. IEEE Trans. Fuzzy Syst., 1993, 1(2): 98-110.
    [92] Jain A K, Law M, Topchy A. Multiobjective data clustering [C]. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004, 424-430.
    [93] Zkim Le. Fuzzy relation compositions and pattern recognition [J]. Information Sciences, 1996, 89: 107-130.
    [94] Sun Z, Li C. A mean approximation approach to a class of.grid-based clustering algorithms [L]. Journal of Software, 2003, 14(7): 1267-1274.
    [95] Yu P S, Liu B, Xia Y. Clustering through decision tree construction [C]. Proceedings of the Ninth International Conference on Information and Knowledge Management, 2000: 20-29.
    [96] Fiedler M. Algebraic connectivity of graphs [J]. Mat h. J., 1973, 23: 298-305.
    [97] Rozonoer L, Aizerman M, Braverman E. Theoretical foundations of the potential function method in pattern recognition learning [J]. Automation and Remote Control, 1964, 25: 821-837.
    [98] MacQueen J. Some methods for classification and analysis of multivariate observations [J]. 5th Berkeley Symp., 1967(1): 281-297.
    [99] Bandyopadhyay, Maulik U. Genetic algorithm-based clustering technique [J]. Pattern Recognition, 2000, 33: 1455-1465.
    [100] McLachlan G, Krishnan T. The em algorithm and extensions [M]. New York: Wiley, 1997.
    [101] Xu L, Meila M. Multiway cut s and spect ral clustering [R]. U. Washington, 2003.
    [102] Mercer J. Functions of positive and negative type and their connection with the theory of integral equations [C]. Proceedings of the Royal Society of London, The Royal Society, 1909: 415-446.
    [103] Milligan G W, Cooper M C. An examination of procedures for determining the number of clusters in a data set [J]. Psychometrika, 1985, 50(2): 159-179.
    [104] Murtagh F. A survey of recent advances in hierarchical clustering algorithms [J]. Computer Journal, 1983, 26(4): 354-359.
    [105] Bezdek J, PalN, Tsao E. Generalized clustering networks and kohonen's self-organizing scheme [J]. IEEE Trans. Neural Netw., 1993(4): 549-557.
    [106] Weiss Y, Ng A Y, Jordan M I. On spectral clustering: Analysis and an algorithm [J]. Advances in Neural Information Processing Systems (Ghahramani T G, Dietterich S, Becker, ed.), Cambridge, MA, MIT Press, 2002, 14: 849-856.
    [107] Martinez R F, Osbourn G C. Empirically defined regions of influence for cluster analysis [J]. Pattern Recognition, 1995, 28(11): 1793-1806.
    [108] Keller J M, Pal N R, Pal K, Bezdek J C. A possibilistic fuzzy c-means clustering algorithm [J]. IEEE Transactions on Fuzzy Systems, 2005, 13(4).
    [109] Urquhart R. Graph theoretical clustering based on limited neighborhood sets [J]. Pattern Recognition, 1982, 15(3): 173-187.
    [110] Xu R, Wunsch D. Survey of clustering algorithms [J]. Transactions on Neural Networks, 2005, 16(3): 645-678.
    [111] Wu X H, Zhou J J. Possibilistic fuzzy c-means clustering model using kernel methods [C]. Computational Intelligence for Modelling, Control and Automation, International Conference on, 2005, 2: 465-470.
    [112] Hart P, Duda R, Stork D. Pattern classification, 2nd ed. [M]. New York: Wiley, 2001.
    [113] Tucker W, Hathaway R, Bezdek J. An improved covergence theorem for the fuzzy c-means clustering algorithms [J]. Analysis of Fuzzy Information,3.
    [114] Miyamoto S, Inokuchi R. Lvq clustering and som using a kernel function [C]. Proceedings of IEEE International Conference on Fuzzy Systems, 2004: 1497-1500.
    [115] Dave R N. Validating fuzzy partition obtained through c-shells clustering [J]. Pattern Recognition Letters, 1996, 17: 613-623.
    [116] Bandyopadhyay S. Simulated annealing using a reversible jump markov chain monte carlo algorithm for fuzzy clustering [J]. IEEE Trans. Knowledge Data Eng., 2005, 17(4): 479-790.
    [117] Saitoh S. Theory of reproducing kernels and its applications [M]. Harlow England: Longman Scientific & Technical, 1988.
    [118] Bertrand L S, Nozha B. Unsupervised robust clustering for image database categorization [C]. IEEE-IAPR International Conference on Pattern Recognition (ICPR ' 2002), August 2002.
    [119] Bittanti S, Gazzaniga G, Savaresi S, Boley D L. Cluster selection in divisive clustering algorithms [C]. Proceedings of the 2nd SLAM International Conference on Data Mining, 2002: 299-314.
    [120] Mika S, Scholkopf B. Input space versus feature space in kernel-based methods [J]. IEEE Transactions on Neural Networks, 1999, 10(5): 1000-1017.
    [121] Martin E, Schikuta E. The bang-clustering system: Grid-based data analysis [C]. Proceedings of the Second International Symposium on Advances in Intelligent Data Analysis, Reasoning about Data, 1997: 513-524.
    [122] Zhang A, Sheikholeslami G, Chatterjee S. Wavecluster: A multi-resolution clustering approach for very large spatial databases [C]. 24rd International Conference on Very Large Data Bases (VLDB'98), New York: Morgan Kaufmann Publishes, 1998: 428- 439.
    [123] Malik J, Shi J. Normalized cut s and image segmentation [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2000, 22(8): 888-905.
    [124] Michener C D, Sokal P R. A statistical method for evaluating systematic relationships [J]. University of Kansas Science Bulletin, 1958, 38: 1409-1438.
    [125] Ghosh J, Strehl A. Cluster ensembles-a knowledge reuse framework for combining multiple partitions [J]. Journal of Machine Learning Research, 2002, 3(3): 583-617.
    [126] Sergios T, Konstantinos K. Pattern recognition 3rd ed. [M]. Publishin House of Electronics Industry, china, 2006.
    [127] Sorensen T. A method of establishing groups of equal amplitude in plant sociology based on similarity of species content and its application to analyses of the vegetation on danish commons [J]. Biologiske Skrifter, 1948: 1-34.
    [128] Obermayer K, Graepel T. Fuzzy topographic kernel clustering [C]. Proceedings of the Fifth GI Workshop Fuzzy Neuro Systems'98 (Brauer W, ed.), 1998: 90-97.
    [129] Duin R P W, Tax D M J. Support vector domain description [J]. Pattern Recognition Letters, 1999, 20: 1191-1199.
    [130] Punch W F, Topchy A P, Jain A K. Combining multiple weak clusterings [C]. Proceedings of the 3rd 1EEE International Conference on Data Mining (ICDM 2003), Melbourne, Florida, USA: IEEE Computer Society, 2003: 331-338.
    [131] Punch W F, Topchy A P, Jain A K. A mixture model for clustering ensembles [C]. Proc . SLAM Conf. on Data Mining, 2004:379-390.
    [132] Punch W F, Topchy A P, Jain A K. Clustering ensembles: Models of consgnsus and weak partitions [J]. IEEE Trans. Pattern Anal.Mach. Intetll., 2005, 27(12): 1866-1881.
    [133] Yang S B, Tseng L Y. A genetic clustering algorithm for data with nonspherical-shape clusters [J]. Pattern Recognition, 2000, 33: 1251-1259.
    [134] Pedrycz W. Clustering and fuzzy clustering, Knowledge-Based Clustering: From Data to Information Granules (Pedrycz W., ed.) [M]. New York: John Wiley & Sons, 2005.
    [135] Qu Fu-heng, Ma Si-liang. Generalized possibilistic c-means clustering based on differential evolution algorithm [C]. 2009 International IEEE Workshop on Intelligent System and Application 2009, vol. 01 (accepted).
    [136] Storn R and Price K. Differential evolution-A simple and efficient heuristic for global optimization over continuous spaces [J]. J. Glob. Optim. 1997, 11(4): 4-359.
    [137] Price K, Storn R, Lampinen J. Differential evolution-a practical approach to global optimization [M]. Springer, Berlin, 2005.
    [138] Ulrike v L. A tutorial on spectral clustering [J]. Statistics and Computing, 2007, 17(4): 395-416.
    [139] Muntz R R, Wang W, Yang J. Sting: A statistical information grid approach to spatial data mining [C]. 23rd International Conference on Very Large Data Bases(VLDB), Athens , Greece: Morgan Kaufmann Publishers, 1997: 186-195.
    [140] Leahy R, Wu Z. An optimal graph theoretic approach to data clustering: theory and it s application to image segmentation [J]. IEEE Trans on PAMI, 1993, 15(11): 1101-1113.
    [141] Palaniappan K, Zhuang X, Huang Y, Zhao Y. Gaussian mixture density modeling, decomposition and applications [J]. IEEE Trans. Image Process., 1996, 5(9): 1293-1302.
    [142] Beni G, Xie X L. A validity measure for fuzzy clustering [J]. IEEE Trans. Pattern Anal. Mach. Intell., 1991, 13(8): 841-847.
    [143] Wu K L, Yang M S. Unsupervised possibilistic clustering [J]. Pattern Recognition, 2006, 39(1): 5-21.
    [144] Michalevitz Z. Genetic algorithms + data structures = evolutionary programming, 2nd ed [M]. Springer-Verlag, 1994.
    [145] Zangwill W. Non-linear programming: A unified approach [M]. Prentice-Hall,Englewood Cliffs, NJ, 1969.
    [146] Yu J P, Wu Z D, Xie W X. Fuzzy c-means clustering algorithm based on kernel method [J]. Comput. Intell. Multimedia Appl., 2003.
    [147] Yeung Y W, Zhang J S. Improved possibilistic c-means clustering algorithms [J]. IEEE Trans Fuzzy Syst, 2004, 12(2): 209-217.

© 2004-2018 中国地质图书馆版权所有 京ICP备05064691号 京公网安备11010802017129号

地址:北京市海淀区学院路29号 邮编:100083

电话:办公室:(+86 10)66554848;文献借阅、咨询服务、科技查新:66554700