用户名: 密码: 验证码:
半监督学习及其应用研究
详细信息    本馆镜像全文|  推荐本文 |  |   获取CNKI官网全文
摘要
本文研究了半监督学习算法及其在数据挖掘技术中的应用。首先,本文对机器学习、数据挖掘的相关知识进行了简单介绍,主要对它们的定义、发展历史和相关流程等进行了讨论。其次,介绍了本文所使用的半监督学习算法。最后,以信用卡数据和望远镜数据为例,进行了数据挖掘技术算法实现。
     在数据挖掘技术中,涉及到机器学习问题。半监督学习是在机器学习领域中同时利用训练样本的类别标记信息和未标记信息的学习方法。监督学习作为机器学习中的一种很主要的方法,基于类别标记已知的前提。在实际问题中,往往需要大的样本集,而提供大量已知类别的样本却存在一定的困难。非监督学习是机器学习中的另一种主要的方法,它不要求类别标记是已知的。但是,与监督学习相比,非监督学习存在着更大的不确定性。由于半监督学习同时利用训练样本的类别标记信息和未标记信息,所以它可以作为传统的监督学习和非监督学习的有益补充。我们所研究的半监督学习算法,就是基于这些基本原理。
     本文的主要创新点以及研究工作如下:
     (1)对半监督学习所涉及的理论基础及相关工作进行了研究,以便后文对算法进行相应的改进研究。目前在机器学习界,主要还是传统的监督学习和非监督学习两大类别,半监督学习还属于一个比较新颖的领域。对数据挖掘所涉及的相关工作领域进行了研究,以便后文对数据挖掘应用进行相应的分析研究。
     (2)给出了一个基于贝叶斯分类的半监督学习算法。该算法基于贝叶斯决策理论,通过概率密度函数进行分布估计,对两类别半监督学习问题进行判定。
     (3)给出了一个基于FCM的半监督学习算法。该算法来源于非监督学习的聚类方法,采用类别分离的间接方法来度量,并且加入了模糊模式识别方法,可以同时进行特征选择,对多类别半监督学习问题进行判定。
     (4)给出了在信用卡数据挖掘模型中,加入半监督学习算法作为技术解决方案的方法。该方案使用的算法就是基于FCM的半监督学习算法,同时可以进行特征选择。并且,因考虑到信用卡审批模型的特点,引入了损失函数,从而得到了一种新的半监督学习算法,来进行不同类别用户的判定。对天文数据分析进行了应用,给出了MAGIC望远镜数据的信息分析,这些数据将高能射线信号与背景区分开来。
The paper is about semi-supervised learning and its applications in data mining technology. The knowledge of machine learning, data mining, their histories, and processes are introduced. Also, we use semi-supervised machine learning algorithms. At last, we give out data mining applications, for useful information.
     Machine learning is used in data mining technology. Semi-supervised learning is exploiting both labeled and unlabeled data. Supervised learning assumes the class labels are known and its algorithms require sufficient training data so that the obtained model can generalize well. However, data in many domains are unlabeled; obtaining unlabeled data is rather easier. Unsupervised learning makes no assumptions about the category structure, which makes it more difficult for machine learning process. Semi-supervised learning can be beneficial to traditional supervised learning and unsupervised learning. The paper describes utilizing both labeled and unlabeled samples in building classifiers and clustering methods. The major innovations and research works are as followed:
     (1) The paper studies the technical bases and related works on semi-supervised learning. Compared with traditional supervised learning and unsupervised learning, semi-supervised learning is in a rather new field. Also, the paper has studied the related works on data mining.
     (2) The paper gives out a semi-supervised learning algorithm for Bayesian classification. The algorithm is based on Bayesian decision machinery, with estimation of the structure from probability density function, for two-class problems.
     (3) The paper gives out a semi-supervised learning algorithm based on FCM. The algorithm uses the idea of unsupervised clustering, indirected measurement of separating clusters, fuzzy method, and feature selection, for multiple-class problems.
     (4) The paper gives out semi-supervised learning algorithm as technical solution for data mining model in credit card application. It uses semi-supervised learning algorithm based on FCM, and for feature selection. Concerned with the characteristics of credit card approval model, cost function is applied, for recognizing different users. Also, the paper gives out Magic telescope data analysis application, for separating high-energy ray signal with background.
引文
[1]边肇祺,张学工.模式识别[M].第二版.北京:清华大学出版社, 2000. 2-31
    [2]孙即祥:现代模式识别[M].长沙:国防科技大学出版社, 2002. 2-10
    [3]杨光正,吴岷,张晓莉:模式识别[M].合肥:中国科学技术大学出版社, 2001. 3-9
    [4]黄同岗,宋克欧.模式识别[M].哈尔滨:哈尔滨工程大学出版社, 1998. 2-15
    [5] Marquess de Sa J P著,吴逸飞译.模式识别——原理、方法及应用[M].北京:清华大学出版社, 2002. 3-21
    [6]孙即祥.模式识别中的特征提取与计算机视觉不变量[M].北京:国防工业出版社, 2001. 3-15
    [7]郭桂荣,庄钊文.信息处理中的模糊技术[M].长沙:国防科技大学出版社, 1993. 2-17
    [8]郭桂蓉.模糊模式识别[M].长沙:国防科技大学出版社, 1992. 2-10
    [9]王士同.模糊系统、模糊神经网络及应用程序设计[M].上海:上海科学技术文献出版社, 1998. 3-25
    [10]殷勤业,杨宗凯,谈正.模式识别与神经网络[M].北京:机械工业出版社, 1992. 4-10
    [11] Gudici P. Applied Data Mining: Statical Methods for Business and Industry [M]. John Wiley & Sons, 2003. 232-255
    [12]闪四清,陈茵,程燕.数据挖掘——概念、模型、方法和算法[M].北京:清华大学出版社, 2002. 5-84
    [13] Han J W, Kamber M. Data Mining: Concepts and Techniques [M]. Morgan Kaufmann, 2001. 5-192
    [14]朱建平.数据挖掘的统计方法及实践[M].北京:中国统计出版社, 2005. 12-17
    [15] Fayyad U M, Piatetsky-Shapiro G, Smyth P, et al. Advances in Knowledge Discovery and Data Mining [M], Cambridge, MA: MIT Prress, 1996. 5-104
    [16]陈安,陈宁,周龙骧.数据挖掘技术及应用[M].北京:科学出版社, 2006. 1-48.
    [17] Michalski R S, Bratko I, Kubat M著,朱明译.机器学习与数据挖掘:方法和应用[M].北京:电子工业出版社, 2004. 5-20
    [18] Mitchell T M著,曾华军译.机器学习[M].北京:机械工业出版社, 2003. 2-19
    [19]杨善林,倪志伟.机器学习与智能决策支持系统[M].北京:科学出版社, 2004. 5-40
    [20] Rosset S, Zhu J, Zou H, et al. A method for inferring label sampling mechanisms in semi-supervised learning [C]. In: Saul L K, Weiss Y, Bottou L, eds. Advances in Neural Information Processing Systems 17. Cambridge, MA: MIT Press, 2005. 1161-1168
    [21] Liu B, Lee W S, Yu P S, et al. Partially supervised classification of text documents [C]. In: Sammut C, Hoffmann A G, eds. Proceedings of the Nineteenth International Conference on Machine Learning. San Francisco, CA: Morgan Kaufmann, 2002. 387-394
    [22] Lu Q, Getoor L. Link-based classification using labeled and unlabeled data [C]. In: Fawcett T, Mishra N, eds. Proceedings of the Twentieth International Conference on Machine Learning. Washington, DC: AAAI Press, 2003. 496-503
    [23] Lawrence N D, Jordan M I. Semi-supervised learning via Gaussian processes [C]. In: Saul L K, Weiss Y, Bottou L, eds. Advances in Neural Information Processing Systems 17. Cambridge, MA: MIT Press, 2005. 753-760
    [24] Inoue M, Ueda N. HMMs for Both Labeled and Unlabeled Time Series Data [C]. In: Miller D, Adali T, Larsen J, et al, eds. Neural Networks for Signal Processing XI, Proceedings of the 2001 IEEE Signal Processing Society Workshop. New York, NY: IEEE, 2001. 93-102
    [25] Denis F, Gilleron R, Tommasi, M. Text classification from positive and unlabeled examples [C]. In: The 9th International Conference on Information Processing and Management of Uncertainty in Knowledge-Based Systems. Annecy, France, 2002. 1927-1934
    [26] Seeger M. Learning with labeled and unlabeled data [R]. Technical Report, Institute for Adaptive and Neural Computation, University of Edinburgh, 2001. 1-7
    [27] Cozman F G, Cohen I. Unlabeled data can degrade classitication performance of generative classifiers [R]. Hewlett Packard Technical Report, HPL-2001-234, 2001.1-16
    [28] Shashahani B, Landgrebe D. The effect of unlabeled samples in reducing the small sample size problem and mitigating the hughes phenomenon [J]. IEEE Transactions on Geoscience and Remote Sensing, 1994, 32(5): 1087-1095
    [29] Elworthy D. Does Baum-Welch re-estimation help taggers? [C]. In: Proceedings of the Fourth Conference on Applied Natural Language Processing. Morristown, NJ, USA: ACL, 1994. 53-58
    [30] Cozman F, Cohen I, Cirelo M. Semi-supervised Learning of Mixture Models [C]. In: Fawcett T, Mishra N, eds. Proceedings of the Twentieth International Conference on Machine Learning. Washington, DC: AAAI Press, 2003. 99-106
    [31] McCallum A K, Nigam K. Employing EM in pool-based active learning for text classification [C]. In: Shavlik J W, ed. Proceedings of the Fifteenth International Conference on Machine Learning. Madison, US: Morgan Kaufmann Publishers, 1998. 350-358
    [32] Muslea I, Minton S, Knoblock C. Active + semi-supervised learning = robust multi-view learning [C]. In: Sammut C, Hoffmann A G, eds. Proceedings of the Nineteenth International Conference on Machine Learning. San Francisco, CA: Morgan Kaufmann, 2002. 435-442
    [33] Zhou Z-H, Chen K-J, Jiang Y. Exploiting unlabeled data in content-based image retrieval [C]. In: Boulicaut J-F, Esposito F, Giannotti F, et al, eds. Proceedings of ECML 2004, 15th European Conference on Machine Learning. Lecture Notes in Computer Science 3201. Berlin Heidelberg: Springer, 2004. 525-536
    [34] Zhu X, Ghahramani Z, Lafferty J. Semi-supervised learning using Gaussian fields and harmonic functions [C]. In: Fawcett T, Mishra N, eds. Proceedings of the Twentieth International Conference on Machine Learning. Washington, DC: AAAI Press, 2003. 912-919
    [35] Zhu X, Lafferty J, Ghahramani Z. Combining active learning and semi-supervised learning using Gaussian fields and harmonic functions [C]. In: Fawcett T, Mishra N, eds. Proceedings of the Twentieth International Conference on Machine Learning.Washington, DC: AAAI Press, 2003. 58-65
    [36] Dempster A P, Laird N M, Rubin D B. Maximum likelihood from incomplete data via the EM algorithm [J]. Journal of the Royal Statistical Society, Series B (Methodological), 1977, 39(1): 1-38
    [37] Bensaid A M, Hall L O, Bezdek J C, et al. Partially supervised clustering for image segmentation [J]. Pattern Recognition, 1996, 29(5): 859-871
    [38] Miller D J, Uyar H S. A mixture of experts classifier with learning based on both labelled and unlabelled data [C]. In: Mozer M, Jordan M I, Petsche T, eds. Advances in Neural Information Processing Systems 9. Cambridge, MA: MIT Press 1997. 571-577
    [39] Chapelle O, Scholkopf B, Zien A. Semi-supervised Learning [M]. Cambridge, MA: MIT Press, 2006. 2-10
    [40] Jaakkola T, Haussler D. Exploiting generative models in discriminative classifiers [C]. In: Kearns M J, Solla S A, Cohn D A, eds. Advances in Neural Information Processing Systems 11. Cambridge, MA: The MIT Press, 1999. 487-489
    [41] Nigam K, McCallum A K, Thrun S, et al. Text classification from labeled and unlabeled documents using EM [J]. Machine Learning, 2000, 39(2-3): 103-134
    [42] Baluja S. Probabilistic modeling for face orientation discrimination: learning from labeled and unlabeled data [C]. In: Kearns M S, Solla S A, Cohn D A, eds. Advances in neural information processing systems 11. Cambridge, MA: MIT Press, 1999. 854-860
    [43] Fujino A, Ueda N, Saito K. A hybrid generative/discriminative approach to semi-supervised classifier design [C]. In: Veloso M M, Kambhampati S, eds. Proceedings of the Twentieth National Conference on Artificial Intelligence and the Seventeenth Innovative Applications of Artificial Intelligence Conference. Menlo Park, California: AAAI Press / The MIT Press, 2005. 764-769
    [44] Ratsaby J, Venkatesh S. Learning from a mixture of labeled and unlabeled examples with parametric side information [C]. In: Proceedings of the Eighth Annual Conference on Computational Learning Theory. New York, NY: ACM Press, 1995. 412-417
    [45] Corduneanu A, Jaakkola T. Stable Mixing of Complete and Incomplete Information [R]. AI Laboratory Memo 2001-30, Massachusetts Institute of Technology, 2001, 1-9
    [46] Castelli V, Cover T. On the exponential value of labeled samples [J]. Pattern Recognition Letters, 1995, 16(1): 105-111
    [47] Castelli V, Cover T. The relative value of labeled and unlabeled samples in pattern recognition with an unknown mixing parameter [J]. IEEE Transactions on Information Theory, 1996, 42(6): 2102-2117
    [48] Nigam K. Using unlabeled data to improve text classification [R]. Technical Report CMU-CS-01-126. Carnegie Mellon University. Doctoral Dissertation, 2001
    [49] Miller D J, Browning J. A Mixture Model and EM-Based Algorithm for Class Discovery, Robust Classification, and Outlier Rejection in Mixed Labeled/Unlabeled Data Sets [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2003, 25(11): 1468-1483
    [50] Yarowsky, D. Unsupervised word sense disambiguation rivaling supervised methods [C]. In: Proceedings of the 33rd Annual Meeting of the Association for Computational Linguistics. Massachusetts, USA: Morgan Kaufmann Publishers, 1995. 189-196
    [51] Riloff E, Wiebe J, Wilson T. Learning subjective nouns using extraction pattern bootstrapping [C]. In: Proceedings of the Seventh Conference on Natural Language Learning. ACL, 2003. 25-32
    [52] Maeireizo B, Litman D, Hwa R. Co-training for predicting emotions with spoken dialogue data [C]. In: The Companion Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics. ACL, 2004. 203-206
    [53] Rosenberg C, Hebert M, Schneiderman H. Semi-supervised selftraining of object detection models [C]. In: Proceedings of the Seventh IEEE Workshop on Applications of Computer Vision. Breckenridge, USA, 2005. 29-36
    [54] Blum A, Mitchell T. Combining labeled and unlabeled data with co-training [C]. In: Bartlett P, Mansour Y, eds. Proceedings of the Eleventh Annual Conference on Computational Learning Theory. New York, NY: ACM Press, 1998. 92-100
    [55] Mitchell T. The role of unlabeled data in supervised learning [C]. In: Proceedings of the Sixth International Colloquium on Cognitive Science. San Sebastian, Spain, 1999. 1-8
    [56] Nigam, K, Ghani R. Analyzing the effectiveness and applicability of co-training [C]. In:Proceedings of the 2000 ACM CIKM International Conference on Information and Knowledge Management, McLean, VA, USA. New York, NY: ACM Press, 2000. 86-93
    [57] Jones R. Learning to extract entities from labeled and unlabeled text [R]. Technical Report CMU-LTI-05-191. Carnegie Mellon University. Doctoral Dissertation, 2005
    [58] Goldman S, Zhou Y. Enhancing supervised learning with unlabeled data [C]. In: Langley P, ed. Proceedings of the Seventeenth International Conference on Machine Learning. Stanford University, Standord, CA, USA: Morgan Kaufmann, 2000. 327-334
    [59] Zhou Y, Goldman S. Democratic co-learing [C]. In: Proceedings of the 16th IEEE International Conference on Tools with Artificial Intelligence. Boca Raton, FL, USA, IEEE Computer Society, 2004. 594-602
    [60] Zhou Z-H, Li M. Tri-Training: Exploiting Unlabeled Data Using Three Classifiers [J]. IEEE Transactions on Knowledge and Data Engineering, 2005, 17(11): 1529-1541
    [61] Balcan M F, Blum A, Yang K. Co-training and expansion: Towards bridging theory and practice [C]. In: Saul L K, Weiss Y, Bottou L, eds. Advances in Neural Information Processing Systems 17. Cambridge, MA: MIT Press, 2005. 89-96
    [62] Leskes, B. The value of agreement - a new boosting algorithm [C]. In: Auer P, Meir R, eds. Proceedings of The Eighteenth Annual Conference on Learning Theory. Lecture Notes in Computer Science 3559, Berlin Heidelberg: Springer, 2005. 95-110
    [63] Joachims T. Transductive inference for text classification using support vector machines [C]. In: Bratko I, Dzeroski S, eds. Proceedings of the Sixteenth International Conference on Machine Learning. San Francisco, CA: Morgan Kaufmann, 1999. 200-209
    [64] Blum A, Chawla S. Learning from labeled and unlabeled data using graph minicuts [C]. In: Brodley C E, Danyluk A P, eds. Proceedings of the Eighteenth International Conference on Machine Learning. San Francisco, CA: Morgan Kaufmann, 2001. 19-26
    [65] Gabrys B, Petrakieva L. Combining labelled and unlabelled data in the design of pattern classification systems [J]. International Journal of Approximate Reasoning, 2004, 35(3): 251-273
    [66] Cohen I, Cozman F G, Sebe N, et al. Semisupervised learning of classifiers: theory, algorithms, and their application to human-computer interaction [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2004, 26(12): 1553-1567
    [67] Klose A. Extracting Fuzzy Classification Rules from Partially Labeled Data [J]. Soft Computing, 2004, 8(6): 417-427
    [68] Klose A, Kruse R. Semi-supervised Learning in Knowledge Discovery [J]. Fuzzy Sets and Systems, 2005, 149(11): 209-233
    [69] Amini M R, Gallinari P. Semi-supervised learning with an imperfect supervisor [J]. Knowledge and Information Systems, 2005, 8(4): 385-413
    [70] Rigutini L, Maggini M. A Semi-supervised Document Clustering Algorithm based on EM [C]. In: Proceedings of the 2005 IEEE/WIC/ACM International Conference on Web Intelligence. Washington, DC, USA: IEEE Computer Society, 2005: 200-206
    [71] Vatsavai, R R, Shekhar S, Burk, T E. A Semi-supervised Learning Method for Remote Sensing Data Mining [C]. In: Proceedings of the 17th IEEE International Conference on Tools with Artificial Intelligence. Washington, DC, USA: IEEE Computer Society, 2005. 207-211
    [72] Gosselin P H, Cord M. Feature-based Approach to Semi-supervised Similarity Learning [J]. Pattern Recognition, 2006, 39(10): 1839-1851
    [73] Grira N, Crucianu M, Boujemaa N. Unsupervised and semisupervised clustering: a brief survey [R]. In: A Review of Machine Learning Techniques for Processing Multimedia Content, Report of the MUSCLE European Network of Excellence, 2004
    [74] Demiriz A, Bennett K, Embrechts M. Semi-supervised clustering using genetic algorithms [C]. In: Proceedings of Artificial Neural Networks in Engineering, ASME Press, 1999. 809-814
    [75] Zeng H-J, Wang X-H, Chen, Z, et al. CBC: Clustering based text classification requiring minimal labeled data [C]. In: Proceedings of the Third IEEE International Conference on Data Mining. Washington, DC, USA: IEEE Computer Society, 2003. 443-450
    [76] Pedrycz W, Vukovich G.. Fuzzy clustering with supervision [J]. Pattern Recognition,2004, 37(7): 1339-1349
    [77] Bouchachia A, Pedrycz W. Data clustering with partial supervision [J]. Data Mining and Knowledge Discovery, 2006, 12(1): 47-78
    [78] Bouchachia A, Pedrycz W. Enhancement of fuzzy clustering by mechanisms of partial supervision [J]. Fuzzy Sets and Systems, 2006, 157(13): 1733-1759
    [79] Zhou Z-H, Li M. Semi-supervised regression with co-training [C]. In: Kaelbling L P, Saffiotti A, eds. Proceedings of the Nineteenth International Joint Conference on Artificial Intelligence. 2005. 908-916
    [80] Brefeld U, Gaertner T, Scheffer T, et al. Efficient co-regularized least squares regression [C]. In: Cohen W, Moore A, eds. Proceedings of the 23rd International Conference on Machine Learning. New York, NY: ACM Press, 2006. 137-144
    [81]宫秀军,史忠植.基于Bayes潜在语义模型的半监督Web挖掘[J].软件学报, 2002, 13(8): 1508-1514
    [82]孙广玲,唐降龙.基于分层高斯混合模型的半监督学习算法[J].计算机研究与发展, 2004, 41(1): 156-161
    [83]徐敏,张丽萍,朱梧.基于ART半监督在线学习的文档分类[J].西南交通大学学报, 2006, 41(3): 335-340
    [84]周艺华,曹元大,魏本杰等.图像检索中基于记忆与半监督的主动相关反馈算法[J].北京理工大学学报, 2006, 26(1): 45-48
    [85]俞研,黄皓.一种半聚类的异常入侵检测算法[J].计算机应用, 2006, 26(7): 1640-1642
    [86]鲁珂,赵继东,叶娅兰等.一种用于图像检索的新型半监督学习算法[J].电子科技大学学报, 2005, 34(5): 669-671
    [87]郭崇慧,田凤占,勒晓明.数据挖掘教程[M].北京:清华大学出版社, 2005. 5-8.
    [88]滕广青,毛英爽.国外数据挖掘应用研究与发展分析[J].统计研究, 2005, 22(12): 68-70
    [89]王学丽,李嘉森.我国近年数据挖掘研究分析[J].中国统计, 2008, 2008(11): 51-52
    [90]焦李成.智能数据挖掘与知识发现[M].西安:西安电子科技大学出版社, 2006.4-21
    [91]马超群,兰秋军,陈为民.金融数据挖掘[M].北京:科学出版社, 2007. 1-56
    [92] Lee H. Justifying database normalization: a cost benefit model [J]. Information Processing & Management, 1995, 31(1): 59-67
    [93] Chae Y M, Ho S H, Cho K W, et al. Data mining approach to policy analysis in a health insurance domain [J]. International Journal of Medical Informatics, 2001, 62(2-3): 103-111
    [94] Knox S. Loyalty-Based Segmentation and the Customer Development Process [J]. European Management Journal, 1998, 16(6): 729-737
    [95] Sung H H, Sang C P. Application of data mining tools to hotel data mart on the Internet for database marketing [J]. Expert Systems With Application, 1998, 15(1): 1-31
    [96] Mani D R, Drew J, Betz A, et al. Statistics and data mining techniques for life time value modeling [C]. In: Proceedings of the ACM 5th International Conference on Knowledge Discovery and Data Mining. New York, NY, USA: ACM, 1999. 94-113
    [97] Kitts B, Freed D, Vrieze M. Cross-sell: a fast promotion-tunable customer-item recommendation method based on conditionally independent probabilities [C]. In: Simoff S J, Zaiane O R, eds. Proceedings of the sixth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. New York, NY, USA: ACM, 2000. 437-446
    [98] Berson A, Smith S, Thearling K.构建面向CRM的数据挖掘应用[M].北京:人民邮电出版社, 2001. 3-28
    [99] Berry M J A, Linoff G.著,吴旭智,赖淑贞译.资料采矿理论与实务——顾客关系管理的技巧与科学[M].台北:维科图书有限公司, 2001. 2-10
    [100]陈孟君.顾客关系管理深度解析——第二辑[M].远擎管理顾问公司,台北:远擎出版社, 2001. 47-55
    [101] Swift R S著,赖士奇,吴嘉哲,刘扬恺等译.深化顾客关系管理[M].远擎管理顾问公司,台北:远擎出版社, 2001. 24-37
    [102]杨红.数据挖掘技术在商业银行客户管理关系中的应用[J].特区经济, 2005 (5): 373-374
    [103] Duda R O, Hart P E, Stork D G.. Pattern classification (2nd ed.) [M]. New York, NY: John Wiley & Sons, 2000. 5-124
    [104] Quinlan R. Induction of decision trees [J]. Machine Learning, 1986, 1(1): 81-106
    [105] Quinlan R. C4.5: Programs for machine learning [M]. San Mateo, CA: Morgan Kaufmann Publishers, 1993. 11-50
    [106] Duda R O, Hart P E. Pattern classification and scene analysis [M], New York: John Wiley & Sons, 1973. 4-72
    [107] John G H, Langley P. Estimating continuous distributions in bayesian classifiers [C]. In: Besnard P, Hanks S, eds. Proceedings of the Eleventh Annual Conference on Uncertainty in Artificial Intelligence. San Mateo, CA: Morgan Kaufmann, 1995, 338-345
    [108] Heckerman D. Bayesian Networks for Data Mining [J]. Data Mining and Knowledge Discovery, 1997, 1(1): 79-119
    [109] Bishop C M. Neural networks for pattern recognition [M]. Oxford, England: Oxford University Press, 1995. 2-52
    [110] Rumelhart D E, Hinton G E, Williams R J. Learning internal representations by error propagation [C]. In: Parallel distributed processing: Explorations in the microstructure of cognition, Vol. 1: Foundations. Cambridge, MA: MIT Press, 1986, 318-362
    [111] Cover T M, Hart P E. Nearest neighbor pattern classification [J]. IEEE Transactions on Information Theory, 1967, 13(1): 21-27
    [112] Cortes C, Vapnik V. Support-vector networks [J]. Machine Learning, 1995, 20(3): 273-297
    [113] Vapnik V. Statistical learning theory [M]. Wiley, 1998. 2-150
    [114] Platt J. Fast training of support vector machines using sequential minimal optimization [C]. In: Advances in Kernel Methods - Support Vector Learning. Cambridge, MA: MIT Press, 1999, 185-208
    [115] Dietterich T. G.: Ensemble methods in machine learning [C]. In: Proceedings of the First International Workshop on Multiple Classifier Systems. New York, NY: SpringerVerlag, 2000, 1-15
    [116] Devroye L, Gyorfi L, Lugosi G. A probabilistic theory of pattern recognition [M]. New York: Springer Verlag, 1996. 2-40
    [117] Izenman, A J. Recent developments in nonparametric density estimation [J]. Journal of the American Statistical Association, 1991, 86(413): 205-224
    [118] Parzen, E. On estimation of a probability density function and model [J]. Annals of Math Statistics, 1962, 33(3): 1065-1076
    [119] Figueiredo M A T. Adaptive Sparseness for Supervised Learning [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2003, 25(9): 1150-1159
    [120] Krishnapuram B, Hartemink A J, Carin L, e al. A Bayesian Approach to Joint Feature Selection and Classifier Design [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2004, 26(9): 1105-1111
    [121] Law M H C, Figueiredo M A T, Jain A K. Simultaneous Feature Selection and Clustering Using Mixture Models [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2004, 26(9), 1154-1166
    [122] Girolami M, He C. Probability density estimation from optimally condensed data samples [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2003, 25(10): 1253-1264
    [123] Heckerman D, Geiger, D, Chickering, D M. Learning bayesian networks: the combination of knowledge and statistical data [J]. Machine Learning, 1995, 20(3): 197-243
    [124] Cheeseman P, Kelly J, Self M, et al. AutoClass: a bayesian classification system [C]. In: Laird J, ed. Proceedings of the Fifth International Conference on Machine Learning. Michigan: Morgan Kaufmann, 1988. 54-64
    [125] Rabiner L R. A tutorial on Hidden Markov Models and selected applications in speech recognition [J]. Proceedings of the IEEE, 1989, 77(2): 257-286
    [126] McLachlan G, Krishnan T. The EM algorithm and extensions [M]. New York: John Wiley & Sons, 1997. 2-100
    [127] Ripley, B D. Pattern Recognition and Neural Networks [M]. Cambridge:Cambridge University Press, 1996. 4-201
    [128] Kaufman L, Rousseeuw P J. Finding Groups in Data: An Introduction to Cluster Analysis [M]. New York: John Wiley & Sons, 1990. 4-71
    [129] Bezdek J C. Pattern recognition with fuzzy objective function algorithms [M]. New York: Plenum, 1981. 2-268
    [130] Pedrycz W. Algorithms of fuzzy clustering with partial supervision [J]. Pattern Recognition Letters, 1985, 3(1): 13-20
    [131] Pedrycz W; Waletzky J. Fuzzy clustering with partial supervision [J]. IEEE Transactions on Systems, Man and Cybernetics - Part B: Cybernetics, 1997, 27(5): 787-795
    [132] Bouchachia A, Pedrycz W. A semi-supervised clustering algorithm for data exploration [C]. In: Bilgic T, ed. Proceedings of 10th International Fuzzy Systems Association World Congress. Berlin Heidelberg: Springer, 2003. 328-337
    [133] Yang M-S. On a class of fuzzy classification maximum likelihood procedures [J]. Fuzzy Sets and Systems, 1993, 57(3): 365-357
    [134] Yang M-S, Su C-F. On parameter estimation for normal mixtures based on fuzzy clustering algorithms [J]. Fuzzy Sets and Systems, 1994, 68(1): 13-28
    [135] Yang M-S, Wu K-L. A Similarity-based robust clustering method [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2004, 26(4): 434-448
    [136] Yang M-S, Yu N-Y. Estimation of parameters in latent class models using fuzzy clustering algorithms [J]. European Journal of Operational Research, 2005, 160(2): 515-531
    [137] Yeung D S, Wang X Z. Improving performance of similarity-based clustering by feature weight learning [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2002, 24(4): 556-561
    [138] Wolf L, Shashua A. Feature selection for unsupervised and supervised inference: the emergence of sparsity in a weighted-based approach [C]. In: Proceedings of the Ninth IEEE International Conference on Computer Vision. Nice, France, IEEE Computer Society, 2003. 378-384
    [139] Huang J, Ng M, Rong H, et al. Automated variable weighting in k-means type clustering [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2005, 27(5): 1-12
    [140] Blum A, Langley P. Selection of relevant features and examples in machine learning [J]. Artificial Intelligence, 1997, 97(1-2): 245-271
    [141] Jain A, Zongker D. Feature selection: evaluation, application, and small sample performance [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1997, 19(2): 153-157
    [142] Kohavi R, John G.. Wrappers for feature subset selection [J]. Artificial Intelligence, 1997, 97(1-2): 273-324
    [143] Koller D, Sahami M. Toward optimal feature selection [C]. In: Saitta L, ed. Proceedings of 13th International Conference on Machine Learning, Morgan Kaufmann Publishers, 1996. 284-292

© 2004-2018 中国地质图书馆版权所有 京ICP备05064691号 京公网安备11010802017129号

地址:北京市海淀区学院路29号 邮编:100083

电话:办公室:(+86 10)66554848;文献借阅、咨询服务、科技查新:66554700