模糊分类新方法及应用
详细信息    本馆镜像全文|  推荐本文 |  |   获取CNKI官网全文
摘要
模糊集理论是一种重要的处理不精确,不确定信息的智能计算方法.它扩展了过去经典集合论中对精确数字和精确关系的要求,将模糊性和不明确性引入到精确的数据关系中,从而为传统的机器学习算法(例如,粗糙集,神经网络,聚类算法等)研究和处理的更多的信息提供了更加灵活,宽泛的条件.模糊粗糙集(与粗糙集相结合)和模糊神经网络(与神经网络相结合)就是两种颇受关注的基于模糊理论的算法.
     粗糙集理论也是一种处理不完备,不完整数据的数学工具.与模糊集理论相比,它更强调数据间的不可分辨性.但是由于定义的原因,使得粗糙集只能处理离散的数据集.通过使用模糊等价关系和模糊逻辑算子代替精确等价关系和经典逻辑算子,模糊粗糙集可以有效的分析离散数据或者连续数据(或是二者的混合数据)之间的关系.在本文中,从算法发展和运算机制的两个角度研究了两种基于模糊粗糙集理论的近邻分类算法:模糊粗糙近邻算法和不确定量化近邻算法.从算法上,文中提出了一种基于核函数的模糊粗糙集,并得到相应的两种基于核函数的模糊粗糙近邻算法.实验给出了该算法在乳癌风险评估方面的表现.与一些常用分类算法相比,基于核函数的模糊近邻算法都得到了更好更稳定的结果.另一方面,从理论上,文中分析了模糊粗糙近邻算法和不明确量化近邻算法的运算机制,指出他们的分类结果分别是由与待分类样本有最大相似度和与待分类样本的最大的相似度和决定.根据这一结论提出了两种分别与模糊粗糙近邻算法和不明确量化近邻算法等价的基于模糊相似关系的近邻算法.进一步的,文中提出了一种高度灵活的通用的基于模糊相似关系的近邻分类结构.并从理论上和实验上都证明了,这种分类结构的两种特例分别与模糊粗糙近邻算法和不明确量化近邻算法在分类结果上等价.
     模糊神经网络作为模糊系统和神经网络相结合的一种人工智能系统,主要有两种形式,一种是将模糊理论引入神经网络,实现结构上的模糊化.这样一来神经网络处理信息的范围和能力就得到了拓宽和提高.另一方面,使用神经网络的结构处理模糊信息.利用神经网络的训练方法,模糊规则和模糊化方法可以从结构上和方法上进行自动提取和优化.本文中采用进化极端学习机算法训练零阶TSK模糊推理系统,得到了一种进化模糊极端学习机算法.并同样使用乳癌诊断数据集对算法进行评估.这一算法也取得了比较好的实验结果.另外对一种具有模糊化性质的局部耦合前向神经网络进行了研究.一种改进的梯度学习算法被用来训练这一网络以提高学习效率,同时给出了使用这种改进的梯度学习算法的误差函数的单调性集算法的弱收敛和强收敛性定理.
Fuzzy set theory (FST) plays an important role in dealing with imprecise, and uncertain information. It relaxes the precise number and exact relation constraints of classical set theory, by introducing fuzziness to the relation between the data. Fuzzy sets provide a wider and more flexible framework for dealing with data than traditional machine learning methods (e.g. neural networks, clustering methods, etc). FST is a mature research area, and has also been widely applied in the areas of mechanical control, pattern recognition, and decision support systems. Fuzzy-rough sets (a hybridisation of rough and fuzzy sets) and fuzzy neural networks (a hybridi-sation of neural networks and fuzzy sets) have enjoyed much attention as two fuzzy set theory based approaches.
     Rough set theory was also proposed as a mathematical tool for dealing with imperfect and incomplete knowledge. Compared with FST, rough set theory is more concerned with a dif-ferent type of uncertainty:indiscernibility. However,due to its definition, rough sets can only operate effectively with datasets containing discrete values. By employing a fuzzy equivalence relation and fuzzy logical operators instead of a crisp equivalence relation and classical logical operators,respectively, fuzzy-rough sets provide a means by which the relationship between dis-crete data or real-valued data (or a mixture of both) can be effectively analysed. In this thesis, fuzzy-rough nearest-neighbour classification algorithms are studied from both methodological and theoretical perspectives. From theoretical development, a kernel-based fuzzy-rough set tech-nique and associated nearest-neighbour algorithms are proposed. Real-world medical datasets for the task of mammographic risk assessment are employed in order to evaluate such classi-fication approaches. The experimental results demonstrate that such kernel-based fuzzy-rough nearest-neighbour approaches offer improved and more robust performances over other classi-fiers. Theoretically, the underlying mechanism of fuzzy-rough nearest-neighbour (FRNN) and vaguely quantified nearest-neighbour (VQNN) algorithms are explored. The research shows that the resulting classification of FRNN and VQNN depends only upon the highest similarity and greatest summation of the similarities of each class, respectively. This fact is exploited in or-der to formulate two novel fuzzy similarity-based parallel methods. Furthermore, a generalised fuzzy similarity-based nearest-neighbour framework is presented. The theoretical proof and em-pirical evaluation demonstrate that FRNN and VQNN can be considered as the special cases of the proposed new framework.
     As the combination of fuzzy systems with neural networks, fuzzy neural networks have two main categories:1) The fuzzified structure of neural networks are implemented via introducing the fuzzy sets to neural networks. In so doing, the range and the ability of processing information for neural networks can be widened and improved.2) The fuzzy information is handled under the framework of neural networks. By using the training algorithms for neural networks, the fuzzy rules and the fuzzification approaches can be automatically extracted and optimised from both constructive and methodological perspectives. In this thesis, the hybridisation of the zero-order TSK fuzzy system with the evolutionary extreme learning machine approach leads an evolutionary fuzzy extreme learning machine. This technique is also applied to the task of mammographic risk analysis. The experimental results demonstrate that the evolutionary fuzzy extreme learning machine offers improved classification accuracy, both at the overall image level and at the level of individual risk types. Also, a local coupled feed-forward neural network is also used as the property for fuzzification for this work. In order to enhance the lcarning efficiency, a modified gradient-based learning method is employed to train such neural networks. The monotonicity of the error functions and the weak and strong convergence results for this algorithm are also proven.
引文
[1]PAWLAK Z. Rough Sets:Theoretical Aspects of Reasoning About Data[M]. Norwell, MA:Kluwer,1991.
    [2]ZADEH L A. Fuzzy sets[J]. Information and Control,1965,8:338-353.
    [3]DUBOIS D, PRADE H. Rough fuzzy sets and fuzzy rough sets[J]. International Journal of General Systems, 1990,17(2-3):191-209.
    [4]DAs M, CHAKRABORTY M K, GHOSHAL T K. Fuzzy tolerance relation, fuzzy tolerance space and basis[J]. Fuzzy Sets and Systems,1998,97:361-369.
    [5]YAO Y Y. A comparative study of fuzzy sets and rough sets[J]. Information Sciences,1998,109(l-4):227-242.
    [6]YEUNG D S, CHEN D, TSANG E C C, et al. On the generalization of fuzzy rough sets[J]. IEEE Transactions on Fuzzy Systems,2005,13(3):343-361.
    [7]BELIAKOV G, PRADERA A, CALVO T. Aggregation Functions:A Guide for Practitioners[M]. Berlin: Springer,2007.
    [8]DUBOIS D, PRADE H. Putting rough sets and fuzzy sets together[J]. Intelligent Decision Support, 1992:203-232.
    [9]Wu W Z, ZHANG W X. Constructive and axiomatic approaches of fuzzy approximation operators [J]. Information Sciences,2004,159(3-4):233-254.
    [10]Wu W Z, MI J S, ZHANG W X. Generalized fuzzy rough sets[J]. Information Sciences,2003,151:263-282.
    [11]RADZIKOWSKA A M, KERRE E E. A comparative study of fuzzy rough sets[J]. Fuzzy Sets and Systems, 2002,126(2):137-155.
    [12]COCK M D, CORNELIS C, KERRE E E. Fuzzy rough sets:The forgotten step[J]. IEEE Transactions on Fuzzy Systems,2007,15(1):121-130.
    [13]JENSEN R, CORNELIS C. Fuzzy-rough instance selection[C]//Proceedings of the 19th International Con-ference on Fuzzy Systems. Piscataway, NJ:IEEE Press,2010:1776-1782.
    [14]Yao Y Y. Combination of rough and fuzzy sets based on α-level sets[M]//. LIN T, CEREONE N. Rough Sets and Data Mining:Analysis of Imprecise Data. Norwell, MA:Kluwer,1997:301-321.
    [15]CORNELIS C, COCK M D, RADZIKOWSKA A M. Vaguely Quantified Rough Sets[J]. Lecture Notes in Artificial Intelligence,2007,4482:87-94.
    [16]ZADEHh L A. A computational approach to fuzzy quantifiers in natural languages[J], Computers and Mathematics with Applications,1983,9(1):149-184.
    [17]KELLER J M, GRAY M R, GIVENS J A. A fuzzy K-nearest neighbor algorithm[J]. IEEE Transactions on Systems, Man, And Cybernetics,1985,15(2):580-588.
    [18]SARKAR M. Fuzzy-Rough nearest neighbors algorithm[J]. Fuzzy Sets and Systems,2007,158:2123-2152.
    [19]WANG X, YANG J, TENG X, et al. Fuzzy-Rough Set Based Nearest Neighbor Clustering Classification Algorithm[J], Lecture Notes in Computer Science,2005,3613/2005:370-373.
    [20]LIANG-YAN S, LI C. A Fast and Scalable Fuzzy-rough Nearest Neighbor Algorithm[J]. WRI Global Congress on Intelligent Systems,2009,4:311-314.
    [21]BIAN H, MAZLACK L. Fuzzy-Rough Nearest-Neighbor Classification Approach[C]//Proceeding of the 22nd International Conference of the North American Fuzzy Information Processing Society. Piscat-away, NJ:IEEE Press,2003:500-505.
    [22]JENSEN R, CORNELIS C. A new approach to fuzzy-rough nearest neighbour classification[C]//Proceedings of the 6th International Conference on Rough Sets and Current Trends in Computing. Berlin:Springer, 2008:310-319.
    [23]GENTON M G. Classes of kernels for machine learning:a statistics perspective[J]. Journal of Machine Learning Research,2001,2:299-312.
    [24]MOSER B. On the T-transitivity of kernels[J]. Fuzzy Sets and Systems,2006,157:1787-1796.
    [25]Hu Q, CHEN D, Yu D, et al. Kernelised fuzzy rough sets[C]//Proceedings of the 4th International Conference on Rough Sets and Knowledge Technology. Berlin:Springer,2009:304-311.
    [26]Hu Q, ZHANG L, CHEN D, et al. Gaussian kernel based fuzzy rough sets:Model, uncertainty measures and applications[J]. International Journal of Approximate Reasoning,2010,51(4):453-471.
    [27]SUCKLING J, PARTNER J, DANCE D, et al. The mammographic image analysis society digital mammogram database[C]//2nd International Workshop on Digital Mammography. New York:Elsevier,1994:211-221.
    [28]OLIVER A, FREIXENET J, MARTI R, et al. A novel breast tissue density classification methodology[J]. IEEE Transactions on Information Technology in Biomedicine,2008,12(1):55-65.
    [29]WOLFE J N. Risk for breast cancer development determined by mammographic parenchymal pattern[J]. Cancer,1976,37:2486-2492.
    [30]BOYD N F, BYNG J W, JONG R A, et al. Quantitative classification of mammographic densities and breast cancer risk:results from the Canadian National Breast Screening Study[J]. Journal Of The National Cancer Institute,1995,87(9):670-675.
    [31]TABAR L, TOT T, DEAN P. Breast Cancer:The Art and Science of Early Detection with Mammogra-phy[M]. Stuttgart, DE:Georg Thieme Verlag,2005.
    [32]Illustrated Breast Imaging Reporting and Data System BIRADS[R].3rd. Reston:American College of Radiology,1998.
    [33]LAM P B, VAVEK P M, GELLER B M, et al. The association of increased weight, body mass index, and tissue density with the risk of breast carcinoma in Vermont[J]. Cancer,2000,89:369-375.
    [34]HALL M A. Correlation-based Feature Selection for Discrete and Numeric Class Machine Learn-ing[C]//Proceedings of the 17th International Conference on Machine Learning. San Francisco, US: Morgan Kaufmann,2000:359-366.
    [35]KLEENE S C. Introduction to Metamathematics[M]. New York:Van Nostrand,1952.
    [36]DIENES S P. On an Implication Function in Many-Valued Systems of Logic[J]. Journal of Symbolic Logic,1949,14(2):95-97.
    [37]BENGIO Y, GRANDVALET Y. Bias in Estimating the Variance of K-Fold Cross-Validation[M]//Statistical Modeling and Analysis for Complex Data Problems. New York:Springer,2005:75-95.
    [38]MACPARTHALAIN N, JENSEN R, SHEN Q, et al. Fuzzy-rough approaches for mammographic risk analysis[J]. Intelligent Data Analysis,2010,14(2):225-244.
    [39]COHEN J. A coefficient of agreement for nominal scales[J]. Educational and Psychological Measurement, 1960,20(1):37-46.
    [40]AHA D. Instance-based learning algorithm[J]. Machine Learning,1991,6:37-66.
    [41]JOHN G H, LANGLEY P. Estimating Continuous Distributions in Bayesian Classifiers[C]//Proceedings of the 11th Conference on Uncertainty in Artificial Intelligence. San Mateo, CA:Morgan Kaufmann, 1995:338-345.
    [42]CORTES C, VAPNIK V. Support-vector networks[J]. Machine Learning,1995,20:273-297.
    [43]WITTEN I H, FRANK E. Generating Accurate Rule Sets Without Global Optimisation[C]//Proceedings of the 15th International Conference on Machine Learning. San Mateo, CA:Morgan Kaufmann,1998.
    [44]WITTEN I H, FRANK E. Data Mining:Practical machine learning tools with Java implementations[M]. San Francisco, US:Morgan Kaufmann,2000.
    [45]COHEN W W. Fast Effecitve Rule Induction[C]//Proceedings of the 12th International Conference on Machine Learning. San Mateo, CA:Morgan Kaufmann,1995:115-123.
    [46]BORKOWSKI L. Selected works by Jan Lukasiewicz[M]. Amsterdam:North-Holland Pub. Co.,1970: 87-88.
    [47]BECK J R, SCHULTZ E K. The use of relative operating characteristic (ROC) curves in test performance evaluation[J]. Archives of pathology & laboratory medicine,1986,110:13-20.
    [48]BLAKE C L, MERZ C J. UCI Repository of machine learning databases[DB]. University of California, Irvine, School of Information and Computer Sciences.
    [49]ARMANTNO C, LEARDIA R, LANTERIA S, et al. Chemometric analysis of Tuscan olive oils[J]. Chemometrics and Intelligent Laboratory Systems,1989,5(4):343-354.
    [50]HUANG J, LING C. Using AUC and accuracy in evaluating learning algorithms[J]. IEEE Transactions on Knowledge and Data Engineering,2005,17(3):299-310.
    [51]HAND D, TILL R. A Simple Generalisation of the Area Under the ROC Curve for Multiple Class Classi-fication Problems[J]. Machine Learning,2001,45:171-186.
    [52]WILCOXON F. Individual comparisons by ranking methods[J]. Biometrics Bulletin,1945, 1(6):80-83.
    [53]GOULDEN C H. Methods of Statistical Analysis[M].2nd. New York:Wiley,1956:50-55.
    [54]YAGER R. On ordered weighted averaging aggregation operators in multi-criteria decision making[J]. IEEE Transactions on Systems, Man, and Cybernetics,1988,18:183-190.
    [55]QUINLAN J R. C4.5:Programs for Machine Learning[M]. The Morgan Kaufmann Series in Machine Learning. San Mateo, CA:Morgan Kaufmann,1993.
    [56]SMOLA A J, SCHOLKOPF B. A tutorial on support vector regression[J]. Statistics and Computing,2004, 14(3):199-222.
    [57]Qu Y, SHANG C, SHEN Q, et al. Kernel-Based Fuzzy-Rough Nearest Neighbour Classifica-tion [C]//Proceedings of the 20th International Conference on Fuzzy Systems. Piscataway, NJ:IEEE Press,2011:1523-1529.
    [58]ZHU Q, QIN A, SUGANTHAN P N, et al. Evolutionary Extreme Learning Machine[J]. Pattern Recognition, 2005,38:1759-1763.
    [59]TAKAGI T, SUGENO M. Fuzzy identification of systems and its applications to modeling and control[J]. IEEE Transactions on Systems, Man, and Cybernetics,1985,15(1):116-132.
    [60]HUANG G, ZHU Q, SIEW C K. Extreme learning machine:Theory and applications [J]. Neurocomputing, 2006,70(1-3):489-501.
    [61]STORN R, PRICE K. Differential evolution-a simple and efficient heuristic for global optimization over continuous spaces[J]. Journal of Global Optimization,1997,11:341-359.
    [62]RAO C, MITRA S K. Generalized Inverse of Matrices and its Applications[M]. New York:Wiley,1971.
    [63]SERRE D. Matrices:Theory and Applications[M]. New York:Springer,2002.
    [64]GHOSH R, VERMA B. A hierarchical method for finding optimal architecture andweights using evolution-ary least square based learning[J]. International Journal of Neural Systems,2003,12(1):13-24.
    [65]JANG J S R. ANFIS:adaptive-network-based fuzzy inference system[J]. IEEE Transactions on Systems, Man and Cybernetics,1993,23(3):665-685.
    [66]JUANG C, CHEN J. Water bath temperature control by a recurrent fuzzy controller and its FPGA imple-mentation[J]. International Journal of Approximate Reasoning,2006,53(3):941-949.
    [67]ER M J, LIU F, LI M B. Self-constructing Fuzzy Neural Networks with Extended Kalman Filter[J]. International Journal of Fuzzy Systems,2010,12(1):66-72.
    [68]K. CHATURVED M P, SRIVASTAVA L. Modified neo-fuzzy neuron-based approach for economic and envi- ronmental optimal power dispatch[J]. Applied Soft Computing,2008,8:1428-1438.
    [69]Wu W, LONG L, YANG J, et al. A modified gradient-based neuro-fuzzy learning algorithm and its conver-gence[J]. Information Sciences,2010,180(9):1630-1642.
    [70]XU Z, ZHANG R, JING W. When does online BP training converge?[J]. IEEE Transactions on Neural Networks,2009,20(10):1529-1539.
    [71]Wu W, FENG G, LI Z, et al. Deterministic Convergence of an Online Gradient Method for BP Neural Networks[J]. IEEE Transactions on Neural Networks,2005,16(3):533-540.
    [72]ZHANG H, Wu W, LIu F, et al. Boundedness and convergence of online gradient method with penalty for feedforward neural networks[J]. IEEE Transactions on Neural Networks,2009,20(6):1050-1054.
    [73]ZHANG N, Wu W, ZHENG G. Convergence of gradient method with momentum for two-Layer feedforward neural networks[J]. IEEE Transactions on Neural Networks,2006,17(2):522-525.
    [74]GORI M, MAGGINI M. Optimal convergence of on-line backpropagation[J]. IEEE Transactions on Neural Networks,1996,7(1):251-254.
    [75]Wu W, YANG J, CHENG M, et al. Convergence analysis of online gradient method for BP neural network-s[J]. Neural Networks,2011,24:91-98.
    [76]WANG J, YANGJ, Wu W. Convergence of Cyclic and Almost-Cyclic Learning with Momentum for Feedforward Neural Networks[J]. IEEE Transactions on Neural Networks,2011,22:1297-1306.
    [77]YUAN Y, SUN W. Optimization Theory and Methods[M]. Beijing:Science Press,2001.
    [78]JENSEN R, SHEN Q. Computational Intelligence and Feature Selection:Rough and Fuzzy Approaches[M]. Washington, US:IEEE Press and Wiley,2008.
    [79]BOONGOEN T, SHEN Q. Nearest-neighbor guided evaluation of data reliability and its applications[J]. IEEE Transactions on Systems, Man and Cybernetics, Part B:Cybernetics,2010,40(6):1622-1633.
    [80]SHANG C, BARNES D, SHEN Q. Facilitating efficient Mars terrain image classification with fuzzy-rough feature selection[J]. International Journal of Hybrid Intelligent Systems,2011,8(1):3-13.

© 2004-2018 中国地质图书馆版权所有 京ICP备05064691号 京公网安备11010802017129号

地址:北京市海淀区学院路29号 邮编:100083

电话:办公室:(+86 10)66554848;文献借阅、咨询服务、科技查新:66554700