JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE) ›› 2017, Vol. 52 ›› Issue (3): 91-96.doi: 10.6040/j.issn.1671-9352.4.2016.080

Previous Articles     Next Articles

Cost-sensitive feature selection via manifold learning

HUANG Tian-yi, ZHU William*   

  1. Laboratory of Granular Computing, Minnan Normal University, Zhangzhou 363000, Fujian, China
  • Received:2016-06-01 Online:2017-03-20 Published:2017-03-20

Abstract: In order to get a low-cost subset of original features, we define the cost-distance among the samples and joint it to existing feature selection framework. We combine manifold learning into cost-sensitive feature selection model and develop a corresponding method, namely, cost-sensitive feature selection via manifold learning(CFSM). Most previous cost-sensitive feature selection algorithms rank features individually and select features just using correlation the between the cost and the features. Our cost-sensitive feature selection algorithm selects features not only using the correlation the between the cost and the features but also using the discriminative information implied within data to improve the features selection performance. Experimental results on different real world datasets show the promising performance of CFSM outperforms the state-of-the-arts.

Key words: cost-sensitive, manifold learning, feature selection, supervised learning

CLC Number: 

  • O151.26
[1] SAITTA L. Machine learning — a technological roadmap[M]. Amsterdam: University of Amsterdam, 2001.
[2] FRASCA M, BASSIS S. Gene-disease prioritization through cost-Sensitive graph-based methodologies[C] //International Work-Conference on Bioinformatics and Biomedical Engineering. Berlin: Springer International Publishing, 2016:739-751.
[3] WEI Fan, STOLFO S J, ZHANG Jingdan, et al. Adacost: misclassification cost-sensitive boosting[C] //Sixteenth International Conference On Machine Learning. Burlington: Morgan Kaufmann Publishers Inc, 1999:97-105.
[4] TURNEY P D. Types of cost in inductive concept learning[C] //The Workshop on Cost-Sensitive Learning at the Seventeenth International Conference on Machine Learning. S. l: s. n, 2002:15-21.
[5] LU Jiwen, TAN Y P. Cost-Sensitive subspace analysis and extensions for face recognition[J]. IEEE Transactions on Information Forensics and Security, 2013, 8(3):510-519.
[6] LU Jiwen, ZHOU Xiuzhuang, TAN Y P, et al. Cost-sensitive semi-supervised discriminant analysis for face recognition[J]. IEEE Transactions on Information Forensics and Security, 2012, 7(3):944-953.
[7] ZADROZNY B, ELKAN C. Learning and making decisions when costs and probabilities are both unknown[C] //Seventh Acm Sigkdd International Conference on Knowledge Discovery and Data Mining. S. l: s. n, 2001:204-213.
[8] DOMINGOS P. MetaCost: a general method for making classifiers cost-sensitive[C] //Proceedings of the Fifth International Conference on Knowledge Discovery and Data Mining. S. l: s. n, 1999:155-164.
[9] MIAO Linsong, LIU Mingxia, ZHANG Daoqiang. Cost-sensitive feature selection with application in software defect prediction[C]. IEEE International Conference on Pattern Recognition, 2012:967-970.
[10] LU Jiwen, TAN Y P. Regularized locality preserving projections and its extensions for face recognition[J]. IEEE Transactions on Systems, Man, and Cybernetics, Part B(Cybernetics), 2009, 40(3): 958-963.
[11] BELKIN M, NIYOGI P. Laplacian eigenmaps and spectral techniques for embedding and clustering[J]. Advances in Neural Information Processing Systems, 2002, 14(6):585-591.
[12] CAI Deng, ZHANG Chiyuan, HE Xiaofei. Unsupervised feature selection for multi-cluster data[C] //ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. Washington Dc: s. n, 2010:333-342.
[13] ZHANG Yin, ZHOU Zhihua. Cost-sensitive face recognition[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2010, 32(10): 1758-1769.
[14] SHI Jianbo, MALIK J. Normalized cuts and image segmentation[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2000, 22(8): 888-905.
[15] ROWEIS S T, SAUL L K. Nonlinear dimensionality reduction by locally linear embedding[J]. Science, 2000, 290(5500): 2323-2326.
[16] NIE Feiping, HUANG Heng, CAI Xiao, et al. Efficient and robust feature selection via joint l2, 1-norms minimization[C] //Advances in Neural Information Processing Systems 23: Conference on Neural Information Processing Systems 2010. Proceedings of a Meeting Held 6-9 December 2010. Vancouver: s. n, 2010:1813-1821.
[17] EFRON B, HASTIE T, JOHNSTONE I, et al. Least angle regression[J]. The Annals of Statistics, 2004, 32(2):407-499.
[18] ZHAO Hong, MIN Fan, ZHU W. Cost-sensitive feature selection of data with errors[J]. Journal of Applied Mathematics, Article ID, 2013, 754698: 18.
[19] ZHU Pengfei, ZUO Wangmeng, ZHANG Lei, et al. Unsupervised feature selection by regularized self-representation[J]. Pattern Recognition, 2015, 48(2): 438-446.
[1] ZHANG Peng, WANG Su-ge, LI De-yu, WANG Jie. A semi-supervised spam review classification method based on heuristic rules [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2017, 52(7): 44-51.
[2] HUANG Wei-ting, ZHAO Hong, ZHU William. Adaptive divide and conquer algorithm for cost-sensitive attribute reduction [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2016, 51(8): 98-104.
[3] WAN Zhong-ying, WANG Ming-wen, ZUO Jia-li, WAN Jian-yi. Feature selection combined with the global and local information(GLFS) [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2016, 51(5): 87-93.
[4] SU Feng-long, XIE Qing-hua, HUANG Qing-quan, QIU Ji-yuan, YUE Zhen-jun. Semi-supervised method for attribute extraction based on transductive learning [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2016, 51(3): 111-115.
[5] LI Zhao,SUN Zhan-,LI Xiao,LI Cheng,. Study on feature selection method based on information loss [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2016, 51(11): 7-12.
[6] DU Hong-le, ZHANG Yan, ZHANG Lin. Intrusion detection on imbalanced dataset [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2016, 51(11): 50-57.
[7] CHEN Ya-dong, HONG Yu, YANG Xue-rong, WANG Xiao-bin, YAO Jian-min, ZHU Qiao-ming. Automatic target identification in frame semantic parsing [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2015, 50(07): 45-53.
[8] XIA Meng-nan, DU Yong-ping, ZUO Ben-xin. Micro-blog opinion analysis based on syntactic dependency and feature combination [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2014, 49(11): 22-30.
[9] ZHENG Yan, PANG Lin, BI Hui, LIU Wei, CHENG Gong. Feature selection algorithm based on sentiment topic model [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2014, 49(11): 74-81.
[10] PAN Qing-qing, ZHOU Feng, YU Zheng-tao, GUO Jian-yi, XIAN Yan-tuan. Recognition method of Vietnamese named entity based on#br# conditional random fields [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2014, 49(1): 76-79.
[11] ZHANG Li-bo, LI Hua-xiong, ZHOU Xian-zhong, HUANG Bing. Multi-granularity cost-sensitive three-way decision for face recognition [J]. JOURNAL OF SHANDONG UNIVERSITY(NATURAL SCIENCE), 2014, 49(08): 48-57.
[12] YU Ran 1,2, LIU Chun-yang3*, JIN Xiao-long 1, WANG Yuan-zhuo 1, CHENG Xue-qi 1. Chinese spam microblog filtering based on the fusion of
multi-angle features
[J]. J4, 2013, 48(11): 53-58.
[13] DU Shi-qiang1, SHI Yu-qing2, WANG Wei-lan1, MA Ming1. Manifold regularized-based discriminant concept factorization [J]. J4, 2013, 48(05): 63-69.
[14] ZENG Weng-fu1, HUANG Tian-qiang1,2, LI Kai1, YU YANG-qiang1, GUO Gong-de1,2. A local linear emedding agorithm based on harmonicmean geodesic kernel [J]. J4, 2010, 45(7): 55-59.
[15] YI Chao-qun, LI Jian-ping, ZHU Cheng-wen. A kind of feature selection based on classification accuracy of SVM [J]. J4, 2010, 45(7): 119-121.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!