计算机科学 ›› 2020, Vol. 47 ›› Issue (11A): 474-478.doi: 10.11896/jsjkx.200100037

• 大数据&数据科学 • 上一篇    下一篇

联合多流形结构和自表示的特征选择方法

易玉根1, 李世成1, 裴洋1, 陈磊1, 代江艳2   

  1. 1 江西师范大学软件学院 南昌 330022
    2 潍坊学院计算机工程学院 山东 潍坊 261061
  • 出版日期:2020-11-15 发布日期:2020-11-17
  • 通讯作者: 代江艳(daijyan@163.com)
  • 作者简介:yiyg510@jxnu.edu.cn
  • 基金资助:
    国家自然科学基金(61602221,61806126);江西省自然科学基金(20171BAB212009);江西省教育厅科技项目(GJJ160315,GJJ170234);山东省高等学校青创科技支持计划(2019KJN012)

Feature Selection Method Combined with Multi-manifold Structures and Self-representation

YI Yu-gen1, LI Shi-cheng1, PEI Yang1, CHEN Lei1, DAI Jiang-yan2   

  1. 1 School of Software,Jiangxi Normal University,Nanchang 330022,China
    2 School of Computer Engineering,Weifang University,Weifang,Shandong 261061,China
  • Online:2020-11-15 Published:2020-11-17
  • About author:YI Yu-gen,born in 1986,Ph.D,lecturer.His research interests include artificial intelligence,computer vision,and machine learning.
    DAI Jiang-yan,born in 1985,postdoctor,associate professor.Her main research interests include image inpain-ting and computer vision.
  • Supported by:
    This work was supported by the National Natural Science Foundation of China (61602221,61806126),Natural Science Foundation of Jiangxi Province (20171BAB212009),Science and Technology Research Project of Jiangxi Provincial Department of Education (GJJ160315,GJJ170234) and Youth Innovation Science and Technology Support Plan of Shandong Provincial Institution of Higher Education (2019KJN012).

摘要: 特征选择是一种通过去除不相关和冗余的特征来降低数据维数和提高后续学习算法效率的数据处理方法。无监督特征选择已经成为维数约简中具有挑战性的问题之一。首先,通过结合特征自表示能力和流形结构,提出了一种联合多流形结构和自表示(Joint Multi-Manifold Structures and Self-Representation,JMMSSR)的无监督特征选择方法。不同于现有的方法,为了更准确地刻画特征的流形结构,引入一种自适应加权策略来融合特征的多个流形结构。然后,提出了一种简单且有效的迭代优化算法来求解JMMSSR方法的目标函数,并利用数值实验验证了优化算法的收敛性。最后,分别在JAFFE,ORL和COIL20 3个数据集上进行聚类实验,实验结果验证了与现有的无监督特征选择方法相比,JMMSSR方法具有较好的性能。

关键词: 多流形结构, 特征选择, 自表示, 自适应加权

Abstract: Feature selection is to reduce the dimension of data by removing irrelevant and redundant features and improve the efficiency of learning algorithm.Unsupervised feature selection has become one of the challenging problems in dimensionality reduction.Firstly,combining self-representation and manifold structure of features,a Joint Multi-Manifold Structures and Self-Representation (JMMSSR) unsupervised feature selection algorithm is proposed.Different from the existing approaches,our approach designs an adaptive weighted strategy to integrate multi-manifold structures to describe the structure of features accurately.Then,a simple and effective iterative updating algorithm is proposed to solve the objective function,and the convergence of the optimization algorithm is also verified by numerical experiments.Finally,experimental results on three datasets (such as JAEEF,ORL and COIL20) show that the proposed approach exhibits better performance than the existing unsupervised feature selection approaches.

Key words: Adaptive weighted, Feature selection, Multi-manifold structures, Self-representation

中图分类号: 

  • TP391
[1] LI Y,LI T,LIU H.Recent advances in feature selection and its applications [J].Knowledge and Information Systems,2017,53(3):551-577.<br /> [2] LI J,LIU H.Challenges of feature selection for big data analy-tics [J].IEEE Intelligent Systems,2017,32(2):9-15.<br /> [3] GUI J,SUN Z,JI S,et al.Feature selection based on structured sparsity:A comprehensive study [J].IEEE Transactions on Neural Networks and Learning Systems,2016,28(7):1490-1507.<br /> [4] LI Z Q,DU J Q,NIE B,et al.Summary of Feature Selection Methods.CEA,2019,55(24):10-19.<br /> [5] FANG B,CHEN H M,WANG S W.Feature Selection Algo-rithm Based on Rough Sets and Fruit Fly Optimization[J].Computer Science,2019,46(7):157-164.<br /> [6] LI J,CHENG K,WANG S,et al.Feature selection:A data perspective [J].ACM Computing Surveys (CSUR),2018,50(6):94.<br /> [7] RIPLEY B D,HJORT N L.Pattern recognition and neural networks [M].Cambridge University press,1996.<br /> [8] HE X,CAI D,NIYOGI P.Laplacian score for feature selection [C]//Advances in Neural Information Processing Systems.2006:507-514.<br /> [9] ZHAO Z,LIU H.Spectral feature selection for supervised and unsupervised learning [C]//Proceedings of the 24th International Conference on Machine Learning.ACM,2007:1151-1157.<br /> [10] CAI D,ZHANG C,HE X.Unsupervised feature selection formulti-cluster data [C]//Proceedings of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.ACM,2010:333-342.<br /> [11] LI Z,YANG Y,LIU J,et al.Unsupervised feature selection using nonnegative spectral analysis [C]//Twenty-Sixth AAAI Conference on Artificial Intelligence.2012.<br /> [12] NIE F,HUANG H,CAI X,et al.Efficient and robust feature selection via joint l2,1-norms minimization [C]//Advances in Neural Information Processing Systems.2010:1813-1821.<br /> [13] QIAN M,ZHAI C.Robust unsupervised feature selection [C]//Twenty-Third International Joint Conference on Artificial Intelligence.2013.<br /> [14] ZHU P,ZUO W,ZHANG L,et al.Unsupervised feature selection by regularized self-representation [J].Pattern Recognition,2015,48(2):438-446.<br /> [15] LIANG S,XU Q,ZHU P,et al.Unsupervised feature selection by manifold regularized self-representation [C]//2017 IEEE International Conference on Image Processing (ICIP).IEEE,2017:2398-2402.<br /> [16] TANG C,ZHU X,CHEN J,et al.Robust graph regularized unsupervised feature selection [J].Expert Systems with Applications,2018,96:64-76.<br /> [17] QIAO L,ZHANG L,CHEN S,et al.Data-driven graph construction and graph learning:A review [J].Neurocomputing,2018,312:336-351.<br /> [18] BELKIN M,NIYOGI P.Laplacian eigenmaps and spectral techniques for embedding and clustering [C]//Advances in Neural Information Processing Systems.2002:585-591.<br /> [19] ROWEIS S T,SAUL L K.Nonlinear dimensionality reduction by locally linear embedding [J].Science,2000,290(5500):2323-2326.<br /> [20] CHENG B,YANG J,YAN S,et al.Learning With l1-Graph for Image Analysis [J].IEEE Transactions on Image Processing,2009,19(4):858-866.<br /> [21] LIU G,LIN Z,YAN S,et al.Robust recovery of subspace structures by low-rank representation [J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2013,35(1):171-184.<br /> [22] LU C Y,MIN H,ZHAO Z Q,et al.Robust and efficient subspace segmentation via least squares regression [C]//European Conference on Computer Vision.Springer,2012:347-360.<br /> [23] LYONS M,AKAMATSU S,KAMACHI M,et al.Coding facial s with gabor wavelets [C]//Proceedings Third IEEE International Conference on Automatic Face and Gesture Recognition.IEEE,1998:200-205.<br /> [24] SAMARIA F S,HARTER A C.Parameterisation of a stochastic model for human face identification [C]//Proceedings of 1994 IEEE Workshop on Applications of Computer Vision.1994:138-142.<br /> [25] NENE S A,NAYAR S K,MURASE H.Columbia object image library (COIL-20):CUCS-005-96 [R].1996.<br /> [26] FANG X,XU Y,LI X,et al.Orthogonal self-guided similaritypreserving projection for classification and clustering [J].Neural Networks,2017,88:1-8.
[1] 李斌, 万源.
基于相似度矩阵学习和矩阵校正的无监督多视角特征选择
Unsupervised Multi-view Feature Selection Based on Similarity Matrix Learning and Matrix Alignment
计算机科学, 2022, 49(8): 86-96. https://doi.org/10.11896/jsjkx.210700124
[2] 胡艳羽, 赵龙, 董祥军.
一种用于癌症分类的两阶段深度特征选择提取算法
Two-stage Deep Feature Selection Extraction Algorithm for Cancer Classification
计算机科学, 2022, 49(7): 73-78. https://doi.org/10.11896/jsjkx.210500092
[3] 康雁, 王海宁, 陶柳, 杨海潇, 杨学昆, 王飞, 李浩.
混合改进的花授粉算法与灰狼算法用于特征选择
Hybrid Improved Flower Pollination Algorithm and Gray Wolf Algorithm for Feature Selection
计算机科学, 2022, 49(6A): 125-132. https://doi.org/10.11896/jsjkx.210600135
[4] 储安琪, 丁志军.
基于灰狼优化算法的信用评估样本均衡化与特征选择同步处理
Application of Gray Wolf Optimization Algorithm on Synchronous Processing of Sample Equalization and Feature Selection in Credit Evaluation
计算机科学, 2022, 49(4): 134-139. https://doi.org/10.11896/jsjkx.210300075
[5] 孙林, 黄苗苗, 徐久成.
基于邻域粗糙集和Relief的弱标记特征选择方法
Weak Label Feature Selection Method Based on Neighborhood Rough Sets and Relief
计算机科学, 2022, 49(4): 152-160. https://doi.org/10.11896/jsjkx.210300094
[6] 李宗然, 陈秀宏, 陆赟, 邵政毅.
鲁棒联合稀疏不相关回归
Robust Joint Sparse Uncorrelated Regression
计算机科学, 2022, 49(2): 191-197. https://doi.org/10.11896/jsjkx.210300034
[7] 张叶, 李志华, 王长杰.
基于核密度估计的轻量级物联网异常流量检测方法
Kernel Density Estimation-based Lightweight IoT Anomaly Traffic Detection Method
计算机科学, 2021, 48(9): 337-344. https://doi.org/10.11896/jsjkx.200600108
[8] 杨蕾, 降爱莲, 强彦.
基于自编码器和流形正则的结构保持无监督特征选择
Structure Preserving Unsupervised Feature Selection Based on Autoencoder and Manifold Regularization
计算机科学, 2021, 48(8): 53-59. https://doi.org/10.11896/jsjkx.200700211
[9] 侯春萍, 赵春月, 王致芃.
基于自反馈最优子类挖掘的视频异常检测算法
Video Abnormal Event Detection Algorithm Based on Self-feedback Optimal Subclass Mining
计算机科学, 2021, 48(7): 199-205. https://doi.org/10.11896/jsjkx.200800146
[10] 胡艳梅, 杨波, 多滨.
基于网络结构的正则化逻辑回归
Logistic Regression with Regularization Based on Network Structure
计算机科学, 2021, 48(7): 281-291. https://doi.org/10.11896/jsjkx.201100106
[11] 周钢, 郭福亮.
基于特征选择的高维数据集成学习方法研究
Research on Ensemble Learning Method Based on Feature Selection for High-dimensional Data
计算机科学, 2021, 48(6A): 250-254. https://doi.org/10.11896/jsjkx.200700102
[12] 丁思凡, 王锋, 魏巍.
一种基于标签相关度的Relief特征选择算法
Relief Feature Selection Algorithm Based on Label Correlation
计算机科学, 2021, 48(4): 91-96. https://doi.org/10.11896/jsjkx.200800025
[13] 滕俊元, 高猛, 郑小萌, 江云松.
噪声可容忍的软件缺陷预测特征选择方法
Noise Tolerable Feature Selection Method for Software Defect Prediction
计算机科学, 2021, 48(12): 131-139. https://doi.org/10.11896/jsjkx.201000168
[14] 张亚钏, 李浩, 宋晨明, 卜荣景, 王海宁, 康雁.
混合人工化学反应优化和狼群算法的特征选择
Hybrid Artificial Chemical Reaction Optimization with Wolf Colony Algorithm for Feature Selection
计算机科学, 2021, 48(11A): 93-101. https://doi.org/10.11896/jsjkx.210100067
[15] 董明刚, 黄宇扬, 敬超.
基于遗传实例和特征选择的K近邻训练集优化方法
K-Nearest Neighbor Classification Training Set Optimization Method Based on Genetic Instance and Feature Selection
计算机科学, 2020, 47(8): 178-184. https://doi.org/10.11896/jsjkx.190700089
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!