Computer Science ›› 2023, Vol. 50 ›› Issue (11A): 220900003-9.doi: 10.11896/jsjkx.220900003

• Big Data & Data Science • Previous Articles     Next Articles

Orthogonal Locality Preserving Projection Unsupervised Feature Selection Based on Graph Embedding

ZHU Jianyong1,2, LI Zhaoxiang1,2, XU Bin1,2, YANG Hui1,2, NIE Feiping3   

  1. 1 School of Electrical and Automation Engineering,East China Jiaotong University,Nanchang 330013,China
    2 Key Laboratory of Advanced Control & Optimization of Jiangxi Province(East China Jiaotong University),Nanchang 330013,China
    3 School of Artificial Intelligence,OPtics and ElectroNics(iOPEN),Northwestern Polytechnical University,Xi'an 710072,China
  • Published:2023-11-09
  • About author:ZHU Jianyong,born in 1977,Ph.D,associate professor,master supervisor.His main research interests include data analysis,random distribution control and predictive control.
    NIE Feiping,born in 1977,Ph.D,professor,Ph.D supervisor.His main research interests include machine learning and its applications,such as pattern recognition,data mining,computer vision,image processing and information retrieval.
  • Supported by:
    National Natural Science Foundation of China(61963015,61733005).

Abstract: The traditional unsupervised feature selection algorithm based on graph learning often adopts sparse regularization method.However,this approach relies too heavily on the efficiency of graph learning,and it is not easy to tune regularization parameters.To solve this problem,an unsupervised feature selection algorithm based on graph embedding learning with orthogonal locality preserving projection is proposed in this paper.Firstly,we utilize locality preserving projection method to enhance the linear mapping ability that can maintain the local geometric manifold structure of the data,and orthogonal projection mode brings convenience to data reconstruction.Moreover,we use graph embedding learning method to quickly learn the similarity matrix of data.Then,$\ell$2,0-norm constrained projection matrix to select discriminative features.Finally,a new nonparametric algorithm is used to efficiently solve the model problem iteratively since $\ell$2,0-norm belongs to NP problem.Experimental results prove the effectiveness and superiority of the proposed algorithm.

Key words: Unsupervised feature selection, Orthogonal locality preserving projection, Graph embedding learning, $\ell$2, 0-norm, Nonparametric iterative algorithm

CLC Number: 

  • TP391
[1]WANG Y Z,JIN X L,CHENG X Q.Network Big Data:Present and Future[J].Chinese Journal of Computers,2013,36(6):1125-1138.
[2]YAO X,WANG X D,ZHENG Y X,et al.Summary of feature selection algorithms[J].Control and Decision,2012,27:161-166,192.
[3]ZHU X,HUANG Z,YANG Y,et al.Self-taught dimensionality reduction on the high-dimensional small-sized data[J].Pattern Recognition,2013,46(1):215-229.
[4]GU Q,LI Z,HAN J.Generalized Fisher score for feature selec-tion[C]//Proceedings of the 27th Conference on Uncertainty in Artificial Intelligence.Barcelona,2011:266-273.
[5]XU Z,KING I,LYU M R T,et al.Discriminative semi-super-vised feature selection via manifold regularization[J].IEEE Transactions on Neural Networks,2010,21:1033-1047.
[6]LIU Y F,LI W B,GAO Y.Adaptive Neighborhood Embedding Based Unsupervised Feature Selection[J].Journal of Computer Research and Development,2020,57(8):11639-11649.
[7]CAI D,ZHANG C Y,HE X F.Unsupervised feature selection for Multi-Cluster data[C]//Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.2010:333-342
[8]LUO Y,TAO D,XU C,et al.Vector-valued multi-view semi-supervsed learning for multi-label image classification[C]//27th AAAI Conference on Artificial Intelligence.Bellevue,2013:647-653.
[9]SOLORIO-FERNÁNDEZ S,MARTÍNEZ-TRINIDAD J F,CARRASCO-OCHOA J A.A new unsupervised spectral feature selection method for mixed data:A filter approach[J].Pattern Recognition,2017,72:314-326.
[10]LIU S W,CHEN X,GU Q,et al.A cluster-analysis-based feature-selection method for software defect prediction[J].Scientia Sinica Informationis,2016,46(9),1298-1320.
[11]GUYON I,WESTON J,BARNHILL S,et al.Gene selection for cancer classification using support vector machines[J].Machine Learning,2002,46(1/2/3):389-422.
[12]BELKIN M,NIYOGI P.Laplacian Eigenmaps and SpectralTechniques for Embedding and Clustering[J].Advances in Neural Information Processing Systems,2001,14(6):585-591.
[13]YANG Y,SHEN H T,MA Z G,et al.L2,1-Norm Regularized Discriminative Feature Selection for Unsupervised Learning[C]//22th AAAI Conference on Artificial Intelligence.2011:1589-1594.
[14]HAN K,WANG Y H;ZHANG C,Li C,Xu C.AutoEncoder Inspired Unsupervised Feature Selection[C]//International Conference on Acoustics,Speech and Signal Processing.2018.
[15]ZHANG R,NIE F P,WANG Y H,et al.Unsupervised Feature Selection via Adaptive Multimeasure Fusion[J].IEEE Transactions on Neural Networks and Learning Systems,2019,30:2886-2892.
[16]GUI J,SUN Z,JI S,et al.Feature selection based on structured sparsity:A comprehensive study[J].IEEE Transactions on Neural Network and Learning System,2017,28:1490-1507.
[17]NIE F P,WEI Z,LI X L.Unsupervised feature selection with structured graph optimization[C]//Proceedings of the Thirtieth AAAI Conference on Artificial.Palo Alto:AAAI,2016:1302-1308.
[18]QIAN B,TANG Z M,LI X,et al.Feature selection based on manifold discriminant information and its structured sparse representation[J].Control and Decision,2016,31(7):1272-1278.
[19]HUANG D,CAI X,WANG C D.Unsupervised Feature Selection with Multi-Subspace Randomization and Collaboration[J].Knowledge-Based Systems,2019,182:104856.1-104856.15.
[20]CAI X,NIE F P,HUANG H.Exact top-k feature selection via L2,0-norm constraint[C]//Proceedings of the Twenty-Third International Joint Conference on Artificial Intelligence.2013:240-1246.
[21]DU X Z,NIE F P,WANG W Q,et al.Exploiting Combination Effect for Unsupervised Feature Selection by L2,0 Norm[J].IEEE Transactions on Neural Networks and Learning Systems,2018,30:201-214.
[22]YANG W,XU H.Streaming sparse principal component analysis[C]//International Conference on Machine Learning.2015:494-503.
[23]KUNDU A,DRINEAS P,MAGDON-ISMAIL M.Recoveringpca and sparse pca via hybrid-(L1,L2) sparse sampling of data elements[J].The Journal of Machine Learning Research,2017,18:2558-2591.
[24]PANG T J,NIE F P,HAN J W.Efficient Feature Selection via L2,0-norm Constrained Sparse Regression[J].IEEE Transactions on Knowledge and Data Engineering,2019,31:880-893.
[25]HE X F,NIYOGI P.Locality preserving projections[J].Advances in Neural Information Processing Systems,2003,16(1):186-197.
[26]ROWEIS S,SAUL L.Nonlinear dimensionality reductionby locally linear embedding[J].Science,2000,290(5000):2323-2326.
[27]NG A Y,JORDAN M I,WEISS Y.On spectral clustering:Ana-lysis and an algorithm[C]//Adv.Neural Inf.Process.Syst,2001,14:849-856.
[28]SHI J,MALIK J M.Normalized Cuts and Image Segmentation[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,2000,22:888-905.
[29]NIE F P,WANG X,JORDAN M I,et al.The constrained laplacian rank algorithm for graph-based clustering[C]//Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence.2016:1023-1031.
[30]LUO Z Q,YU W.An introduction to convex optimization forcommunications and signal processing[J].IEEE Journal on Selected Areas in Communications,2006,24:1426-1438.
[31]HE X F,CAI D,NIYOGI P.Laplacian score for feature selection[C]//Advances in Neural Information Processing Systems.Vancouver,2005:507-514.
[1] WANG Xiaofeng, LI Chaoran, LU Kunfeng, LUAN Tianjiao, YAO Na, ZHOU Hui, XIE Yujia. Acceleration Design and FPGA Implementation of CNN Scene Matching Algorithm [J]. Computer Science, 2023, 50(11): 8-14.
[2] DING Yue, XU Chuanfu, QIU Haozhong, DAI Weixi, WANG Qingsong, LIN Yongzhen, WANG Zhenghua. Study on Cross-platform Heterogeneous Parallel Computing for Lattice Boltzmann Multi-phase Flow Simulations Based on SYCL [J]. Computer Science, 2023, 50(11): 32-40.
[3] DONG Hongbin, HAN Shuang, FU Qiang. Geo-sensory Time Series Prediction Based on Joint Model of Auto Regression and Deep NeuralNetwork [J]. Computer Science, 2023, 50(11): 41-48.
[4] XIE Zhuo, KANG Le, ZHOU Lijuan, ZHANG Zhihong. Clustering Method Based on Contrastive Learning for Multi-relation Attribute Graph [J]. Computer Science, 2023, 50(11): 62-70.
[5] HE Wenhao, WU Chunjiang, ZHOU Shijie, HE Chaoxin. Study on Short Text Clustering with Unsupervised SimCSE [J]. Computer Science, 2023, 50(11): 71-76.
[6] ZHENG Wenping, WANG Fumin, LIU Meilin, YANG Gui. Graph Clustering Algorithm Based on Node Clustering Complexity [J]. Computer Science, 2023, 50(11): 77-87.
[7] LIU Qidong, LIU Chaoyue, QIU Zixin, GAO Zhimin, GUO Shuai, LIU Jizhao, FU Mingsheng. Time-aware Transformer for Traffic Flow Forecasting [J]. Computer Science, 2023, 50(11): 88-96.
[8] WU Yushan, XU Zengmin, ZHANG Xuelian, WANG Tao. Self-supervised Action Recognition Based on Skeleton Data Augmentation and Double Nearest Neighbor Retrieval [J]. Computer Science, 2023, 50(11): 97-106.
[9] ZHAO Xingwang, XUE Jinfang. Community Discovery Algorithm for Attributed Networks Based on Bipartite Graph Representation [J]. Computer Science, 2023, 50(11): 107-113.
[10] NI Hongjie, LIU Jiawei, ZHENG Haibin, CHEN Yipeng, CHEN Jinyin. NeuronSup:Deep Model Debiasing Based on Bias Neuron Suppression [J]. Computer Science, 2023, 50(11): 122-131.
[11] WANG Shangshang, JIN Cheng. Surface Anomaly Detection Based on Image Reconstruction and Semantic Difference Discrimination [J]. Computer Science, 2023, 50(11): 151-159.
[12] YU Yang, YUAN Jiabin, CAI Jiyuan, ZHA Keke, CHEN Zhangyu, DAI Jiawei, FENG Yuxiang. Deepfake Face Tampering Video Detection Method Based on Non-critical Masks and AttentionMechanism [J]. Computer Science, 2023, 50(11): 160-167.
[13] TAO Xinyu, XIONG Lizhi, ZHANG Xiang. Robust Video Watermarking Scheme Based on QDCT Global Equalization Strategy [J]. Computer Science, 2023, 50(11): 168-176.
[14] WANG Xueguang, ZHU Junwen, ZHANG Aixin. Three-dimensional AI Clone Speech Source Identification Method Based on Improved MFCCFeature Model [J]. Computer Science, 2023, 50(11): 177-184.
[15] KANG Mengyao, LIU Yang, HUANG Junheng, WANG Bailing, LIU Shulong. Chat Dialogue Summary Model Based on Multi-granularity Contrastive Learning [J]. Computer Science, 2023, 50(11): 192-200.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!