Computer Science ›› 2021, Vol. 48 ›› Issue (10): 185-190.doi: 10.11896/jsjkx.200800219

• Database & Big Data & Data Science • Previous Articles     Next Articles

Sample Feature Kernel Matrix-based Sparse Bilinear Regression

SHAO Zheng-yi1, CHEN Xiu-hong1,2   

  1. 1 School of Artificial Intelligence and Computer Science,Jiangnan University,Wuxi,Jiangsu 214122,China
    2 Jiangsu Key Laboratory of Media Design and Software Technology,Wuxi,Jiangsu 214122,China
  • Received:2020-08-30 Revised:2020-11-11 Online:2021-10-15 Published:2021-10-18
  • About author:SHAO Zheng-yi,born in 1996,M.S candidate.His main research interests include pattern recognition,linear regression,etc.

Abstract: There are a large number of redundant data in many real applications,which may be high dimensional.In this case,there will be many problems in regression prediction,such as overfitting and low prediction accuracy.In addition,most regression methods are based on vectors,ignoring the relationship between the original positions of matrix data.To this end,a sample kernel matrix-based sparse bilinear regression (KMSBR) method is proposed.The KMSBR model which use the sample feature kernel matrix and L2,1-norm is established through the left and right regression coefficient matrix.Thus,the KMSBR can implement the selection of samples and its features simultaneously.Experimental results on several data sets show that KMSBR can effectively select samples and its features,thus improve the efficiency of the algorithm,and the prediction accuracy is better than the existing regression models.

Key words: Feature kernel matrix, Left and right regression matrix, Linear regression, Sample and feature extraction, Sparsity

CLC Number: 

  • TP391.4
[1]CHATTERJEE S,HADI A S.Regression analysis by example(5th edition)[M].John Wiley & Sons Inc,2012:31-34.
[2]DAVID J O.Linear regression [M].Springer,2017:1-15.
[3]RHINEHART R R.Nonlinear regression modeling for engi-neering applications:modeling,model validation,and enabling design of experiments [M].John Wiley & Sons,2016:12-30.
[4]BASAK D,SRIMANTA P,PATRANBIS D C.Support vectorregression[J].Neural Information Processing Letters & Reviews,2007,11(10):203-224.
[5]CHU W,KEERTHI S S,CHONG J O.Bayesian support vector regression using a unified loss function[J].IEEE Transactions on Neural Networks,2004,15(1):29-44.
[6]ZHAO Y,SUN J.Robust support vector regression in the primal[J].Neural Networks,2008,21(10):1548-1555.
[7]SUYKENS J A K,GESTEL V T,BRABANTER J D,et al.Least squares support vector machines [M].World Scientific,2002:10-12.
[8]CHEN C F.A robust weighted least squares support vector regression based on least trimmed squares[J].Neurocomputing,2015,168:941-946.
[9]TIAN Y,GAO G,YUE J,et al.Two-phase multi-kernel LP-SVR for feature sparsification and forecasting[J].Neurocompu-ting,2016,214:594-606.
[10]ZHANG Z W,HE J,GAO G X,et al.Bi-sparse optimization-based least squares regression[J].Applied Soft Computing,2019,77:300-315.
[11]ZHANG Z,GAO G,YAO T,et al.An interpretable regression approach based on bi-sparse optimization[J].Applied Intelligence,2020,50(11):4117-4112.
[12]STORK D G,HART P E,PATTERN R O.Classification[M].Hoboken,New York,USA,2001:307-321.
[13]LIU J,JI S W,YE J P.Multi-Task Feature Learning Via Efficient l2,1-Norm Minimization [C]//Conference on Uncertrainty in Artificial Intelligence.2009:339-348.
[14]NIE F,HUANG H,CAI X,et al.Efficient and robust feature selection via joint ℓ2,1-norms minimization[C]//International Conference on Neural Information Processing Systems.Curran Associates Inc.2010:1813-1821.
[15]CAI X,NIE F,HUANG H.Exact Top-k Feature Selection via l2,0-Norm Constraint[C]//International Joint Conference on Artificial Intelligence.AAAI Press,2013:1240-1246.
[16]LIU X,WANG L,ZHANG J,et al.Global and Local Structure Preservation for Feature Selection[J].IEEE Transactions on Neural Networks & Learning Systems,2014,25(6):1083-1095.
[17]HOU C,JIAO Y,NIE F,et al.2D Feature Selection by Sparse Matrix Regression[J].IEEE Transactions on Image Processing,2017,26(9):4255-4268.
[18]GABRIEL K R.Generalised bilinear regression[J].Biometrika,1998,85:689-700.
[19]TIBSHIRANI R.Regression shrinkage and selection via theLasso:a retrospective[J].Journal of the Royal Statistical Society:Series B,1994,73(3):267-288.
[20]Репозиторий UCI.UCI Machine Learning Repository:Data Sets [EB/OL].[2018-09-24].http://archive.ics.uci.edu/ml.
[1] LYU You, WU Wen-yuan. Privacy-preserving Linear Regression Scheme and Its Application [J]. Computer Science, 2022, 49(9): 318-325.
[2] LI Xia, MA Qian, BAI Mei, WANG Xi-te, LI Guan-yu, NING Bo. RIIM:Real-Time Imputation Based on Individual Models [J]. Computer Science, 2022, 49(8): 56-63.
[3] SUN Xiao-han, ZHANG Li. Collaborative Filtering Recommendation Algorithm Based on Rating Region Subspace [J]. Computer Science, 2022, 49(7): 50-56.
[4] WANG Mei-ling, LIU Xiao-nan, YIN Mei-juan, QIAO Meng, JING Li-na. Deep Learning Recommendation Algorithm Based on Reviews and Item Descriptions [J]. Computer Science, 2022, 49(3): 99-104.
[5] CHEN Chang-wei, ZHOU Xiao-feng. Fast Local Collaborative Representation Based Classifier and Its Applications in Face Recognition [J]. Computer Science, 2021, 48(9): 208-215.
[6] MA Feng-fei, LIN Su-zhen, LIU Feng, WANG Li-fang, LI Da-wei. Semantic-contrast Generative Adversarial Network Based Highly Undersampled MRI Reconstruction [J]. Computer Science, 2021, 48(4): 169-173.
[7] YANG Zhang-jing, WANG Wen-bo, HUANG Pu, ZHANG Fan-long, WANG Xin. Local Weighted Representation Based Linear Regression Classifier and Face Recognition [J]. Computer Science, 2021, 48(11A): 351-359.
[8] XU Bing, YI Pei-yu, WANG Jin-ce, PENG Jian. High-order Collaborative Filtering Recommendation System Based on Knowledge Graph Embedding [J]. Computer Science, 2021, 48(11A): 244-250.
[9] LU Ai-hong, GUO Yan, LI Ning, WANG Meng, LIU Jie. Direction-of-arrival Estimation with Two-dimensional Sparse Array Based on Atomic NormMinimization [J]. Computer Science, 2020, 47(5): 271-276.
[10] ZHAO Nan, PI Wen-chao, XU Chang-qiao. Video Recommendation Algorithm for Multidimensional Feature Analysis and Filtering [J]. Computer Science, 2020, 47(4): 103-107.
[11] HOU Ming-xing,QI Hui,HUANG Bin-ke. Data Abnormality Processing in Wireless Sensor Networks Based on Distributed Compressed Sensing [J]. Computer Science, 2020, 47(1): 276-280.
[12] LU Zhu-bing, LI Yu-zhou. Recommendation Strategy Based on Trust Model via Emotional Analysis of Online Comment [J]. Computer Science, 2019, 46(6): 75-79.
[13] ZHOU Ming,JIA Yan-ming,ZHOU Cai-lan,XU Ning. English Automated Essay Scoring Methods Based on Discourse Structure [J]. Computer Science, 2019, 46(3): 234-241.
[14] WANG Yong, WANG Yong-dong, DENG Jiang-zhou, ZHANG Pu. Recommendation Algorithm Based on Jensen-Shannon Divergence [J]. Computer Science, 2019, 46(2): 210-214.
[15] ZHU Zhi-cheng, LIU Jia-wei, YAN Shao-hong. Research and Application of Multi-label Learning in Intelligent Recommendation [J]. Computer Science, 2019, 46(11A): 189-193.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!