计算机科学 ›› 2020, Vol. 47 ›› Issue (9): 190-197.doi: 10.11896/jsjkx.200700077

• 人工智能 • 上一篇    下一篇

基于边际概率分布匹配的主动标记分布学习

董心悦, 范瑞东, 侯臣平   

  1. 国防科技大学文理学院 长沙410008
  • 收稿日期:2020-05-05 发布日期:2020-09-10
  • 通讯作者: 侯臣平(hcpnudt@hotmail.com)
  • 作者简介:dongxinyue96@163.com
  • 基金资助:
    国家自然科学基金(61922087,61906201);湖南省杰出青年自然科学基金(2019JJ20020)

Active Label Distribution Learning Based on Marginal Probability Distribution Matching

DONG Xin-yue, FAN Rui-dong, HOU Chen-ping   

  1. College of Liberal Arts and Sciences,National University of Defense Technology,Changsha 410008,China
  • Received:2020-05-05 Published:2020-09-10
  • About author:DONG Xin-yue,born in 1996,postgra-duate.Her main research interests include statistical data analysis and machine learning.
    HOU Chen-ping,born in 1982,Ph.D,professor,Ph.D supervisor,is a member of China Computer Federation.His main research interests include machine learning,statistical data analysis,pattern recognition and computer vision.
  • Supported by:
    National Natural Science Foundation of China (61922087,61906201) and Natural Science Foundation for Distinguished Young Scholars of Hunan Province (2019JJ20020).

摘要: 标记分布学习是在以标记分布标注的示例上学习的新型学习范式,近年来已成功应用于面部年龄估计、头部姿势估计和情感识别等实际场景中。在标记分布学习中,需要足够多的标记分布数据才能训练出预测性能好的模型。然而,标记分布学习有时会面临标记数据不足和注释成本太高的困境。基于边际概率分布匹配的主动标记分布学习(Active Label Distribution Learning Based on Marginal Probability Distribution Matching,ALDL-MMD)算法是针对标记分布学习注释成本过高的问题而设计的,以减少训练模型所需的标注数据量,从而降低注释成本。ALDL-MMD算法训练了一个线性回归模型,在保证其训练误差最小的同时,学习一个反映未标记数据上选点需求的稀疏向量,使选点后的训练集和未标记集的数据分布尽量相似,并对这个向量做松弛化处理,以简计算。在多个标记分布数据集上的实验结果表明,在“Canberra Metric”和“Intersection”这两个衡量标记分布的指标上,ALDL-MMD算法优于已有的主动示例选择方法,体现了其在降低注释成本方面的有效性。

关键词: 边际概率分布匹配, 标记分布学习, 线性模型, 主动学习, 最大平均差异

Abstract: Label distribution learning (LDL)is a new learning paradigm for learning on instances labeled with label distribution,and has been successfully applied to real world scenes such as face age estimation,head pose estimation,and emotion recognition in recent years.In label distribution learning,enough data labeled by label distribution is needed when people train a model with good prediction performance.However,label distribution learning sometimes faces the dilemma that labeled data is insufficient and that marking enough label distribution data means high annotation cost.The Active label distribution learning based on marginal probability distribution matching (ADLD-MMD)algorithm is designed to solve the problem of high annotation cost for label distribution learning,by reducing the amount of labeled data required to train the model and reducing the annotation cost accor-dingly.The ALDL-MMD algorithm trains a linear regression model.While ensuring the minimum training error of the linear regression model,it learns a sparse vector that reflects that which instance in the unlabeled data set are selected,so that the data distribution of the training data set and unlabeled data set after instance selection is as similar as possible.We relax the vector for easy calculation.An effective method to optimize the objective function in ALDL-MMD is given,and proof for the convergence of ALDL-MMD is also provided.The experimental results on multiple label distribution data sets show that the ALDL-MMD algorithm is superior to the existing active example selection methods on the two evaluation measures of "Canberra Metric" (distance) and “Intersection” (similarity) to measure that what degree of the label distribution of the instance is accurate,which reflects its effectiveness in reducing annotation costs.

Key words: Active learning, Label distribution learning, Linear model, Marginal probability distribution matching, Maximum mean discrepancy

中图分类号: 

  • TP391
[1] ZHANG M,ZHOU Z.A Review on Multi-Label Learning Algorithms[J].IEEE Transactions on Knowledge & Data Enginee-ring,2014,26(8):1819-1837.
[2] GENG X,JI R.Label Distribution Learning[C]//IEEE International Conference on Data Mining Workshops.IEEE Computer Society,2013.
[3] HE Z,LI X,ZHANG Z.Data-Dependent Label DistributionLearning for Age Estimation[J].IEEE Trans. Image Process.,2017,26(8):3846-3858.
[4] KONG S,OYINI MBOUNA R.Head Pose Estima-tion from a 2-D Face Image using 3-D Face Morphing with Depth Parameters [J].IEEE Transactions on Image Processing,2015,24(6):1-1.
[5] ZHANG Z L,LAI C H,LIU H,et al.Infrared Facial Expression Recognition via Gaussian-based Label Distribution Learning in the Dark Illumination Environment[J].Neurocomputing,2020,409:341-350.
[6] ZHOU D Y,ZHANG X,ZHOU Y,et al.Emotion Distribution Learning from Texts[C]//Conference on Empirical Methods in Natural Language Processing.2016.
[7] SEUNG H S.Query by Committee[C]//Workshop on Computational Learning Theory.ACM,1992.
[8] FREUND Y,SEUNG H S,SHAMIR E,et al.Selective Sampling Using the Query by Committee Algorithm[J].Machine Learning,1997,28(2/3):133-168.
[9] GU S,CAI Y,SHAN J,et al.Active Learning with Error-Correcting Output Codes[J].Neurocomputing,2019,364:182-191.
[10] LINDLEY D V.On a Measure of the Information Provided by an Experiment[J].Annals of Mathematical Statistics,1956,27(4):986-1005.
[11] YU K,BI J B,TRESP V,et al.Active learning via transductive experimental design[C]//International Conference on Machine Learning.2006:1081-1088.
[12] PATRA S,BRUZZONE L.A cluster-assumption based batchmode active learning technique[J].Pattern Recognition Letters,2012,33(9):1042-1048.
[13] BURGES C J.A Tutorial on Support Vector Machines for Pattern Recognition[J].Data Mining and Knowledge Discovery,1998,2(2):121-167.
[14] MCCALLUM A,NIGAM K.A comparison of event models for naive bayes text classification[C]//National Conference on Artificial Intelligence.1998:41-48.
[15] PIETRA S D,PIETRA V D,LAFFERTY J,et al.Inducing features of random fields[J].IEEE Transactions on Pattern Analysis and Machine Intelligence,1997,19(4):380-393.
[16] NOCEDAL J,WRIGHT S.Numerical optimization[M].Springer Science & Business Media,2006:61-63.
[17] PRINCE M.Does Active Learning Work? A Review of the Research[J].Journal of Engineering Education,2004,93(3):223-231.
[18] CAI D,HE X.Manifold Adaptive Experimental Design for Text Categorization[J].IEEE Transactions on Knowledge and Data Engineering,2011,24(4):707-719.
[19] SINDHWANI V,NIYOGI P,BELKIN M,et al.Beyond thepoint cloud:from transductive to semi-supervised learning[C]//International Conference on Machine Learning.2005:824-831.
[20] PAN S J,TSANG I W,KWOK J T,et al.Domain Adaptation via Transfer Component Analysis[J].IEEE Transactions on Neural Networks,2011,22(2):199-210.
[21] GRANT M,BOYD S.CVX:Matlab software for disciplined convex programming[J].International Journal of Communications,Network and System Science,2008,1(1).
[22] EISEN M B,SPELLMAN P T,BROWN P O,et al.Clusteranalysis and display of genome-wide expression patterns[J].Proceedings of the National Academy of Sciences of the United States of America,1998,95(25):14863-14868.
[23] CHA S H.Comprehensive Survey on Distance/ Similarity Mea-sures Between Probability Density Functions[J].International Journal of Mathematical Models and Methods in Applied Sciences,2007,1(4):300-307.
[1] 周慧, 施皓晨, 屠要峰, 黄圣君.
基于主动采样的深度鲁棒神经网络学习
Robust Deep Neural Network Learning Based on Active Sampling
计算机科学, 2022, 49(7): 164-169. https://doi.org/10.11896/jsjkx.210600044
[2] 侯夏晔, 陈海燕, 张兵, 袁立罡, 贾亦真.
一种基于支持向量机的主动度量学习算法
Active Metric Learning Based on Support Vector Machines
计算机科学, 2022, 49(6A): 113-118. https://doi.org/10.11896/jsjkx.210500034
[3] 张人之, 朱焱.
基于主动学习的社交网络恶意用户检测方法
Malicious User Detection Method for Social Network Based on Active Learning
计算机科学, 2021, 48(6): 332-337. https://doi.org/10.11896/jsjkx.200700151
[4] 王学光, 张爱新, 窦炳琳.
复杂网络上的非线性负载容量模型
Non-linear Load Capacity Model of Complex Networks
计算机科学, 2021, 48(6): 282-287. https://doi.org/10.11896/jsjkx.200700040
[5] 王体爽, 李培峰, 朱巧明.
基于数据增强的中文隐式篇章关系识别方法
Chinese Implicit Discourse Relation Recognition Based on Data Augmentation
计算机科学, 2021, 48(10): 85-90. https://doi.org/10.11896/jsjkx.200800115
[6] 姚成亮, 朱庆生.
基于自然邻居的标记分布学习
Label Distribution Learning Based on Natural Neighbors
计算机科学, 2020, 47(8): 132-136. https://doi.org/10.11896/jsjkx.190700012
[7] 李翼宏, 刘方正, 杜镇宇.
一种改进主动学习的恶意代码检测算法
Malware Detection Algorithm for Improving Active Learning
计算机科学, 2019, 46(5): 92-99. https://doi.org/10.11896/j.issn.1002-137X.2019.05.014
[8] 赵海燕, 汪静, 陈庆奎, 曹健.
主动学习在推荐系统中的应用
Application of Active Learning in Recommendation System
计算机科学, 2019, 46(11A): 153-158.
[9] 孙金, 陈若煜, 罗恒利.
基于主动学习的人脸标注研究
Research on Face Tagging Based on Active Learning
计算机科学, 2018, 45(9): 299-302. https://doi.org/10.11896/j.issn.1002-137X.2018.09.050
[10] 吕巨建, 赵慧民, 陈荣军, 李键红.
基于自适应稀疏邻域重构的无监督主动学习算法
Unsupervised Active Learning Based on Adaptive Sparse Neighbors Reconstruction
计算机科学, 2018, 45(6): 251-258. https://doi.org/10.11896/j.issn.1002-137X.2018.06.045
[11] 李昌利, 张琳, 樊棠怀.
基于自适应主动学习与联合双边滤波的高光谱图像分类
Hyperspectral Image Classification Based on Adaptive Active Learning and Joint Bilateral Filtering
计算机科学, 2018, 45(12): 223-228. https://doi.org/10.11896/j.issn.1002-137X.2018.12.037
[12] 汤颖, 孙康高, 秦绪佳, 周建美.
基于局部模型加权融合的Top-N电影推荐算法
Local Model Weighted Ensemble for Top-N Movie Recommendation
计算机科学, 2018, 45(11A): 439-444.
[13] 李锋,万小强.
基于关联矩阵的短信自动分类
SMS Automatic Classification Based on Relational Matrix
计算机科学, 2017, 44(Z6): 428-432. https://doi.org/10.11896/j.issn.1002-137X.2017.6A.096
[14] 王长宝,李青雯,于化龙.
面向类别不平衡数据的主动在线加权极限学习机算法
Active,Online and Weighted Extreme Learning Machine Algorithm for Class Imbalance Data
计算机科学, 2017, 44(12): 221-226. https://doi.org/10.11896/j.issn.1002-137X.2017.12.040
[15] 翟俊海,臧立光,张素芳.
在线序列主动学习方法
Online Sequential Active Learning Approach
计算机科学, 2017, 44(1): 37-41. https://doi.org/10.11896/j.issn.1002-137X.2017.01.007
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!