计算机科学 ›› 2019, Vol. 46 ›› Issue (11A): 246-250.

• 模式识别与图像处理 • 上一篇    下一篇

基于深度学习的非实验室场景人脸属性识别

葛宏孔, 罗恒利, 董佳媛   

  1. (南京航空航天大学计算机科学与技术学院 南京211106)
  • 出版日期:2019-11-10 发布日期:2019-11-20
  • 通讯作者: 葛宏孔(1994-),男,硕士生,主要研究方向为人脸属性识别,E-mail:18549830154@163.com。
  • 作者简介:罗恒利(1994-),男,硕士生,主要研究方向为人脸聚类;董佳媛(1995-),女,硕士生,主要研究方向为图像分类。
  • 基金资助:
    本文受国家自然科学基金(61772268)资助。

Face Attributes in Wild Based on Deep Learning

GE Hong-kong, LUO Heng-li, DONG Jia-yuan   

  1. (School of Computer Science and Technology,Nanjing University of Aeronautics and Astronautics,Nanjing 211106,China)
  • Online:2019-11-10 Published:2019-11-20

摘要: 非实验室场景下的人脸图片数量巨大,更加贴近生活,对其进行识别具有较大的研究价值。文中对非实验室环境下的人脸属性识别问题进行了研究,提出了一种人脸属性识别网络(Regional Multiple Layer Attributes Related Net,RMLARNet),不仅对人脸特征的提取方式进行了研究,还挖掘了人脸属性间的关系。该网络由3个部分组成:1)将人脸图像分割成包含属性部位的多个局部区域,并将这些局部区域作为输入提取特征信息;2)以Inception V3 为迁移模型,采取多个不相邻卷积层迁移方式提取人脸特征;3)搭建了一个以人脸属性关系为约束的属性识别网络。实验结果表明,对CelebA数据集进行筛选处理,创建属性样本较平衡的CelebA-数据集,并在该数据集上设计实验将取得优于现有方法的实验效果。

关键词: 多标签任务, 迁移学习, 人脸属性识别, 深度学习, 属性约束

Abstract: Faces in the wild are huge in number and more close to life,and the recognition of facial attributes is a valuable research.A face attributes recognition method named RMLARNet (Regional Multiple Layer Attributes Related Net) was proposed for faces in the wild,which explores a new feature extraction method and attributes relationship.The processing steps of this method are as follows:1)Feature extraction is based on the regional parts of image.2)Features are extracted from different layer of Inception V3,and they are concatenated to get the final face feature.3)An attributes relationship related network is used for attributes recognition.The experiment is conducted on a balanced CelebA- data set which is a subset of CelebA,and this method outperforms state-of-the art methods.

Key words: Attributes constraint, Deep learning, Face attributes recognition, Multi-label task, Transfer learning

中图分类号: 

  • TP391.4
[1]KUMAR N,BERG A C,BELHUMEUR P N,et al.Attribute and simile classifiers for face verification[C]∥IEEE International Conference on Computer Vision.IEEE,2009:365-372.
[2]SONG F,TAN X,CHEN S.Exploiting relationship between attributes for improved face verification [J].Computer Vision & Image Understanding,2014,122(4):143-154.
[3]BOURDEV L,MAJI S,MALIK J.Describing people:A poselet-based approach to attribute classification[C]∥IEEE International Conference on Computer Vision.IEEE,2011:1543-1550.
[4]ZHANG N,PALURI M,RANZATO M,et al.PANDA:Pose Aligned Networks for Deep Attribute Modeling[C]∥IEEE Conference on Computer Vision and Pattern Recognition.IEEE,2014:1637-1644.
[5]LiU Z,LUO P,WANG X,et al.Deep Learning Face Attributes in the Wild[C]∥IEEE International Conference on Computer Vision.IEEE,2015:3730-3738.
[6]DONAHUE J,JIA Y,VINYALS O,et al.DeCAF:a deep convolutional activation feature for generic visual recognition[C]∥International Conference on International Conference on Machine Learning.JMLR.org,2014:I-647.
[7]SRINIVAS N,ATWAL H,ROSE D C,et al.Age,Gender,and Fine-Grained Ethnicity Prediction Using Convolutional Neural Networks for the East Asian Face Dataset[C]∥IEEE International Conference on Automatic Face & Gesture Recognition.IEEE,2017:953-960.
[8]ZENG T,JI S.Deep Convolutional Neural Networks for Multi-instance Multi-task Learning[C]∥IEEE International Conference on Data Mining.IEEE,2016:579-588.
[9]DONG M,PANG K,WU Y,et al.Transferring CNNS to multi-instance multi-label classification on small datasets[C]∥IEEE International Conference on Image Processing.Beijing:IEEE,2017:1332-1336.
[10]GHOSH S,LAKSANA E,SCHERER S,et al.A multi-labelconvolutional neural network approach to cross-domain action unit detection[C]∥International Conference on Affective Computing and Intelligent Interaction.IEEE,2015:609-615.
[11]SZEGEDY C,VANHOUCKE V,IOFFE S,et al.Rethinking the Inception Architecture for Computer Vision[C]∥Computer Vision and Pattern Recognition.IEEE,2016:2818-2826.
[12]SIMONYAN K,ZISSERMAN A.Very deep convolutional networks for large-scale image recognition[J].arXiv:1409.1556,2014.
[13]ZOU W Y,ZHU S,NG A Y,et al.Deep learning of invariant features via simulated fixations in video[C]∥International Conference on Neural Information Processing Systems.Curran Associates Inc,2012:3203-3211.
[14]COATES A,NG A Y.Selecting receptive fields in deep net-works[C]∥International Conference on Neural Information Processing Systems.Curran Associates Inc,2011:2528-2536.
[15]ZHANG M L,ZHANG K.Multi-label learning by exploiting label dependency[C]∥ACM SIGKDD International Conference on Knowledge Discovery and Data Mining.ACM,2010:999-1008.
[16]HUANG S J,ZHOU Z H.Multi-label learning by exploiting label correlations locally[C]∥Twenty-Sixth AAAI Conference on Artificial Intelligence.AAAI Press,2012:949-955.
[17]KUMAR N,BELHUMEUR P,NAYAR S.FaceTracer:ASearch Engine for Large Collections of Images with Faces[C]∥Europican Conference on Computer Vision.Springer-Verlag,2008:340-353.
[18]LI J,ZHANG Y.Learning SURF Cascade for Fast and Accurate Object Detection[C]∥Computer Vision and Pattern Recognition.IEEE,2013:3468-3475.
[19]SUN Y,WANG X,TANG X.Deep Convolutional Network Cascade for Facial Point Detection[C]∥IEEE Conference on Computer Vision and Pattern Recognition.IEEE Computer Society,2013:3476-3483.
[1] 徐涌鑫, 赵俊峰, 王亚沙, 谢冰, 杨恺.
时序知识图谱表示学习
Temporal Knowledge Graph Representation Learning
计算机科学, 2022, 49(9): 162-171. https://doi.org/10.11896/jsjkx.220500204
[2] 饶志双, 贾真, 张凡, 李天瑞.
基于Key-Value关联记忆网络的知识图谱问答方法
Key-Value Relational Memory Networks for Question Answering over Knowledge Graph
计算机科学, 2022, 49(9): 202-207. https://doi.org/10.11896/jsjkx.220300277
[3] 汤凌韬, 王迪, 张鲁飞, 刘盛云.
基于安全多方计算和差分隐私的联邦学习方案
Federated Learning Scheme Based on Secure Multi-party Computation and Differential Privacy
计算机科学, 2022, 49(9): 297-305. https://doi.org/10.11896/jsjkx.210800108
[4] 方义秋, 张震坤, 葛君伟.
基于自注意力机制和迁移学习的跨领域推荐算法
Cross-domain Recommendation Algorithm Based on Self-attention Mechanism and Transfer Learning
计算机科学, 2022, 49(8): 70-77. https://doi.org/10.11896/jsjkx.210600011
[5] 孙奇, 吉根林, 张杰.
基于非局部注意力生成对抗网络的视频异常事件检测方法
Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection
计算机科学, 2022, 49(8): 172-177. https://doi.org/10.11896/jsjkx.210600061
[6] 王剑, 彭雨琦, 赵宇斐, 杨健.
基于深度学习的社交网络舆情信息抽取方法综述
Survey of Social Network Public Opinion Information Extraction Based on Deep Learning
计算机科学, 2022, 49(8): 279-293. https://doi.org/10.11896/jsjkx.220300099
[7] 郝志荣, 陈龙, 黄嘉成.
面向文本分类的类别区分式通用对抗攻击方法
Class Discriminative Universal Adversarial Attack for Text Classification
计算机科学, 2022, 49(8): 323-329. https://doi.org/10.11896/jsjkx.220200077
[8] 姜梦函, 李邵梅, 郑洪浩, 张建朋.
基于改进位置编码的谣言检测模型
Rumor Detection Model Based on Improved Position Embedding
计算机科学, 2022, 49(8): 330-335. https://doi.org/10.11896/jsjkx.210600046
[9] 侯钰涛, 阿布都克力木·阿布力孜, 哈里旦木·阿布都克里木.
中文预训练模型研究进展
Advances in Chinese Pre-training Models
计算机科学, 2022, 49(7): 148-163. https://doi.org/10.11896/jsjkx.211200018
[10] 周慧, 施皓晨, 屠要峰, 黄圣君.
基于主动采样的深度鲁棒神经网络学习
Robust Deep Neural Network Learning Based on Active Sampling
计算机科学, 2022, 49(7): 164-169. https://doi.org/10.11896/jsjkx.210600044
[11] 苏丹宁, 曹桂涛, 王燕楠, 王宏, 任赫.
小样本雷达辐射源识别的深度学习方法综述
Survey of Deep Learning for Radar Emitter Identification Based on Small Sample
计算机科学, 2022, 49(7): 226-235. https://doi.org/10.11896/jsjkx.210600138
[12] 胡艳羽, 赵龙, 董祥军.
一种用于癌症分类的两阶段深度特征选择提取算法
Two-stage Deep Feature Selection Extraction Algorithm for Cancer Classification
计算机科学, 2022, 49(7): 73-78. https://doi.org/10.11896/jsjkx.210500092
[13] 程成, 降爱莲.
基于多路径特征提取的实时语义分割方法
Real-time Semantic Segmentation Method Based on Multi-path Feature Extraction
计算机科学, 2022, 49(7): 120-126. https://doi.org/10.11896/jsjkx.210500157
[14] 王君锋, 刘凡, 杨赛, 吕坦悦, 陈峙宇, 许峰.
基于多源迁移学习的大坝裂缝检测
Dam Crack Detection Based on Multi-source Transfer Learning
计算机科学, 2022, 49(6A): 319-324. https://doi.org/10.11896/jsjkx.210500124
[15] 楚玉春, 龚航, 王学芳, 刘培顺.
基于YOLOv4的目标检测知识蒸馏算法研究
Study on Knowledge Distillation of Target Detection Algorithm Based on YOLOv4
计算机科学, 2022, 49(6A): 337-344. https://doi.org/10.11896/jsjkx.210600204
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!