Computer Science ›› 2022, Vol. 49 ›› Issue (10): 214-223.doi: 10.11896/jsjkx.210900080

• Computer Graphics& Multimedia • Previous Articles     Next Articles

Face Image Synthesis Driven by Geometric Feature and Attribute Label

DAI Fu-yun, CHI Jing, REN Ming-guo, ZHANG Qi-dong   

  1. School of Computer Science and Technology,Shandong University of Finance and Economics,Jinan 250014,China
    Shandong Provincial Key Laboratory of Digital Media Technology,Jinan 250014,China
  • Received:2021-09-10 Revised:2022-03-10 Online:2022-10-15 Published:2022-10-13
  • About author:DAI Fu-yun,born in 1996,postgra-duate.Her main research interests include face image synthesis and so on.
    CHI Jing,born in 1980,Ph.D,professor,Ph.D supervisor.Her main research interests include computer animation,digital image processing and information visualization.
  • Supported by:
    Key Research and Development Program of Shandong Province(2019GSF109112),Natural Science Foundation of Shandong Province(ZR2019MF016),Science and Technology Plan for Young Talents in Colleges and Universities of Shandong Province(2020KJN007) and Scientific Research Studio in Colleges and Universities of Jinan City(2021GXRC092).

Abstract: Aiming at the problems in current face image synthesis,such as the lack of diversity of synthetic appearances and expressions,the low reality of the facial expressions and the low synthesis efficiency,this paper proposes a novel face synthesis network model driven by facial geometric feature and attribute label.Given a source face image,a target face image and the attribute(e.g.,hair color,gender,age) label,the new face synthesis model can generate a highly realistic face image which owns the expression of the source face,the identity of the target face and the specified attribute.The new model consists of two parts:facial landmark generator(FLMG) and geometry and attribute aware generator(GAAG).FLMG uses the facial geometric feature points to encode the expression information,and transfers the expression from the source to the target face in the form of feature points.Combining the transferred feature points,the specified attribute label and the target face image,GAAG generates a face image with specified appearance and expression.A novel soft margin triplet perception loss is introduced to GAAG,which can make the synthesized face more natural and keep the identity of the target face well,and makes the GAAG converge faster.Experimental results show that the face images generated by our approach have more diverse appearances and more realistic expressions.In addition,our model only needs to be trained once to realize the transfer between any arbitrary different expressions,so its efficiency is high.

Key words: Face image synthesis, Expression transfer, Face editing, Generative adversarial network

CLC Number: 

  • TP317
[1]SUSSKIND J M,HINTON G E,MOVELLAN J R,et al.Gene-rating facial expressions with deep belief nets[M]//Affective Computing,Emotion Modelling,Synthesis and Recognition.2008:421-440.
[2]YAN Y F,LU K,XUE J,et al.Face animation method based on deep learning and expression AU parameters [J].Journal of Computer Aided Design & Graphics,2019,31(11):1973-1980.
[3]GOODFELLOW I J,POUGET-ABADIE J,MIRZA M,et al.Generative adversarial nets[C]//Neural Information Processing Systems.2014:2672-2680.
[4]ARJOVSKY M,CHINTALA S,BOTTOU L.Wasserstein GAN[J].arXiv:1701.07875,2017.
[5]ISOLA P,ZHU J Y,ZHOU T H,et al.Image-to-image translation with conditional adversarial networks[C]//Computer Vison and Pattern Recognition.2017:5967-5976.
[6]YI Z L,ZHANG H,TAN P,et al.DualGAN:unsupervised dual learning for image-to-image translation[C]//IEEE International Conference on Computer Vision.2017:2868-2876.
[7]HUANG H B.Face image synthesis and analysis based on gene-rative model[D].Beijing:University of Chinese Academy of Sciences,2019.
[8]CHEN F J,ZHU F,WU Q X,et al.A Survey of Generative Adversarial Networks and Its Application in Image Generation[J].Chinese Journal of Computers,2021,44(2):347-369.
[9]PERARNAU G,VAN J,RADUCANU B,et al.Invertible conditional GANs for image editing[C]//NIPS Workshop on Adversarial Training.2016.
[10]DING H,SRICHARAN K,CHELLAPPA R.ExprGAN:Facial expression editing with controllable expression intensity[C]//AAAI Conference on Artificial Intelligence.2018:6665-7655.
[11]KARRAS T,LAINE S,AILA T.A Style-Based Generator Architecture for Generative Adversarial Networks[C]//Confe-rence on Computer Vision and Pattern Recognition.2019:4401-4410.
[12]TEWARI A,ELGHARIB M,BHARAJ G.StyleRig:RiggingStyleGAN for 3D Control over Portrait Images[C]//Confe-rence on Computer Vision and Pattern Recognition.2020:6141-6150.
[13]SONG L X,LU Z H,HE R,et al.Geometry Guided Adversarial Facial Expression Synthesis[C]//ACM International Confe-rence on Multimedia.2018:627-635.
[14]QIAO F C,YAO N M,JIAO Z R,et al.Emotional facial expression transfer from a single image via generative adversarial nets[J].Computer Animation and Virtual Worlds,2018,29(3/4):1/10-10/10.
[15]OLSZEWSKI Z,LI Z M,YANG C,et al.Realistic Dynamic Facial Textures from a Single Image Using GANs[C]//IEEE International Conference on Computer Vision.2017:5429-5438.
[16]ZHU J Y,PARK T,ISOLA P,et al.Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks[C]//IEEE International Conference on Computer Vision.2017:2232-2969.
[17]CHOI Y J,CHOI M J,KIM M Y,et al.StarGAN:Unified Gene-rative Adversarial Networks for Multi-Domain Image-to-Image Translation[C]//IEEE Conference on Computer Vision and Pattern Recognition.2018:8789-8797.
[18]ZHANG J N,ZENG X F,WANG M M,et al.FReeNet:Multi-Identity Face Reenactment[C]//Computer Vison and Pattern Recognition.2020:5326-5335.
[19]WANG X P,WANG Y H,LI W X.U-Net Conditional GANs for Photo-Realistic and Identity-Preserving Facial Expression Synthesis[J].ACM Transactions on Multimedia Computing,Communications,and Applications,2019,15(3s):1-23.
[20]SCHROFF F,KALENICHENKO D,PHILBIN J.FaceNet:AUnified Embedding for Face Recognition and Clustering[C]//Computer Vison and Pattern Recognition.2015:815-823.
[21]JUSTIN J,ALEXANDRE A,LI F F.Perceptual Losses forReal-Time Style Transfer and Super-Resolution[C]//European Conference on Computer Vision.2016:694-711.
[22]GUO X J,LI S Y,ZHANG J W,et al.PFLD:A Practical Facial Landmark Detector[J].arXiv:1902.10859,2019.
[23]ODENA A,OLAH C,SHLENS J.Conditional Image Synthesis with Auxiliary Classifier GANs[C]//The International Confe-rence on Machine Learning.2017:2642-2651.
[24]SIMONYAN K,ZISSERMAN A.Very Deep Convolutional Networks for Large-Scale Image Recognition[C]//International Conference on Learning Representations.2015.
[25]LANGNER O,DOTSCH R,BIJLSTRA G,et al.Presentation and validation of the Radboud Faces Database [J].Cognition and Emotion,2010,24(8):1377-1388.
[26]Karolinska Institute,Stockholm,Sweden.The Karolinska Di-rected Emotional Faces[EB/OL].https://www.kdef.se/index.html.
[27]DIEDERIK P,KINGMA D,JIMMY B.Adam:A Method for Stochastic Optimization[C]//International Conference on Learning Representations.2015.
[28]AMAZON.Amazon Mechanical Turk(AMT)[EB/OL].https://www.mturk.com.
[1] ZHANG Jia, DONG Shou-bin. Cross-domain Recommendation Based on Review Aspect-level User Preference Transfer [J]. Computer Science, 2022, 49(9): 41-47.
[2] SUN Qi, JI Gen-lin, ZHANG Jie. Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection [J]. Computer Science, 2022, 49(8): 172-177.
[3] DAI Zhao-xia, LI Jin-xin, ZHANG Xiang-dong, XU Xu, MEI Lin, ZHANG Liang. Super-resolution Reconstruction of MRI Based on DNGAN [J]. Computer Science, 2022, 49(7): 113-119.
[4] XU Guo-ning, CHEN Yi-peng, CHEN Yi-ming, CHEN Jin-yin, WEN Hao. Data Debiasing Method Based on Constrained Optimized Generative Adversarial Networks [J]. Computer Science, 2022, 49(6A): 184-190.
[5] YIN Wen-bing, GAO Ge, ZENG Bang, WANG Xiao, CHEN Yi. Speech Enhancement Based on Time-Frequency Domain GAN [J]. Computer Science, 2022, 49(6): 187-192.
[6] XU Hui, KANG Jin-meng, ZHANG Jia-wan. Digital Mural Inpainting Method Based on Feature Perception [J]. Computer Science, 2022, 49(6): 217-223.
[7] GAO Zhi-yu, WANG Tian-jing, WANG Yue, SHEN Hang, BAI Guang-wei. Traffic Prediction Method for 5G Network Based on Generative Adversarial Network [J]. Computer Science, 2022, 49(4): 321-328.
[8] DOU Zhi, WANG Ning, WANG Shi-jie, WANG Zhi-hui, LI Hao-jie. Sketch Colorization Method with Drawing Prior [J]. Computer Science, 2022, 49(4): 195-202.
[9] LI Si-quan, WAN Yong-jing, JIANG Cui-ling. Multiple Fundamental Frequency Estimation Algorithm Based on Generative Adversarial Networks for Image Removal [J]. Computer Science, 2022, 49(3): 179-184.
[10] SHI Da, LU Tian-liang, DU Yan-hui, ZHANG Jian-ling, BAO Yu-xuan. Generation Model of Gender-forged Face Image Based on Improved CycleGAN [J]. Computer Science, 2022, 49(2): 31-39.
[11] TANG Yu-xiao, WANG Bin-jun. Research Progress of Face Editing Based on Deep Generative Model [J]. Computer Science, 2022, 49(2): 51-61.
[12] LI Jian, GUO Yan-ming, YU Tian-yuan, WU Yu-lun, WANG Xiang-han, LAO Song-yang. Multi-target Category Adversarial Example Generating Algorithm Based on GAN [J]. Computer Science, 2022, 49(2): 83-91.
[13] TAN Xin-yue, HE Xiao-hai, WANG Zheng-yong, LUO Xiao-dong, QING Lin-bo. Text-to-Image Generation Technology Based on Transformer Cross Attention [J]. Computer Science, 2022, 49(2): 107-115.
[14] CHEN Gui-qiang, HE Jun. Study on Super-resolution Reconstruction Algorithm of Remote Sensing Images in Natural Scene [J]. Computer Science, 2022, 49(2): 116-122.
[15] HUANG Zhong-hao, YANG Xing-yao, YU Jiong, GUO Liang, LI Xiang. Mutual Learning Knowledge Distillation Based on Multi-stage Multi-generative Adversarial Network [J]. Computer Science, 2022, 49(10): 169-175.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!