Computer Science ›› 2021, Vol. 48 ›› Issue (4): 151-156.doi: 10.11896/jsjkx.200500049

Special Issue: Medical Imaging

• Computer Graphics & Multimedia • Previous Articles     Next Articles

Transverse Section Recognition Algorithm Based on BCNN for Fetal Craniocerebral Ultrasound

SHU Xin1, CHANG Feng1, ZHANG Xin2, DU Rui2, YU Zhuan2   

  1. 1 School of Computer Science,Jiangsu University of Science and Technology,Zhenjiang,Jiangsu 212003,China
    2 Department of Medical Ultrasound,Affiliated Hospital of Jiangsu University,Zhenjiang,Jiangsu 212003,China
  • Received:2020-06-24 Revised:2020-09-18 Online:2021-04-15 Published:2021-04-09
  • About author:SHU Xin,born in 1979.Ph.D,associate professor,is a member of China Computer Federation.His main research interests include medical image analysis,face recognition and pattern recognition.
  • Supported by:
    National Natural Science Foundation of China(61876072),Maternal and Child Health Research Project of Jiangsu Province(F201822) and Key Research and Development Project of Zhenjiang(Social Development)(SH2019038).

Abstract: Ultrasound examination during pregnancy is an important step to discover whether the fetus is abnormal,so it is of great value to carry out accurate and efficient examination and diagnosis on fetus.In this paper,bilinear convolutional neural network(BCNN) is used to identify the transverse section of the fetal head.On this basis,we propose BCNN-R and BCNN-S.The BCNN model takes the fetal craniocerebral ultrasound image as input,and firstly preprocesses the input data.Secondly,the two parallel sub-networks can extract the cross-sectional features with high identification and strong robustness from the input,after that the model integrates the extracted features,which is helpful to extract the fine features for recognition.Finally,the linear layer outputs the result of classification.In order to verify the effectiveness of the proposed algorithm,this paper makes contrast experiment on self-built fetal ultrasound dataset JTU19.The experimental results show that the proposed algorithms have ob-vious improvement on classification performance compared with the basic network(GoogleNet,DenseNet,SeNet,etc.),the overall accuracy of BCNN-S reaches 88.95%,and the precision and recall of BCNN-R in horizontal cross-sections achieves 97.22% and 88.61%.In addition,we also use the public dataset HC18 to conduct classification experiments.The accuracy,precision and recall of BCNN reach 89.48%,87.66% and 87.71% respectively,which further verifies the effectiveness of the proposed algorithms.

Key words: BCNN, Craniocerebrum image, Deep learning, Fetal ultrasound, Identification of transverse section

CLC Number: 

  • TP391
[1]VAN DEN HEUVEL T L A,PETROS H,SANTINI S,et al.Automated fetal head detection and circumference estimation from free-hand ultrasound sweeps using deep learning in resource-limited countries[J].Ultrasound in Medicine & Biology,2019,45(3):773-785.
[2]YAQUB M,KELLY B,PAPAGEORGHIOU A T,et al.A deep learning solution for automatic fetal neurosonographic diagnostic plane verification using clinical standard constraints[J].Ultrasound in Medicine & Biology,2017,43(12):2925-2933.
[3]MURTY M N,DEVI V S.Introduction to pattern recognitionand machine learning[M].Co-Published with Indian Institute of Science(IISc),Bangalore,India,2015.
[4]SHAN J,ALAM S K,GARRA B,et al.Computer-aided diagnosis for breast ultrasound using computerized BI-RADS features and machine learning methods[J].Ultrasound in Medicine & Bio-logy,2016,42(4):980-988.
[5]LEI B,ZHUO L,CHEN S,et al.Automatic recognition of fetal standard plane in ultrasound image[C]//2014 IEEE 11th International Symposium on Biomedical Imaging(ISBI).IEEE,2014:85-88.
[6]OJALA T,PIETIKAINEN M,MAENPAA T.Multiresolution gray-scale and rotation invariant texture classification with local binary patterns[J].IEEE Transactions on Pattern Analysis & Machine Intelligence,2002,24(7):971-987.
[7]SONG L,LIU X,MA L,et al.Using HOG-LBP features and MMP learning to recognize imaging signs of lung lesions[C]//2012 25th IEEE International Symposium on Computer-Based Medical Systems(CBMS).IEEE,2012:1-4.
[8]LEI B,TAN E L,CHEN S,et al.Automatic recognition of fetal facial standard plane in ultrasound image via fisher vector[J].PloS one,2015,10(5):e0121838.
[9]LEI B,YAO Y,CHEN S,et al.Discriminative learning for automatic staging of placental maturity via multi-layer fisher vector[J].Scientific Reports,2015,5(1):1-11.
[10]MA Y,WU L,LI Z.A novel face presentation attack detection scheme based on multi-regional convolutional neural networks[J].Pattern Recognition Letters,2020,131:261-267.
[11]CHEN H,NI D,QIN J,et al.Standard Plane Localization in Fetal Ultrasound via Domain Transferred Deep Neural Networks[J].IEEE Journal of Biomedical and Health Informatics,2015,19(5):1627-1636.
[12]YU Z,WU L Y,NI D,et al.Automatic Recognition of Fetal Facial Ultrasound Standard Section based on Deep Learning[J].Chinese Journal of Biomedical Engineering,2017(3):267-275.
[13]LIN T Y,ROYCHOWDHURY A,MAJI S.Bilinear cnn models for fine-grained visual recognition[C]//Proceedings of the IEEE International Conference on Computer Vision.2015:1449-1457.
[14]YOSINSKI J,CLUNE J,BENGIO Y,et al.How transferableare features in deep neural networks?[C]//Advances in Neural Information Processing Systems.2014:3320-3328.
[15]MALLAT S.Understanding deep convolutional networks[J].Philosophical Transactions of the Royal Society A:Mathematical,Physical and Engineering Sciences,2016,374(2065):20150203.
[16]HE K,ZHANG X,REN S,et al.Delving deep into rectifiers:Surpassing human-level performance on imagenet classification[C]//Proceedings of the IEEE International Conference on Computer Vision.2015:1026-1034.
[17]VAN DEN HEUVEL T L A,PETROS H,SANTINI S,et al.Automated fetal head detection and circumference estimation from free-hand ultrasound sweeps using deep learning in resource-limited countries[J].Ultrasound in Medicine & Biology,2019,45(3):773-785.
[18]HE K,ZHANG X,REN S,et al.Deep residual learning for ima-ge recognition[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2016:770-778.
[19]MAAS A L,HANNUN A Y,NG A Y.Rectifier nonlinearities improve neural network acoustic models[C]//Proc.ICML.2013.
[20]IANDOLA F N,HAN S,MOSKEWICZ M W,et al.Squeeze-Net:AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size[J].arXiv:1602.07360,2016.
[21]SZEGEDY C,LIU W,JIA Y,et al.Going deeper with convolutions[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2015:1-9.
[22]VAN D H T L A,DAGMAR D B,et al.Automated measurement of fetal head circumference using 2D ultrasound images[J].PloS One,13.8(2018):e0200412.
[23]VAN D H T L A,DAGMAR D B,et al.Automated measurement of fetal head circumference using 2D ultrasound images [DB/OL].Zenodo.http://doi.org/10.5281/zenodo.1322001.
[24]SZEGEDY C,LIU W,JIA Y,et al.Going deeper with convolutions[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2015:1-9.
[25]HU J,SHEN L,SUN G.Squeeze-and-excitation networks[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2018:7132-7141.
[26]HUANG G,LIU Z,VAN DER MAATEN L,et al.Densely connected convolutional networks[C]//Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2017:4700-4708.
[27]KINGMA D P,BA J.Adam:A method for stochastic optimization[J].arXiv:1412.6980,2014.
[28]HE K,GIRSHICK R,DOLLÁR P.Rethinking imagenet pre-training[C]//Proceedings of the IEEE International Conference on Computer Vision.2019:4918-4927.
[1] XU Yong-xin, ZHAO Jun-feng, WANG Ya-sha, XIE Bing, YANG Kai. Temporal Knowledge Graph Representation Learning [J]. Computer Science, 2022, 49(9): 162-171.
[2] RAO Zhi-shuang, JIA Zhen, ZHANG Fan, LI Tian-rui. Key-Value Relational Memory Networks for Question Answering over Knowledge Graph [J]. Computer Science, 2022, 49(9): 202-207.
[3] TANG Ling-tao, WANG Di, ZHANG Lu-fei, LIU Sheng-yun. Federated Learning Scheme Based on Secure Multi-party Computation and Differential Privacy [J]. Computer Science, 2022, 49(9): 297-305.
[4] WANG Jian, PENG Yu-qi, ZHAO Yu-fei, YANG Jian. Survey of Social Network Public Opinion Information Extraction Based on Deep Learning [J]. Computer Science, 2022, 49(8): 279-293.
[5] HAO Zhi-rong, CHEN Long, HUANG Jia-cheng. Class Discriminative Universal Adversarial Attack for Text Classification [J]. Computer Science, 2022, 49(8): 323-329.
[6] JIANG Meng-han, LI Shao-mei, ZHENG Hong-hao, ZHANG Jian-peng. Rumor Detection Model Based on Improved Position Embedding [J]. Computer Science, 2022, 49(8): 330-335.
[7] SUN Qi, JI Gen-lin, ZHANG Jie. Non-local Attention Based Generative Adversarial Network for Video Abnormal Event Detection [J]. Computer Science, 2022, 49(8): 172-177.
[8] HU Yan-yu, ZHAO Long, DONG Xiang-jun. Two-stage Deep Feature Selection Extraction Algorithm for Cancer Classification [J]. Computer Science, 2022, 49(7): 73-78.
[9] CHENG Cheng, JIANG Ai-lian. Real-time Semantic Segmentation Method Based on Multi-path Feature Extraction [J]. Computer Science, 2022, 49(7): 120-126.
[10] HOU Yu-tao, ABULIZI Abudukelimu, ABUDUKELIMU Halidanmu. Advances in Chinese Pre-training Models [J]. Computer Science, 2022, 49(7): 148-163.
[11] ZHOU Hui, SHI Hao-chen, TU Yao-feng, HUANG Sheng-jun. Robust Deep Neural Network Learning Based on Active Sampling [J]. Computer Science, 2022, 49(7): 164-169.
[12] SU Dan-ning, CAO Gui-tao, WANG Yan-nan, WANG Hong, REN He. Survey of Deep Learning for Radar Emitter Identification Based on Small Sample [J]. Computer Science, 2022, 49(7): 226-235.
[13] ZHU Wen-tao, LAN Xian-chao, LUO Huan-lin, YUE Bing, WANG Yang. Remote Sensing Aircraft Target Detection Based on Improved Faster R-CNN [J]. Computer Science, 2022, 49(6A): 378-383.
[14] WANG Jian-ming, CHEN Xiang-yu, YANG Zi-zhong, SHI Chen-yang, ZHANG Yu-hang, QIAN Zheng-kun. Influence of Different Data Augmentation Methods on Model Recognition Accuracy [J]. Computer Science, 2022, 49(6A): 418-423.
[15] MAO Dian-hui, HUANG Hui-yu, ZHAO Shuang. Study on Automatic Synthetic News Detection Method Complying with Regulatory Compliance [J]. Computer Science, 2022, 49(6A): 523-530.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!