Computer Science ›› 2025, Vol. 52 ›› Issue (1): 221-231.doi: 10.11896/jsjkx.240400108

• Computer Graphics & Multimedia • Previous Articles     Next Articles

Contact-free IR-UWB Human Motion Recognition Based on Dual-stream Fusion Network

ZHANG Chuanzong, WANG Dongzi, GUO Zhengxin, GUI Linqing, XIAO Fu   

  1. School of Computer Science,Nanjing University of Posts and Telecommunications,Nanjing 210023,China
    Jiangsu High Technology Research Key Laboratory for Wireless Sensor Networks,Nanjing 210023,China
  • Received:2024-04-15 Revised:2024-07-02 Online:2025-01-15 Published:2025-01-09
  • About author:ZHANG Chuanzong,born in 1997,postgraduate.His main research interests include IR-UWB and mobile computing.
    GUO Zhengxin,born in 1993,Ph.D,lecturer.His main research interests include wireless sensing,mobile computing,deep learning and Internet of Things.
  • Supported by:
    National Science Fund for Distinguished Young Scholars of China(62125203), National Natural Science Foundation of China(61932013),Key Research and Development Program of Jiangsu Province(BE2022798) and Natural Science Research Start-up Foundation of Recruiting Talents of Nanjing University of Posts and Telecommunications(NY223041).

Abstract: With the rapid development of intelligent sensing technology,the field of human computer interaction(HCI) has entered a new era.Traditional HCI methods,predominantly reliant on wearable devices and cameras to collect user behavior data,have significant limitations despite their precise recognition capabilities.Wearable devices,for instance,impose additional burden on users,whereas camera-based solutions are susceptible to ambient lighting conditions and pose significant privacy concerns.These challenges considerably restrict their applicability in daily life.To solve these challenges,we utilize the exceptional sensiti-vity and spatial resolution of impulse radio ultra-wideband(IR-UWB) in the field of radio frequency(RF) to propose a novel and contact-free method for human motion recognition based on a dual-stream fusion network.This method adeptly captures the temporal signal variations caused by target movements and extracts the corresponding frequency-domain features by analyzing Doppler frequency shift(DFS) changes on the time-domain signals.Subsequently,a sophisticated dual-stream network model,integrating multi-dimensional convolutional neural networks(CNNs) and GoogLeNet modules,is developed to facilitate precise action recognition.Through extensive experimental tests,the results show that the proposed method achieves an average accuracy of 94.89% for eight common daily human actions and maintains an accuracy of over 90% under varying test conditions,thereby va-lidating the robustness of the proposed method.

Key words: Human computer interaction, Wireless sensing, Impulse radio ultra-wideband(IR-UWB), Motion recognition

CLC Number: 

  • TP391
[1]YANG X,TIAN Y L.Effective 3d action recognition usingeigenjoints[J].Journal of Visual Communication and Image Representation,2014,25(1):2-11.
[2]XU Y,CHEN J,YANG Q,et al.Human posture recognition and fall detection using Kinect V2 camera[C]//2019 Chinese Control Conference(CCC).IEEE,2019:8488-8493.
[3]LI X,WANG Y,ZHANG B,et al.PSDRNN:An efficient and effective HAR scheme based on feature extraction and deep learning[J].IEEE Transactions on Industrial Informatics,2020,16(10):6703-6713.
[4]HE X,ZHU J,SU W,et al.RFID based non-contact human activity detection exploiting cross polarization[J].IEEE Access,2020,8:46585-46595.
[5]CASTILLO-CARA M,LOVÓN-MELGAREJO J,BRAVORO-CCA G,et al.An empirical study of the transmission power setting for bluetooth-based indoor localization mechanisms[J].Sensors,2017,17(6):1318.
[6]KJARGAARD M B,WIRZ M,ROGGEN D,et al.Mobile sen-sing of pedestrian flocks in indoor environments using wifi signals[C]//2012 IEEE International Conference on Pervasive Computing and Communications.IEEE,2012:95-102.
[7]WANG W,LIU A X,SHAHZAD M,et al.Device-free human activity recognition using commercial WiFi devices[J].IEEE Journal on Selected Areas in Communications,2017,35(5):1118-1131.
[8]YAN H,ZHANG Y,WANG Y,et al.WiAct:A passive WiFi-based human activity recognition system[J].IEEE Sensors Journal,2019,20(1):296-305.
[9]ZHOU Q,YANG Q,XING J.Enabling efficient WiFi-based occupant behavior recognition using insufficient samples[J].Buil-ding and Environment,2022,212:108806.
[10]XIAO C,LEI Y,LIU C,et al.Mean teacher-based cross-domain activity recognition using WiFi signals[J].IEEE Internet of Things Journal,2023,10(14):12787-12797.
[11]XIA Z,LIU J,GUO S.Rf-care:Rfid-based human pose estimation for nursing-care applications[C]//2021 IEEE International Conference on Robotics and Biomimetics(ROBIO).IEEE,2021:1384-1389.
[12]OGUNTALA G A,HU Y F,ALABDULLAH A A S,et al.Passive RFID module with LSTM recurrent neural network activity classification algorithm for ambient-assisted living[J].IEEE Internet of Things Journal,2021,8(13):10953-10962.
[13]LI L,BAI R,XIE B,et al.R&P:an low-cost device-free activity recognition for E-health[J].IEEE Access,2017,6:81-90.
[14] LI J,PHUNG S L,TIVIVE F H C,et al.Automatic classification of human motions using Doppler radar[C]//The 2012 International Joint Conference on Neural Networks(IJCNN).IEEE,2012:1-6.
[15]KIM Y,MOON T.Human detection and activity classification based on micro-Doppler signatures using deep convolutional neural networks[J].IEEE Geoscience and Remote Sensing Letters,2015,13(1):8-12.
[16]KIM Y,ALNUJAIM I,OH D.Human activity classificationbased on point clouds measured by millimeter wave MIMO radar with deep recurrent neural networks[J].IEEE Sensors Journal,2021,21(12):13522-13529.
[17]ABEDI H,ANSARIYAN A,MORITA P P,et al.AI-powered non-contact in-home gait monitoring and activity recognition system based on mm-wave FMCW radar and cloud computing[J].arXiv:2208.05905,2022.
[18]BRYAN J D,KWON J,LEE N,et al.Application of ultra-wide band radar for classification of human activities[J].IET Radar,Sonar & Navigation,2012,6(3):172-179.
[19]AHMED S,CHO S H.Hand gesture recognition using an IR-UWB radar with an inception module-based classifier[J].Sensors,2020,20(2):564.
[20]DING C,ZHANG L,GU C,et al.Non-contact human motionrecognition based on UWB radar[J].IEEE Journal on Emerging and Selected Topics in Circuits and Systems,2018,8(2):306-315.
[21]JIANG L B,LI C,CHE L.Using two-dimensional wavelet pac-ket decomposition for human action recognition ultra-wideband radar [J].Journal of electronic measurement and instrument,2018,32(8):69-75.
[23]JIANG J.Human posture recognition based on UWB and support Vector Machine [D].Beijing:Beijing University of Posts and Telecommunications,2015.
[24]AL-QIZWINI M,BARJASTEH I,AL-QASSAB H,et al.Deep learning algorithm for autonomous driving using googlenet[C]//2017 IEEE Intelligent Vehicles Symposium(IV).IEEE,2017:89-96.
[25]LEE J,ABDEL-ATY M,CHOI K,et al.Multi-level hot zoneidentification for pedestrian safety[J].Accident Analysis & Prevention,2015,76:64-73.
[26]MAO M,ZHANG R,ZHENG H,et al.Dual-stream network for visual recognition[J].Advances in Neural Information Proces-sing Systems,2021,34:25346-25358.
[27]KIM Y,TOOMAJIAN B.Hand gesture recognition using micro-Doppler signatures with convolutional neural network[J].IEEE Access,2016,4:7125-7130.
[28]SKARIA S,AL-HOURANI A.,LECH,M,et al.Hand-Gesture Recognition Using Two-Antenna Doppler Radar with Deep Convolutional Neural Networks[J].IEEE Sensors Journal,2019,19:3041-3048.
[1] ZHANG Jiahao, ZHANG Zhaohui, YAN Qi, WANG Pengwei. Speech Emotion Recognition Based on Voice Rhythm Differences [J]. Computer Science, 2024, 51(4): 262-269.
[2] LI Zhengping, LI Hanwen, WANG Lijun. High-generalization Ability EEG Emotion Recognition Model with Differential Entropy [J]. Computer Science, 2024, 51(11A): 231200066-7.
[3] ZHANG Xiaoyun, ZHAO Hui. Study on Multi-task Student Emotion Recognition Methods Based on Facial Action Units [J]. Computer Science, 2024, 51(10): 105-111.
[4] ZHANG Liguo, XU Xin, DONG Yuxin. Eye Emotion Recognition and Visualization in Smart Classrooms Based on ConvNeXt [J]. Computer Science, 2024, 51(10): 112-118.
[5] CUI Lin, CUI Chenlu, LIU Zhengwei, XUE Kai. Speech Emotion Recognition Based on Improved MFCC and Parallel Hybrid Model [J]. Computer Science, 2023, 50(6A): 220800211-7.
[6] LUO Ruiqi, YAN Jinlin, HU Xinrong, DING Lei. EEG Emotion Recognition Based on Multiple Directed Weighted Graph and ConvolutionalNeural Network [J]. Computer Science, 2023, 50(6A): 220600128-8.
[7] LEI Ying, LIU Feng. LDM-EEG:A Lightweight EEG Emotion Recognition Method Based on Dual-stream Structure Scaling and Multiple Attention Mechanisms [J]. Computer Science, 2023, 50(6A): 220300262-9.
[8] WANG Xiya, ZHANG Ning, CHENG Xin. Review on Methods and Applications of Text Fine-grained Emotion Recognition [J]. Computer Science, 2023, 50(6A): 220900137-7.
[9] XU Ming-ke, ZHANG Fan. Head Fusion:A Method to Improve Accuracy and Robustness of Speech Emotion Recognition [J]. Computer Science, 2022, 49(7): 132-141.
[10] GAO Yue, FU Xiang-ling, OUYANG Tian-xiong, CHEN Song-ling, YAN Chen-wei. EEG Emotion Recognition Based on Spatiotemporal Self-Adaptive Graph ConvolutionalNeural Network [J]. Computer Science, 2022, 49(4): 30-36.
[11] WEI Yan-tao, LUO Jie-lin, HU Mei-jia, LI Wen-hao, YAO Huang. Online Learning Emotion Recognition Based on Videos [J]. Computer Science, 2022, 49(11A): 211000049-6.
[12] ZHU Chen-shuang, CHENG Shi-wei. Prediction and Assistance of Navigation Demand Based on Eye Tracking in Virtual Reality Environment [J]. Computer Science, 2021, 48(8): 315-321.
[13] QIAN Tian-tian, ZHANG Fan. Emotion Recognition System Based on Distributed Edge Computing [J]. Computer Science, 2021, 48(6A): 638-643.
[14] CHAI Bing, LI Dong-dong, WANG Zhe, GAO Da-qi. EEG Emotion Recognition Based on Frequency and Channel Convolutional Attention [J]. Computer Science, 2021, 48(12): 312-318.
[15] ZHENG Chun-jun, WANG Chun-li, JIA Ning. Survey of Acoustic Feature Extraction in Speech Tasks [J]. Computer Science, 2020, 47(5): 110-119.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!