Computer Science ›› 2019, Vol. 46 ›› Issue (11A): 13-18.

• Intelligent Computing • Previous Articles     Next Articles

Study on Semantic Topology and Supervised Word Sense Disambiguation of Polysemous Words

XIAO Rui1, JIANG Jia-qi2, ZHANG Yun-chun2   

  1. (School of International Education,Yunnan University,Kunming 650091,China)1;
    (School of Software,Yunnan University,Kunming 650091,China)2
  • Online:2019-11-10 Published:2019-11-20

Abstract: Polysemous words have been considered as learning emphasis and major obstacles for foreign students who learn Chinese as a foreign language.Word Sense Disambiguation (WSD),which is mainly used to determine the specific meaning of polysemous words in a given context,has important application in human-computer interaction,machine translation,automatic essay scoring and other emerging applications.It is also the difficulty in teaching Chinese as a foreign language and HSK examination.The existing word sense disambiguation methods are shown to be with low accuracy,lack of corpus,simple features etc.Considering teaching Chinese as foreign language and its evaluation corpus,it is a hot research topic on building Chinese polysemous words WSD based on deep neural networks.It provides necessary technical support for achieving automatic HSK essay scoring.Existing researches assumed that semantic items are mutually independent and thus paid little attention to their evolutional relationships among items.To solve this problem,this paper firstly made research on semantics of typical Chinese polysemous words.Followed by semantic topology construction the basic semantic items and set-phrases were discriminated for supervised classification model training.Based on the semantic topology construction of polysemous words,the corpus samples were collected by web crawling.The supervised deep neural networks,including RNN,LSTM and GRU,were constructed subsequently.By analyzing the scrawled samples,both uni-directional and bi-directional neural networks were designed by choosing 30 words and 60 words length respectively.The final WSD classification models were obtained by multiple rounds of training and optimization to model parameters.The “Yisi”was chosen for example and used for WSD experiments within their contexts.The experimental results shows that RNN,LSTM and GRU all achieve average classification accuracy more than 75% while maximum accuracy is more than 94%.AUC under each model is more than 0.966,which shows its good performance on class imbalance among samples.Both uni-directional and bi-directional RNN models achieve best classification performance under different words length.

Key words: Deep neural networks, Long short-term memory network, Recurrent neural network, Semantic topology, Word sense disambiguation

CLC Number: 

  • TP391
[1]CORREA E A J,LOPES A D A,AMANCIO D R.Word sense disambiguation:a complex network approach[J].Information Sciences,2018,442:103-113.
[2]万富强,吴云芳.基于中文维基百科的词语语义相关度计算[J].中文信息学报,2013,27(6):31-37.
[3]WU Y,ZHANG M.Overview of the NLPCC 2017 Shared Task:Chinese Word Semantic Relation Classification[C]∥ Natural Language Processing and Chinese Computing(NLPCC 2017).Cham:Springer,2017:919-925.
[4]鲁松,白硕,黄雄.基于向量空间模型中义项词语的无导词义消歧[J].软件学报,2002,13(6):1082-1089.
[5] 陈笑蓉,秦进.基于最大熵原理的汉语词义消歧[J].计算机科学,2005,32(5):174-176.
[6]全昌勤,何婷婷,姬东鸿,等.基于多分类器决策的词义消歧方法[J].计算机研究与发展,2006,43(5):933-939.
[7]曹鸿霞.统计与神经网络相结合的词义消歧模型[J].武汉理工大学学报,2006,28(8):131-134.
[8]张雄,陈福才,黄瑞阳.基于双词主题模型的半监督实体消歧方法研究[J].电子学报,2018,46(3):607-613.
[9]JIMENO Y A.Word embeddings and recurrent neural networks based on Long-Short Term Memory nodes in supervised biomedi-cal word sense disambiguation[J].Journal of biomedical informatics,2017,73:137-147.
[10]CORREA JR E A,LOPES A A,AMANCIO D R.Word sense disambiguation:A complex network approach[J].Information Science,2018,442:103-113.
[11]杨安,李素建,李芸.基于领域知识和词向量的词义消歧方法[J].北京大学学报(自然科学版),2017,53(2):204-210.
[12]杨陟卓,黄河燕.基于语言模型的有监督词义消歧模型优化研究[J].中文信息学报,2014,28(1):19-25.
[13]苏新春,洪桂治,唐师瑶.再论义类词典的分类原则与方法[J].世界汉语教学,2010(2):16-27.
[14]WEN L,LI J,JIN Y,et al.A method for Word Sense Disambiguation combining contextual semantic features[C]∥International Conference on Asian Language Processing.IEEE Press,2017:283-287.
[15]LEI X,CAI Y,LI Q,et al.Combining Local and Global Features in Supervised Word Sense Disambiguation[C]∥Web Information Systems Engineering-WISE 2017.Puschino,Russia:Springer,2017:117-131.
[16]PESARANGHADER A,PESARANGHADER A,MATWIN S,et al.One Single Deep Bidirectional LSTM Network for Word Sense Disambiguation of Text Data[C]∥Advances in Artificial Intelligence.Canadian AI 2018.Toronto,Canada:Springer,2018:96-107.
[17]LOPEZ-AREVALO I,SOSA-SOSA V J,ROJAS-LOPEZ F,et al.Improving selection of synsets from WordNet for domain-specific word sense disambiguation[J].Computer Speech & Language,2016,41(C):128-145.
[18]李安.多义词义项的语义关系及其对词义消歧的影响[J].语言文字应用,2014(1):29-37.
[19]廖光蓉.多义词意义关系模式研究[J].外语教学,2005,26(3):56-59.
[20]UNGERER F,SCHMID H J.An Introduction to Cognitive Linguistics[M].London:Addison Welsey,1996.
[1] WANG Xin-tong, WANG Xuan, SUN Zhi-xin. Network Traffic Anomaly Detection Method Based on Multi-scale Memory Residual Network [J]. Computer Science, 2022, 49(8): 314-322.
[2] PENG Shuang, WU Jiang-jiang, CHEN Hao, DU Chun, LI Jun. Satellite Onboard Observation Task Planning Based on Attention Neural Network [J]. Computer Science, 2022, 49(7): 242-247.
[3] WANG Shan, XU Chu-yi, SHI Chun-xiang, ZHANG Ying. Study on Cloud Classification Method of Satellite Cloud Images Based on CNN-LSTM [J]. Computer Science, 2022, 49(6A): 675-679.
[4] YU Xin, LIN Zhi-liang. Novel Neural Network for Dealing with a Kind of Non-smooth Pseudoconvex Optimization Problems [J]. Computer Science, 2022, 49(5): 227-234.
[5] AN Xin, DAI Zi-biao, LI Yang, SUN Xiao, REN Fu-ji. End-to-End Speech Synthesis Based on BERT [J]. Computer Science, 2022, 49(4): 221-226.
[6] SHI Yu-tao, SUN Xiao. Conversational Comprehension Model for Question Generation [J]. Computer Science, 2022, 49(3): 232-238.
[7] PAN Zhi-hao, ZENG Bi, LIAO Wen-xiong, WEI Peng-fei, WEN Song. Interactive Attention Graph Convolutional Networks for Aspect-based Sentiment Classification [J]. Computer Science, 2022, 49(3): 294-300.
[8] LI Hao, CAO Shu-yu, CHEN Ya-qing, ZHANG Min. User Trajectory Identification Model via Attention Mechanism [J]. Computer Science, 2022, 49(3): 308-312.
[9] XIAO Ding, ZHANG Yu-fan, JI Hou-ye. Electricity Theft Detection Based on Multi-head Attention Mechanism [J]. Computer Science, 2022, 49(1): 140-145.
[10] FAN Hong-jie, LI Xue-dong, YE Song-tao. Aided Disease Diagnosis Method for EMR Semantic Analysis [J]. Computer Science, 2022, 49(1): 153-158.
[11] ZHOU Xin, LIU Shuo-di, PAN Wei, CHEN Yuan-yuan. Vehicle Color Recognition in Natural Traffic Scene [J]. Computer Science, 2021, 48(6A): 15-20.
[12] ZENG You-yu, XIE Qiang. Fault Prediction Method Based on Improved RNN and VAR for Ship Equipment [J]. Computer Science, 2021, 48(6): 184-189.
[13] WANG Xi, ZHANG Kai, LI Jun-hui, KONG Fang, ZHANG Yi-tian. Generation of Image Caption of Joint Self-attention and Recurrent Neural Network [J]. Computer Science, 2021, 48(4): 157-163.
[14] CHEN Qian, CHE Miao-miao, GUO Xin, WANG Su-ge. Recurrent Convolution Attention Model for Sentiment Classification [J]. Computer Science, 2021, 48(2): 245-249.
[15] LYU Ming-qi, HONG Zhao-xiong, CHEN Tie-ming. Traffic Flow Forecasting Method Combining Spatio-Temporal Correlations and Social Events [J]. Computer Science, 2021, 48(2): 264-270.
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!