Computer Science ›› 2024, Vol. 51 ›› Issue (8): 272-280.doi: 10.11896/jsjkx.230500047
• Artificial Intelligence • Previous Articles Next Articles
GUO Zhiqiang, GUAN Donghai, YUAN Weiwei
CLC Number:
[1]YU L,GUO Z,CHEN G,et al.Review of Knowledge Extraction Technology for Knowledge Graph Construction[J].Journal of the University of Information Engineering,2020,21(2):227-235. [2]ZHONG S S,CHEN X,ZHAO M H,et al.Incorporating word-set attention into Chinese named entity recognition Method[J].Journal of Jilin University(Engineering and Technology Edition),2022,52(5):1098-1105. [3]CHEN Y,XU L,LIU K,et al.Event extraction via dynamic multi-pooling convolutional neural networks[C]//Proceedings of the 53rd Annual Meeting of the Association for Compu-tational Linguistics and the 7th International Joint Conference on Natural Language Processing(Volume 1:Long Papers).2015:167-176. [4]DIEFENBACH D,LOPEZ V,SINGH K,et al.Core techniques of question answering systems over knowledge bases:a survey[J].Knowledge and Information systems,2018,55:529-569. [5]SAITO K,NAGATA M.Multi-language named-entity recognition system based on HMM[C]//Proceedings of the ACL 2003 Workshop on Multilingual and Mixed-language Named Entity Recognition.2003:41-48. [6]CHIEU H L,NG H T.Named entity recognition with a maximum entropy approach[C]//Proceedings of the Seventh Confe-rence on Natural Language Learning at HLT-NAACL 2003.2003:160-163. [7]EKBAL A,BANDYOPADHYAY S.Named entity recognitionusing support vector machine:A language independent approach[J].International Journal of Electrical and Computer Enginee-ring,2010,4(3):589-604. [8]FENG Y,SUN L,LV Y.Chinese word segmentation and named entity recognition based on conditional random fields models[C]//Proceedings of the Fifth SIGHAN Workshop on Chinese Language Processing.2006:181-184. [9]LAMPLE G,BALLESTEROS M,SUBRAMANIAN S,et al.Neural Architectures for Named Entity Recognition[C]//Proceedings of NAACL-HLT.2016:260-270. [10]YANG J,TENG Z,ZHANG M,et al.Combining discrete and neural features for sequence labeling[C]//Computational Linguistics and Intelligent Text Processing:17th International Conference,CICLing 2016.Springer International Publishing,2018:140-154. [11]HE J,WANG H.Chinese named entity recognition and wordsegmentation based on character[C]//Proceedings of the Sixth SIGHAN Workshop on Chinese Language Processing.2008. [12]ZHANG Y,YANG J.Chinese NER Using Lattice LSTM[C]//Proceedings of the 56th Annual Meeting of the Association forComputational Linguistics(Volume 1:Long Papers).2018:1554-1564. [13]LIU W,XU T,XU Q,et al.An encoding strategy based word-character LSTM for Chinese NER[C]//Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies(Long and Short Papers).2019:2379-2389. [14]COLLOBERT R,WESTON J.A unified architecture for natural language processing:Deep neural networks with multitask lear-ning[C]//Proceedings of the 25th International Conference on Machine Learning.2008:160-167. [15]CHEN X,QIU X,ZHU C,et al.Long short-term memory neural networks for chinese word segmentation[C]//Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing.2015:1197-1206. [16]ZHANG B,CAI J,ZHANG H,et al.VisPhone:Chinese named entity recognition model enhanced by visual and phonetic features[J].Information Processing & Management,2023,60(3):103314. [17]MAI C,LIU J,QIU M,et al.Pronounce differently,mean diffe-rently:A multi-tagging-scheme learning method for Chinese NER integrated with lexicon and phonetic features[J].Information Processing & Management,2022,59(5):103041. [18]DONG C,ZHANG J,ZONG C,et al.Character-based LSTM-CRF with radical-level features for Chinese named entity recognition[C]//Natural Language Understanding and Intelligent Applications:5th CCF Conference on Natural Language Processing and Chinese Computing,NLPCC 2016,and 24th International Conference on Computer Processing of Oriental Languages,ICCPOL 2016.Springer International Publishing,2016:239-250. [19]KENTON J D M W C,TOUTANOVA L K.Bert:Pre-training of deep bidirectional transformers for language understanding[C]//Proceedings of naacL-HLT.2019:4171-4186. [20]LI J,FEI H,LIU J,et al.Unified named entity recognition as word-word relation classification[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2022,36(10):10965-10973. [21]LI L,DAI Y,TANG D,et al.MarkBERT:Marking WordBoundaries Improves Chinese BERT[J].arXiv:2203.06378,2022. [22]GU Y,QU X,WANG Z,et al.Delving Deep into Regularity:A Simple but Effective Method for Chinese Named Entity Recognition[C]//Findings of the Association for Computational Linguistics:NAACL 2022.2022:1863-1873. [23]PENG N,DREDZE M.Improving Named Entity Recognition for Chinese Social Media with Word Segmentation Representation Learning[C]//Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics(Volume 2:Short Papers).2016:149-155. [24]LIU W,FU X,ZHANG Y,et al.Lexicon Enhanced Chinese Sequence Labeling Using BERT Adapter[C]//Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing(Volume 1:Long Papers).2021:5847-5858. [25]WANG B,ZHANG Z,XU K,et al.DyLex:Incorporating Dynamic Lexicons into BERT for Sequence Labeling[C]//Procee-dings of the 2021 Conference on Empirical Methods in Natural Language Processing.2021:2679-2693. [26]MENGGE X,YU B,LIU T,et al.Porous lattice transformer en-coder for Chinese NER[C]//Proceedings of the 28th International Conference on Computational Linguistics.2020:3831-3841. [27]HUANG S,SHA Y,LI R.A chinese named entity recognition method for small-scale dataset based on lexicon and unlabeled data[J].Multimedia Tools and Applications,2023,82(2):2185-2206. [28]LIN Z,FENG M,DOS SANTOS C,et al.A structured self-attentive sentence embedding[C]//International Conference on Learning Representations(ICLR).2017. [29]PENG N,DREDZE M.Named entity recognition for chinese social media with jointly trained embeddings[C]//Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing.2015:548-554. [30]HE H,SUN X.A unified model for cross-domain and semi-supervised named entity recognition in chinese social media[C]//Proceedings of the AAAI Conference on Artificial Intelligence.2017. [31]HE H,SUN X.F-Score Driven Max Margin Neural Network for Named Entity Recognition in Chinese Social Media[C]//Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics:Volume 2,Short Papers.2017:713-718. [32]CHEN A,PENG F,SHAN R,et al.Chinese named entity recognition with conditional probabilistic models[C]//Proceedings of the Fifth SIGHAN Workshop on Chinese Language Processing.2006:173-176. [33]ZHANG S,QIN Y,HOU W J,et al.Word segmentation andnamed entity recognition for sighan bakeoff3[C]//Proceedings of the Fifth SIGHAN Workshop on Chinese Language Proces-sing.2006:158-161. [34]ZHOU J,QU W,ZHANG F.Chinese named entity recognition via joint identification and categorization[J].Chinese Journal of Electronics,2013,22(2):225-230. [35]LU Y,ZHANG Y,JI D.Multi-prototype Chinese character embedding[C]//Proceedings of the Tenth International Conference on Language Resources and Evaluation(LREC’16).2016:855-859. [36]CAO P,CHEN Y,LIU K,et al.Adversarial transfer learning for Chinese named entity recognition with self-attention mechanism[C]//Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing.2018:182-192. |
[1] | YANG Binxia, LUO Xudong, SUN Kaili. Recent Progress on Machine Translation Based on Pre-trained Language Models [J]. Computer Science, 2024, 51(6A): 230700112-8. |
[2] | WANG Yingjie, ZHANG Chengye, BAI Fengbo, WANG Zumin. Named Entity Recognition Approach of Judicial Documents Based on Transformer [J]. Computer Science, 2024, 51(6A): 230500164-9. |
[3] | LI Minzhe, YIN Jibin. TCM Named Entity Recognition Model Combining BERT Model and Lexical Enhancement [J]. Computer Science, 2024, 51(6A): 230900030-6. |
[4] | YIN Baosheng, ZHOU Peng. Chinese Medical Named Entity Recognition with Label Knowledge [J]. Computer Science, 2024, 51(6A): 230500203-7. |
[5] | PENG Bo, LI Yaodong, GONG Xianfu, LI Hao. Method for Entity Relation Extraction Based on Heterogeneous Graph Neural Networks and TextSemantic Enhancement [J]. Computer Science, 2024, 51(6A): 230700071-5. |
[6] | LI Bin, WANG Haochang. Implementation and Application of Chinese Grammatical Error Diagnosis System Based on CRF [J]. Computer Science, 2024, 51(6A): 230900073-6. |
[7] | LAI Xin, LI Sining, LIANG Changsheng, ZHANG Hengyan. Ontology-driven Study on Information Structuring of Aeronautical Information Tables [J]. Computer Science, 2024, 51(6A): 230800150-7. |
[8] | YU Bihui, TAN Shuyue, WEI Jingxuan, SUN Linzhuang, BU Liping, ZHAO Yiman. Vision-enhanced Multimodal Named Entity Recognition Based on Contrastive Learning [J]. Computer Science, 2024, 51(6): 198-205. |
[9] | ZHANG Mingdao, ZHOU Xin, WU Xiaohong, QING Linbo, HE Xiaohai. Unified Fake News Detection Based on Semantic Expansion and HDGCN [J]. Computer Science, 2024, 51(4): 299-306. |
[10] | TU Xin, ZHANG Wei, LI Jidong, LI Meijiao , LONG Xiangbo. Study on Automatic Classification of English Tense Exercises for Intelligent Online Teaching [J]. Computer Science, 2024, 51(4): 353-358. |
[11] | LIAO Meng, JIA Zhen, LI Tianrui. Chinese Named Entity Recognition Based on Label Information Fusion and Multi-task Learning [J]. Computer Science, 2024, 51(3): 198-204. |
[12] | ZHENG Cheng, SHI Jingwei, WEI Suhua, CHENG Jiaming. Dual Feature Adaptive Fusion Network Based on Dependency Type Pruning for Aspect-basedSentiment Analysis [J]. Computer Science, 2024, 51(3): 205-213. |
[13] | GE Huibin, WANG Dexin, ZHENG Tao, ZHANG Ting, XIONG Deyi. Study on Model Migration of Natural Language Processing for Domestic Deep Learning Platform [J]. Computer Science, 2024, 51(1): 50-59. |
[14] | GU Shiwei, LIU Jing, LI Bingchun, XIONG Deyi. Survey of Unsupervised Sentence Alignment [J]. Computer Science, 2024, 51(1): 60-67. |
[15] | MAO Xin, LEI Zhanyao, QI Zhengwei. Automated Kaomoji Extraction Based on Large-scale Danmaku Texts [J]. Computer Science, 2024, 51(1): 284-294. |
|