Computer Science ›› 2020, Vol. 47 ›› Issue (3): 231-236.doi: 10.11896/jsjkx.190100108
• Artificial Intelligence • Previous Articles Next Articles
FU Jian,KONG Fang
CLC Number:
[1]HOBBS J R.Resolving pronoun references[J].Lingua,1978,44(4):311-338. [2]LAPPIN S,LEASS H J.An algorithm for pronominal anaphora resolution[J].Computational linguistics,1994,20(4):535-561. [3]MCCORD M C.Slot grammar[M]∥Natural Language and Logic. Berlin:Springer,1990:118-145. [4]KONG F,ZHOU G D.Pronoun Resolution in English and Chinese Languages Based on Tree Kernel[J].Journal of Software,2012,23(5):1085-1099. [5]MIKOLOV T,SUTSKEVER I,CHEN K,et al.Distributed representations of words and phrases and their compositionality[C]∥Advances in Neural Information Processing Systems.Lake Tahoe:NIPS,2013:3111-3119. [6]CLARK K,MANNING C D.Deep Reinforcement Learning for Mention-Ranking Coreference Models[C]∥Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing.Austin:EMNLP,2016:2256-2262. [7]PRADHAN S,MOSCHITTI A,XUE N,et al.CoNLL-2012 shared task:Modeling multilingual unrestricted coreference in OntoNotes[C]∥Joint Conference on EMNLP and CoNLL-Shared Task.Jeju Island:ACL,2012:1-40. [8]WU J L,MA W Y.A deep learning framework for coreference resolution based on convolutional neural network[C]∥2017 IEEE 11th International Conference on Semantic Computing (ICSC).San Diego:IEEE,2017:61-64. [9]LEE K,HE L,LEWIS M,et al.End-to-end Neural Coreference Resolution[C]∥Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing.Copenhagen:ACL,2017:188-197. [10]HOCHREITER S,SCHMIDHUBER J.Long short-term memory[J].Neural Computation,1997,9(8):1735-1780. [11]LEE K,HE L,ZETTLEMOYER L.Higher-Order Coreference Resolution with Coarse-to-Fine Inference[C]∥Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.New Orleans:ACL,2018,2:687-692. [12]PETERS M,NEUMANN M,IYYER M,et al.Deep Contextuali- zed Word Representations[C]∥Proceedings of the 2018 Confe-rence of the North American Chapter of the Association for Computational Linguistics:Human Language Technologies.New Orleans:ACL,2018:2227-2237. [13]LIANG D,XU W,ZHAO Y.Combining word-level and character-level representations for relation classification of informal text[C]∥Proceedings of the 2nd Workshop on Representation Learning for NLP.Vancouver:ACL,2017:43-47. [14]ZHANG X,ZHAO J,LECUN Y.Character-level convolutional networks for text classification[C]∥Advances in Neural Information Processing Systems.Montreal:NIPS,2015:649-657. [15]LING W,DYER C,BLACK A W,et al.Finding Function in Form:Compositional Character Models for Open Vocabulary Word Representation[C]∥Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing.Lisbon:ACL,2015:1520-1530. [16]VILAIN M,BURGER J,ABERDEEN J,et al.A model-theore- tic coreference scoring scheme[C]∥Proceedings of the 6th Conference on Message Understanding.Columbia:ACL,1995:45-52. [17]BAGGA A,BALDWIN B.Algorithms for scoring coreference chains[C]∥The First International Conference on Language Resources and Evaluation Workshop on Linguistics Corefe-rence.Granada:LREC,1998,1:563-566. [18]LUO X.On coreference resolution performance metrics[C]∥ Proceedings of the Conference on Human Language Technology and Empirical Methods in Natural Language Processing.Vancouver:ACL,2005:25-32. [19]NAIR V,HINTON G E.Rectified linear units improve restric- ted boltzmann machines[C]∥Proceedings of the 27th International Conference on International Conference on Machine Learning.Haifa:Omni press,2010:807-814. [20]SRIVASTAVA N,HINTON G,KRIZHEVSKY A,et al.Dropout:a simple way to prevent neural networks from overfitting[J].The Journal of Machine Learning Research,2014,15(1):1929-1958. [21]KINGMA D P,BA J.Adam:A method for stochastic optimization[J].arXiv:1412.6980,2014. [22]AL-RFOU R,PEROZZI B,SKIENA S.Polyglot:Distributed Word Representations for Multilingual NLP [C]∥Proceedings of the Seventeenth Conference on Computational Natural Language Learning.Sofia:ACL,2013:183-192. [23]PENNINGTON J,SOCHER R,MANNING C.Glove:Global vectors for word representation[C]∥Proceedings of the 2014 Conference on Empirical Methods in Natural Language Proces-sing (EMNLP).Doha:ACL,2014:1532-1543. [24]CLARK K,MANNING C D.Improving Coreference Resolution by Learning Entity-Level Distributed Representations[C]∥Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics.Berlin:ACL,2016:643-653. [25]DEVLIN J,CHANG M W,LEE K,et al.Bert:Pre-training of deep bidirectional transformers for language understanding[J].arXiv:1810.04805,2018. |
[1] | JIANG Meng-han, LI Shao-mei, ZHENG Hong-hao, ZHANG Jian-peng. Rumor Detection Model Based on Improved Position Embedding [J]. Computer Science, 2022, 49(8): 330-335. |
[2] | SHUAI Jian-bo, WANG Jin-ce, HUANG Fei-hu, PENG Jian. Click-Through Rate Prediction Model Based on Neural Architecture Search [J]. Computer Science, 2022, 49(7): 10-17. |
[3] | HOU Yu-tao, ABULIZI Abudukelimu, ABUDUKELIMU Halidanmu. Advances in Chinese Pre-training Models [J]. Computer Science, 2022, 49(7): 148-163. |
[4] | HAN Hong-qi, RAN Ya-xin, ZHANG Yun-liang, GUI Jie, GAO Xiong, YI Meng-lin. Study on Cross-media Information Retrieval Based on Common Subspace Classification Learning [J]. Computer Science, 2022, 49(5): 33-42. |
[5] | ZHONG Gui-feng, PANG Xiong-wen, SUI Dong. Text Classification Method Based on Word2Vec and AlexNet-2 with Improved AttentionMechanism [J]. Computer Science, 2022, 49(4): 288-293. |
[6] | LI Yong, WU Jing-peng, ZHANG Zhong-ying, ZHANG Qiang. Link Prediction for Node Featureless Networks Based on Faster Attention Mechanism [J]. Computer Science, 2022, 49(4): 43-48. |
[7] | YANG Hui, TAO Li-hong, ZHU Jian-yong, NIE Fei-ping. Fast Unsupervised Graph Embedding Based on Anchors [J]. Computer Science, 2022, 49(4): 116-123. |
[8] | CHEN Shi-cong, YUAN De-yu, HUANG Shu-hua, YANG Ming. Node Label Classification Algorithm Based on Structural Depth Network Embedding Model [J]. Computer Science, 2022, 49(3): 105-112. |
[9] | GUO Lei, MA Ting-huai. Friend Closeness Based User Matching [J]. Computer Science, 2022, 49(3): 113-120. |
[10] | YANG Xu-hua, WANG Lei, YE Lei, ZHANG Duan, ZHOU Yan-bo, LONG Hai-xia. Complex Network Community Detection Algorithm Based on Node Similarity and Network Embedding [J]. Computer Science, 2022, 49(3): 121-128. |
[11] | LI Yu-qiang, ZHANG Wei-jiang, HUANG Yu, LI Lin, LIU Ai-hua. Improved Topic Sentiment Model with Word Embedding Based on Gaussian Distribution [J]. Computer Science, 2022, 49(2): 256-264. |
[12] | LI Zhao-qi, LI Ta. Query-by-Example with Acoustic Word Embeddings Using wav2vec Pretraining [J]. Computer Science, 2022, 49(1): 59-64. |
[13] | LIU Kai, ZHANG Hong-jun, CHEN Fei-qiong. Name Entity Recognition for Military Based on Domain Adaptive Embedding [J]. Computer Science, 2022, 49(1): 292-297. |
[14] | ZHENG Su-su, GUAN Dong-hai, YUAN Wei-wei. Heterogeneous Information Network Embedding with Incomplete Multi-view Fusion [J]. Computer Science, 2021, 48(9): 68-76. |
[15] | SUN Sheng-zi, GUO Bing-hui , YANG Xiao-bo. Embedding Consensus Autoencoder for Cross-modal Semantic Analysis [J]. Computer Science, 2021, 48(7): 93-98. |
|