计算机科学 ›› 2010, Vol. 37 ›› Issue (1): 245-250.

• 人工智能 • 上一篇    下一篇

基于Fuzzy Rough集模型的汉语人称代词消解

李凡,刘启和,李洪伟   

  1. (电子科技大学计算机科学与工程学院 成都610054)
  • 出版日期:2018-12-01 发布日期:2018-12-01
  • 基金资助:
    本文受国家自然科学基金(60873077)资助。

Pronominal Anaphora Resolution within Chinese Text Based on Fuzzy Rough Sets Model

LI Fan,LIU Qi-he,LI Hong-wei   

  • Online:2018-12-01 Published:2018-12-01

摘要: 指代消解是自然语言处理中重要的研究课题之一。结合基于实例的学习方法,提出了一种基于Fuzzy Rough集模型的中文人称代词消解方法。该方法的第一步过滤掉与人称代词性别和单复数特征不一致的名词短语,构成候选集,然后按照仅步及浅层语义和语法知识的属性集对其中的每个名词短语进行标记。第二步利用Fuzzy Rough集模型中相关概念选择代表性较强的实例,并对其进行属性值约简,以提高这些实例的泛化能力。以上两步即为学习阶段。第三步即可根据这些实例判断新输入的名词短语是否为代词的先行语。该方法用人民日报语料进行了测试,测试结果表明该方法是有效的。

关键词: 指代消解,先行语,Fuzzy Rough集,基于实例的学习

Abstract: Anaphora resolution is an important issue in natural language processing. This paper presented an approach based on Fuzzy Rough sets model combined with instance -based learning approach to resolve pronominal anaphora within Chinese text. The first phase of the presented approach is preprocessing. In this phase,after extracting noun phases and eliminating those whose number and gender features arc inconsistent with pronominal anaphora, the potential antecedents set was formed. hhen, the attri-butc values of every noun phase in this set were computed according to an attribute set which only involves shallow syntactic and semantic information. The second phase aimed to select representative examples from the potential antecedents set and reduce redundant attributes to improve the generalization capability of these examples. hhese tasks were done by using concepts of Fuzzy Rough sets model. The two phases above can be regarded as learning phase. In the last phase, those examples were used to estimate whether a new noun phase is the that antecedent of a pronominal anaphor. The presented approach was tested by People Daily corpus. The results show this approach is effective.

Key words: Anaphora resolution,Antecedent,Fuzzy Rough sets,Instance-based learning

No related articles found!
Viewed
Full text


Abstract

Cited

  Shared   
  Discussed   
No Suggested Reading articles found!