首页|CCL23-Eval任务5总结报告:跨领域句子级别中文省略消解

Overview of CCL23-Eval Task 5: Sentence Level Multi-domain Chinese Ellipsis Resolution

CCL23-Eval任务5总结报告:跨领域句子级别中文省略消解

李炜 邵艳秋 祁佳璐

Overview of CCL23-Eval Task 5: Sentence Level Multi-domain Chinese Ellipsis Resolution

CCL23-Eval任务5总结报告:跨领域句子级别中文省略消解

李炜 1邵艳秋 1祁佳璐1
扫码查看

作者信息

  • 1. 北京语言大学信息科学学院北京市海淀区学院路15号,100083
  • 折叠

摘要

省略是一种会出现在包括中文在内的各种语言中的一种语言现象。虽然人类一般能够正确理解带有省略的文本,但是其对机器在句法、语义等方面的理解却会造成影响。因此自动恢复省略成分对文本自动分析理解具有重要意义。本任务提出一个面向应用的省略恢复任务,旨在恢复在句子句法结构中占据有效位置同时在句子中扮演语义成分的被省略内容。本任务将省略恢复任务划分成两个子任务:省略位置探测和省略内容生成,并分别描述在两个子任务中取得较好结果的基线方法。此外,为了推进对大语言模型的研究,本文还尝试使用场景学习的方法使用ChatGPT来完成本任务,并进行了相关分析。

Abstract

Ellipsis is a linguistic phenomenon that occurs in various languages, including Chinese. Although humans can generally understand text with omissions correctly, it can have an impact on machine understanding in terms of syntax and semantics. Therefore, the automatic recovery of omitted elements is of significant importance for automated text analysis and comprehension. This task proposes a computationally feasible omission recovery task that aims to restore omitted constituents that occupy valid positions in the syntactic structure of a sentence while playing a semantic role. The task is divided into two subtasks: ellipsis position detection and ellipsis content generation. Baseline methods that have achieved good results in both subtasks are described. Additionally, to advance research on large language models, this study also attempts to utilize the approach of in context learning using ChatGPT to perform this task and conducts relevant analysis.

关键词

省略探测/省略恢复

Key words

省略探测/省略恢复

引用本文复制引用

会议名称

Chinese national conference on computational linguistics

会议地点

Harbin(CN)

会议母体文献

22nd Chinese national conference on computational linguistics (CCL 2023): evaluations

页码

159-165

出版时间

2023
段落导航相关论文