面向文本事实库的多证据问答方法研究 | |
朱敏郡![]() | |
2023-05-22 | |
页数 | 76 |
学位类型 | 硕士 |
中文摘要 | 问答系统旨在为用户提出的自然语言问题提供精准答案。问答系统不仅是机器智能水平的重要验证手段,也是智能服务的主要形态,适用于智能助理、自动客服和搜索引擎等广泛的应用场景,具有重要研究意义和应用价值。由于文本具有获取容易、表达自然、覆盖广泛等优势,目前较多问答系统依赖文本数据作为知识来源,这类文本问答系统(Textual Question Answering,TQA)是问答系统的重要分支。 近年来,得益于深度学习技术的发展,文本问答系统得以快速发展。文本问答模型已经具备一些简单的理解文本和回答问题的能力。但是,目前的文本问答系统依然面临以下问题:(1)现有文本问答数据集主要关注简单类型的问题,未能包含更多样的复杂问题,同时也缺乏多个证据的详细求解过程,不能很好反映真实场景中的问答需求。(2)现有文本问答方法依赖于对篇章级文本进行整体建模,没有对多个证据句子的细粒度推理过程的建模,在检索证据回答各类复杂问题时缺乏可解释性推理能力。为了解决上述问题,本文从数据集构建、方法研究等方面对文本问答任务中的复杂问题展开系统性探索。整体而言,我们构造了具有细粒度推理过程描述的多证据文本库问答数据集,同时提出了能够获取链式、图式证据链的文本问答方法,以提高文本问答系统的智能水平和可解释性能力。主要的创新点和研究成果包括: |
英文摘要 | Text Question Answering (QA) system is an important research direction in Natural Language Processing, where knowledge retrieval and reasoning are the core capabilities. Real-world question answering involves various complex question types, requiring QA systems to have the ability to support various types of knowledge reasoning and have clear QA processes. However, existing text QA datasets lack in-depth and comprehensive sorting of question types and do not have clear reasoning processes. This paper focuses on complex questions and evidence retrieval in text QA. We construct the first multi-evidence QA dataset with fine-grained multi-structure reasoning processes and propose multiple targeted methods to solve reasoning problems using text knowledge bases and pre-trained language models. This research is not only important for the intelligence and interpretability of QA systems but also for exploring the knowledge reasoning abilities of pre-trained language models. We conduct our research based on a text fact library. The main achievements of this work are as follows: \textbf{Multi-evidence QA dataset based on Textual Fact Databases}: We construct a database called "Textual Fact Database" where each text contains knowledge fact sentences at the sentence level, converted from knowledge graph triplets. The text facts are mutually independent and non-overlapping, but can be related to each other, forming a structure graph composed of relationships. Complex QA tasks should be able to measure a model's ability to reason based on explicit knowledge and include QA solving processes. We first use structured knowledge to text generation technology to convert a structured knowledge base into a natural language text database. Furthermore, since KBQA datasets contain formal reasoning processes that can align reasoning steps to knowledge graph triplets, we map the triplets to text to provide a reasoning process from the question to the answer for each question. Compared with previous datasets, our dataset has significantly improved in terms of breadth and depth of reasoning and can automatically provide accurate supporting evidence processes for complex questions. \textbf{Chain evidence retrieval method based on path modeling}: Based on the retriever-reader architecture, existing multi-hop retrieval models solve complex multi-hop problems through iterative computation, mostly using discrete retrieval methods that do not fully utilize the complete information of evidence chains and perform poorly for questions with longer reasoning paths. Some methods use pre-defined aggregation modules to combine sub-answers, which heavily rely on pre-set question patterns and can only handle fixed types of questions. To fully utilize the evidence chain information, we model the overall path information from a global perspective and introduce single-step and overall chain loss functions to calculate similarity loss, learning both the overall and partial perspectives. This method supports both unordered unsupervised training and ordered supervised training, improving both evidence chain retrieval accuracy and QA accuracy. \textbf{Evidence graph retrieval method based on bidirectional graph}: Although existing text QA retrieval methods can directly retrieve relevant paragraphs complementarily, they cannot retrieve structured evidence graphs, which limits the model's reasoning and interpretability abilities. Previous models typically use single-directional retrieval to obtain supporting evidence and cannot model the structure of evidence (e.g., evidence order) when answering complex questions. In addition, due to the complex structure of multiple chains and hops, single-directional retrieval models may require a large search space. We propose a graph-jump retrieval and reasoning method that includes bidirectional graph-jump retrieval (BGR) and a subgraph reconstruction module for graph-jump evidence retrieval and explanatory graph generation. This greatly reduces the constraints on the evidence graph structure and can be extended to various reasoning problems. Bidirectional retrieval alleviates the problem of the model's accuracy decreasing as the depth increases in single-directional retrieval and the backward search can provide the model with reverse thinking, greatly improving the accuracy of evidence graph retrieval. |
关键词 | 文本问答,文本事实库,双向图检索 |
语种 | 中文 |
七大方向——子方向分类 | 自然语言处理 |
国重实验室规划方向分类 | 语音语言处理 |
是否有论文关联数据集需要存交 | 否 |
文献类型 | 学位论文 |
条目标识符 | http://ir.ia.ac.cn/handle/173211/52284 |
专题 | 毕业生_硕士学位论文 |
推荐引用方式 GB/T 7714 | 朱敏郡. 面向文本事实库的多证据问答方法研究[D],2023. |
条目包含的文件 | ||||||
文件名称/大小 | 文献类型 | 版本类型 | 开放类型 | 使用许可 | ||
论文_签名版.pdf(4740KB) | 学位论文 | 限制开放 | CC BY-NC-SA |
个性服务 |
推荐该条目 |
保存到收藏夹 |
查看访问统计 |
导出为Endnote文件 |
谷歌学术 |
谷歌学术中相似的文章 |
[朱敏郡]的文章 |
百度学术 |
百度学术中相似的文章 |
[朱敏郡]的文章 |
必应学术 |
必应学术中相似的文章 |
[朱敏郡]的文章 |
相关权益政策 |
暂无数据 |
收藏/分享 |
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。
修改评论