面向知识图谱补全的小样本学习方法研究 | |
李金林 | |
2024-05-18 | |
页数 | 58 |
学位类型 | 硕士 |
中文摘要 | 随着移动互联网的持续发展,数据量呈爆炸式增长,知识图谱成为有效管理 |
英文摘要 | With the continuous development of the internet, the amount of data is experiencing explosive growth, and knowledge graph has become an effective tool for managing these massive data. In a knowledge graph, entities and concepts are represented as nodes, and the connections between these nodes are explicitly presented in the form of edges. A notable characteristic of large-scale knowledge graphs is that their data follows a long-tail distribution, so accurately predicting low-frequency relationships in the knowledge graph is a pressing issue in current research. This paper combines advanced technologies from deep learning, pre-trained language models, and large language models to study a knowledge graph few-shot learning method based on interactive learning, text augmentation, and large model precise ranking. 1. Few-shot learning method based on interactive learning Regarding the issue of the independence of the support set and query set in existing methods during the modeling process, this paper proposes a few-shot learning method based on interactive learning. In current research methods, modeling the support set and query set separately results in the same triplet vector for different relation tasks in the query set, and due to limitations in vector dimensions, it is difficult to achieve multi-angle and multi-level representations of relation features. In addition, there may be multiple relationships between the head and tail entities of a query, and unrelated relationships may interfere with the semantic representation of task relationships. In order to better adapt to various relationship tasks, the model needs to establish dynamic vector representations based on different types of relationships. This semantic information about task relationships is implicitly contained in the support set. This paper explores this implicit information through interactive learning of facts and uses it to adjust the modeling of the query set. To enhance the model's generalization capability, an adaptive loss module is added to the model, which allows the model to pay more attention to difficult-to-classify negative samples, thereby improving the robustness of the model. Experimental validation on two public benchmark datasets confirms the effectiveness of the method. 2. Entity Description-enhanced few-shot Learning Method In order to address the issue of entity representation errors caused by insufficient background knowledge graph information, various methods can be considered to enhance entity expressions. In this paper, a method of providing relevant text descriptions for entities is employed to more comprehensively depict the attributes and relationships of entities. This paper proposes a method of heterogeneous information fusion, which extracts the neighborhood information of entities on the background knowledge graph using a single-layer graph neural network, while also using BERT to extract the textual semantic information of entities. To effectively differentiate the importance of different information features, this paper designs a gating network to adaptively regulate weights during the feature fusion process, thereby better utilizing information from different sources. To validate the model's effectiveness, this paper introduces the FB15K-237-One dataset. On this dataset, the model proposed in this study achieves optimal performance. Through the research presented in this paper, the issue of entity representation errors caused by insufficient background knowledge graph information has been successfully resolved. 3. Few-shot learning method based on large language model fine ranking In the field of few-shot learning in knowledge graph, large models are still in the exploration stage. This article proposes a few-shot learning method based on large-model fine ranking. With the continuous evolution of natural language processing technology, pre-trained language models have gradually developed into large language models, demonstrating outstanding semantic understanding and reasoning capabilities in various domains. Building upon the research on few-shot learning methods with enhanced entity descriptions, this study integrates with an efficient fine-tuning approach to achieve fine-tuning of large models for few-shot learning in knowledge graph on consumer-grade mid-to-low-end devices. The specific implementation involves running a recall model on the FB15K-237-One dataset to obtain the rearranged dataset, which is then transformed into a format that is easy for LLaMa2 to understand, and finally fine-tuning the parameters using efficient fine-tuning methods. Experimental results show that LLaMa2 fine-tuned based on LoRA performs well on the rearranged dataset, demonstrating its effectiveness and superiority in the task of few-shot learning in knowledge graph, providing a new approach and avenue for completing few-shot knowledge graphs. |
关键词 | 小样本学习 知识图谱补全 交互学习 预训练语言模型 大语言模型 |
语种 | 中文 |
文献类型 | 学位论文 |
条目标识符 | http://ir.ia.ac.cn/handle/173211/56648 |
专题 | 毕业生_硕士学位论文 |
推荐引用方式 GB/T 7714 | 李金林. 面向知识图谱补全的小样本学习方法研究[D],2024. |
条目包含的文件 | ||||||
文件名称/大小 | 文献类型 | 版本类型 | 开放类型 | 使用许可 | ||
毕业论文终稿_new.pdf(1855KB) | 学位论文 | 限制开放 | CC BY-NC-SA |
个性服务 |
推荐该条目 |
保存到收藏夹 |
查看访问统计 |
导出为Endnote文件 |
谷歌学术 |
谷歌学术中相似的文章 |
[李金林]的文章 |
百度学术 |
百度学术中相似的文章 |
[李金林]的文章 |
必应学术 |
必应学术中相似的文章 |
[李金林]的文章 |
相关权益政策 |
暂无数据 |
收藏/分享 |
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。
修改评论