TCKGE: Transformers with contrastive learning for knowledge graph embedding
Zhang, Xiaowei1; Fang, Quan2; Hu, Jun2; Qian, Shengsheng2; Xu, Changsheng2
发表期刊INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL
ISSN2192-6611
2022-11-27
页码9
通讯作者Fang, Quan(qfang@nlpr.ia.ac.cn)
摘要Representation learning of knowledge graphs has emerged as a powerful technique for various downstream tasks. In recent years, numerous research efforts have been made for knowledge graphs embedding. However, previous approaches usually have difficulty dealing with complex multi-relational knowledge graphs due to their shallow network architecture. In this paper, we propose a novel framework named Transformers with Contrastive learning for Knowledge Graph Embedding (TCKGE), which aims to learn complex semantics in multi-relational knowledge graphs with deep architectures. To effectively capture the rich semantics of knowledge graphs, our framework leverages the powerful Transformers to build a deep hierarchical architecture to dynamically learn the embeddings of entities and relations. To obtain more robust knowledge embeddings with our deep architecture, we design a contrastive learning scheme to facilitate optimization by exploring the effectiveness of several different data augmentation strategies. The experimental results on two benchmark datasets show the superior of TCKGE over state-of-the-art models.
关键词Augmentation Contrastive learning Knowledge graph Transformer
DOI10.1007/s13735-022-00256-3
收录类别SCI
语种英语
资助项目National Natural Science Foundation of China[62036012] ; National Natural Science Foundation of China[62072456] ; National Natural Science Foundation of China[62106262] ; Open Research Projects of Zhejiang Lab[2021KE0AB05]
项目资助者National Natural Science Foundation of China ; Open Research Projects of Zhejiang Lab
WOS研究方向Computer Science
WOS类目Computer Science, Artificial Intelligence ; Computer Science, Software Engineering
WOS记录号WOS:000889032300001
出版者SPRINGER
引用统计
被引频次:2[WOS]   [WOS记录]     [WOS相关记录]
文献类型期刊论文
条目标识符http://ir.ia.ac.cn/handle/173211/50780
专题多模态人工智能系统全国重点实验室_多媒体计算
通讯作者Fang, Quan
作者单位1.Zhengzhou Univ, Henan Inst Adv Technol, Zhengzhou 450001, Peoples R China
2.Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing 100190, Peoples R China
通讯作者单位模式识别国家重点实验室
推荐引用方式
GB/T 7714
Zhang, Xiaowei,Fang, Quan,Hu, Jun,et al. TCKGE: Transformers with contrastive learning for knowledge graph embedding[J]. INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL,2022:9.
APA Zhang, Xiaowei,Fang, Quan,Hu, Jun,Qian, Shengsheng,&Xu, Changsheng.(2022).TCKGE: Transformers with contrastive learning for knowledge graph embedding.INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL,9.
MLA Zhang, Xiaowei,et al."TCKGE: Transformers with contrastive learning for knowledge graph embedding".INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL (2022):9.
条目包含的文件
条目无相关文件。
个性服务
推荐该条目
保存到收藏夹
查看访问统计
导出为Endnote文件
谷歌学术
谷歌学术中相似的文章
[Zhang, Xiaowei]的文章
[Fang, Quan]的文章
[Hu, Jun]的文章
百度学术
百度学术中相似的文章
[Zhang, Xiaowei]的文章
[Fang, Quan]的文章
[Hu, Jun]的文章
必应学术
必应学术中相似的文章
[Zhang, Xiaowei]的文章
[Fang, Quan]的文章
[Hu, Jun]的文章
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。