Knowledge Commons of Institute of Automation,CAS
面向交互场景的情感识别研究 | |
连政![]() | |
2021 | |
页数 | 106 |
学位类型 | 博士 |
中文摘要 | 情感识别是一项通过分析情感表达时所产生的生理反应和行为反应来识别 |
英文摘要 | Emotion recognition is a technology that recognizes emotional states by analyzing the physiological and behavioral responses generated from emotion expression. As an important branch of artifcial intelligence, emotion recognition can be widely utilized in interaction, education, security and fnance. With the widespread applications of human-computer interaction systems and social network platforms, the use of smart devices for human-computer and human-to-human interaction has become a part of our daily life. However, existing interaction systems pay more attention to the speech content. They fail to fully utilize emotional information, which affects the naturalness and humanity. Therefore, emotion recognition technology oriented to interactive scenes has received numerous attentions from researches of home and abroad. This paper conducts research on emotion recognition methods in interactive scenes from three aspects: emotion feature extraction, multimodal fusion, and individual information modeling. The main contribution of this paper can be summarized as follows: 1. In the aspect of emotional feature extraction, this paper aims to improve the performance of emotion recognition by learning distinguishing emotional features. Firstly, considering the shortcoming of poor discrimination of different emotion states, this paper utilizes the discriminative loss function for emotional feature extraction. The proposed method uses the contrastive loss and supervised cross-entropy loss to jointly optimize the trainable parameters. The contrastive loss can reduce the intra-class distance and increase the inter-class distance, thus learning distinguishing emotional features. 2. In the aspect of multimodal fusion, this paper proposes an emotion recognition system based on cross-modal interactive modeling. This paper takes word-level lexical features and segment-level acoustic features as the inputs. These features have different sequence length and exhibit “unaligned” nature. This paper proposes a transformerbased multimodal fusion strategy to obtain optimal mapping between these features, and then learns cross-modal interactions to improve the performance of multimodal emotion recognition systems.
|
关键词 | 交互场景 情感识别 情感特征提取 多模态融合 个体信息建模 |
语种 | 中文 |
七大方向——子方向分类 | 多模态智能 |
文献类型 | 学位论文 |
条目标识符 | http://ir.ia.ac.cn/handle/173211/44747 |
专题 | 多模态人工智能系统全国重点实验室_智能交互 |
推荐引用方式 GB/T 7714 | 连政. 面向交互场景的情感识别研究[D]. 中国科学院自动化研究所. 中国科学院自动化研究所,2021. |
条目包含的文件 | ||||||
文件名称/大小 | 文献类型 | 版本类型 | 开放类型 | 使用许可 | ||
Thesis-Zheng Lian.pd(4140KB) | 学位论文 | 开放获取 | CC BY-NC-SA |
个性服务 |
推荐该条目 |
保存到收藏夹 |
查看访问统计 |
导出为Endnote文件 |
谷歌学术 |
谷歌学术中相似的文章 |
[连政]的文章 |
百度学术 |
百度学术中相似的文章 |
[连政]的文章 |
必应学术 |
必应学术中相似的文章 |
[连政]的文章 |
相关权益政策 |
暂无数据 |
收藏/分享 |
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。
修改评论