情感计算与理解研究发展概述
An overview of research development of affective computing and understanding
查看参考文献225篇
文摘
|
情感在感知、决策、逻辑推理和社交等一系列智能活动中起到核心作用,是实现人机交互和机器智能的重要元素。近年来,随着多媒体数据爆发式增长及人工智能的快速发展,情感计算与理解引发了广泛关注。情感计算与理解旨在赋予计算机系统识别、理解、表达和适应人的情感的能力来建立和谐人机环境,并使计算机具有更高、更全面的智能。根据输入信号的不同,情感计算与理解包含不同的研究方向。本文全面回顾了多模态情感识别、孤独症情感识别、情感图像内容分析以及面部表情识别等不同情感计算与理解方向在过去几十年的研究进展并对未来的发展趋势进行展望。对于每个研究方向,首先介绍了研究背景、问题定义和研究意义;其次从不同角度分别介绍了国际和国内研究现状,包括情感数据标注、特征提取、学习算法、部分代表性方法的性能比较和分析以及代表性研究团队等;然后对国内外研究进行了系统比较,分析了国内研究的优势和不足;最后讨论了目前研究存在的问题及未来的发展趋势与展望,例如考虑个体情感表达差异问题和用户隐私问题等。 |
其他语种文摘
|
Humans are emotional creatures. Emotion plays a key role in various intelligent actions, including perception, decision-making, logical reasoning, and social interaction. Emotion is an important and dispensable component in the realization of human-computer interaction and machine intelligence. Recently, with the explosive growth of multimedia data and the rapid development of artificial intelligence, affective computing and understanding has attracted much research attention. It aims to establish a harmonious human-computer environment by giving the computing machines the ability to recognize, understand, express, and adapt to human emotions, and to make computers have higher and more comprehensive intelligence. Based on the input signals, such as speech, text, image, action and gait, and physiological signals, affective computing and understanding can be divided into multiple research topics. In this paper, we will comprehensively review the development of four important topics in affective computing and understanding, including multi-modal emotion recognition, autism emotion recognition, affective image content analysis, and facial expression recognition. For each topic, we first introduce the research background, problem definition, and research significance. Specifically, we introduce how such topics were proposed, what the corresponding tasks do, and why it is important in different applications. Second, we introduce the international and domestic research on emotion data annotation, feature extraction, learning algorithms, performance comparison and analysis of some representative methods, and famous research teams. Emotion data annotation is conducted to evaluate the performances of affective computing and understanding algorithms. We briefly summarize how categorical and dimensional emotion representation models in psychology are used to construct datasets and the comparisons between these datasets. Feature extraction aims to extract discriminative features to represent emotions. We summarize both hand-crafted features in the early years and deep features in the deep learning era. Learning algorithms aim to learn a mapping between extracted features and emotions. We also summarize and compare both traditional and deep models. For a better understanding of how existing methods work, we report the emotion recognition results of some representative and influential methods on multiple datasets and give some detailed analysis. To better track the latest research for beginners, we briefly introduce some famous research teams with their research focus and main contributions. After that, we systematically compare the international and domestic research, and analyze the advantages and disadvantages of domestic research, which would motivate and boost the future research for domestic researchers and engineers. Finally, we discuss some challenges and promising research directions in the future for each topic, such as 1) image content and context understanding, viewer contextual and prior knowledge modeling, group emotion clustering, viewer and image interaction, and efficient learning for affective image content analysis; 2) data collection and annotation, real-time facial expression analysis, hybrid expression recognition, personalized emotion expression, and user privacy. Since emotion is an abstract, subjective, and complex high-level semantic concept, there are still some limitations of existing methods, and many challenges still remain unsolved. Such promising future research directions would help to reach the emotional intelligence for a better human-computer interaction. |
来源
|
中国图象图形学报
,2022,27(6):2008-2035 【核心库】
|
DOI
|
10.11834/jig.220085
|
关键词
|
情感计算
;
情感识别
;
孤独症
;
图像识别
;
表情识别
|
地址
|
1.
哈尔滨工业大学, 哈尔滨, 150006
2.
北京邮电大学, 北京, 100876
3.
中国科学院心理研究所, 北京, 100083
4.
南开大学, 天津, 300071
5.
美国哥伦比亚大学, 美国, 纽约, 10032
|
语种
|
中文 |
文献类型
|
研究性论文 |
ISSN
|
1006-8961 |
学科
|
自动化技术、计算机技术 |
基金
|
国家自然科学基金项目
|
文献收藏号
|
CSCD:7242309
|
参考文献 共
225
共12页
|
1.
Ahsan U. Towards using visual attributes to infer image sentiment of social events.
Proceedings of 2017 International Joint Conference on Neural Networks,2017:1372-1379
|
CSCD被引
1
次
|
|
|
|
2.
Alameda-Pineda X. Recognizing emotions from abstract paintings using non-linear matrix completion.
Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition,2016:5240-5248
|
CSCD被引
2
次
|
|
|
|
3.
Balouchian P. LUCFER: a large-scale context-sensitive image dataset for deep learning of visual emotions.
Proceedings of 2019 IEEE Winter Conference on Applications of Computer Vision,2019:1645-1654
|
CSCD被引
1
次
|
|
|
|
4.
Barsoum E. Training deep networks for facial expression recognition with crowd-sourced label distribution.
Proceedings of the 18th ACM International Conference on Multimodal Interaction,2016:279-283
|
CSCD被引
18
次
|
|
|
|
5.
Ben X Y. Video-based facial micro-expression analysis: a survey of datasets, features and algorithms.
IEEE Transactions on Pattern Analysis and Machine Intelligence,2021:#3067464
|
CSCD被引
1
次
|
|
|
|
6.
Benitez-Quiroz C F. EmotioNet:an accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild.
Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition,2016:5562-5570
|
CSCD被引
4
次
|
|
|
|
7.
Borth D. Large-scale visual sentiment ontology and detectors using adjective noun pairs.
Proceedings of the 21st ACM International Conference on Multimedia,2013:223-232
|
CSCD被引
18
次
|
|
|
|
8.
Burkhardt F. A database of German emotional speech.
Proceedings of the 9th European Conference on Speech Communication and Technology,2005:1-4
|
CSCD被引
2
次
|
|
|
|
9.
Busso C. IEMOCAP: interactive emotional dyadic motion capture database.
Language Resources and Evaluation,2008,42(4):335-359
|
CSCD被引
94
次
|
|
|
|
10.
Centers for Disease Control and Prevention. Key Findings from the ADDM Network: A Snapshot of Autism Spectrum Disorder.
Community Report on Autism,2016
|
CSCD被引
1
次
|
|
|
|
11.
Chen F H. Predicting microblog sentiments via weakly supervised multimodal deep learning.
IEEE Transactions on Multimedia,2018,20(4):997-1007
|
CSCD被引
5
次
|
|
|
|
12.
Chen J W. VGAN-based image representation learning for privacy-preserving facial expression recognition.
Proceedings of 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops,2018:1683-1692
|
CSCD被引
1
次
|
|
|
|
13.
Chen M. Learning deep features for image emotion classification.
Proceedings of 2015 IEEE International Conference on Image Processing,2015:4491-4495
|
CSCD被引
1
次
|
|
|
|
14.
Chen T.
DeepSentiBank: visual sentiment concept classification with deep convolutional neural networks,2014
|
CSCD被引
8
次
|
|
|
|
15.
Chen Y L. Understanding and mitigating annotation bias in facial expression recognition.
Proceedings of 2021 IEEE/CVF International Conference on Computer Vision,2021:14960-14971
|
CSCD被引
1
次
|
|
|
|
16.
Conner C M. The role of emotion regulation and core autism symptoms in the experience of anxiety in autism.
Autism,2020,24(4):931-940
|
CSCD被引
2
次
|
|
|
|
17.
Dan-Glauser E S. The Geneva affective picture database (GAPED): a new 730-picture database focusing on valence and normative significance.
Behavior Research Methods,2011,43(2):468-477
|
CSCD被引
4
次
|
|
|
|
18.
Darwin C.
The Expression of the Emotions in Man and Animals,2015
|
CSCD被引
1
次
|
|
|
|
19.
Davison A K. SAMM: a spontaneous micro-facial movement dataset.
IEEE Transactions on Affective Computing,2018,9(1):116-129
|
CSCD被引
30
次
|
|
|
|
20.
Dhall A. EmotiW 2019: automatic emotion, engagement and cohesion prediction tasks.
Proceedings of 2019 International Conference on Multimodal Interaction,2019:546-550
|
CSCD被引
1
次
|
|
|
|
|