详细信息
VAD-Net: Multidimensional Emotion Recognition from Facial Expression Images ( CPCI-S收录 EI收录)
文献类型:会议论文
英文题名:VAD-Net: Multidimensional Emotion Recognition from Facial Expression Images
作者:Huo, Yi[1];Ge, Yun[2]
第一作者:霍奕
通讯作者:Huo, Y[1]
机构:[1]Beijing Union Univ, Educ Informat Technol, Teachers Coll, Beijing, Peoples R China;[2]Univ Chinese Acad Social Sci, Dept Comp Teaching & Res, Beijing, Peoples R China
第一机构:北京联合大学师范学院
通讯机构:[1]corresponding author), Beijing Union Univ, Educ Informat Technol, Teachers Coll, Beijing, Peoples R China.|[1141711]北京联合大学师范学院;[11417]北京联合大学;
会议论文集:International Joint Conference on Neural Networks (IJCNN)
会议日期:JUN 30-JUL 05, 2024
会议地点:Yokohama, JAPAN
语种:英文
外文关键词:VAD Benchmark; VAD Predication Baseline; Facial Expression Recognition; Affective Computing; Intelligent Education System
摘要:Current FER (Facial Expression Recognition) dataset is mostly labeled by emotion categories, such as happy, angry, sad, fear, disgust, surprise, and neutral which are limited in expressiveness. However, future affective computing requires more comprehensive and precise emotion metrics which could be measured by VAD( Valence-Arousal-Dominance) multidimension parameters. To address this, AffectNet has tried to add VA (Valence and Arousal) information, but still lacks D(Dominance). Thus, the research introduces VAD annotation on FER2013 dataset, takes the initiative to label D(Dominance) dimension. Then, to further improve VAD prediction accuracy, it enforces orthogonalized convolution on regression network to extract more diverse and expressive features. Experiment results show that D dimension could be measured but is difficult to obtain compared with V and A dimension, no matter in manual annotation or regression model prediction. Furthermore, the ablation test is carried out by introducing orthogonal convolution whose results verifies that better VAD prediction could be achieved under the configuration of orthogonalized convolution. Therefore, the research provides an initial annotation work for D(Dominance) dimension on FER dataset, and proposes a better regression network for VAD prediction through orthogonalized operation. The newly built VAD annotated FER2013 dataset could act as a benchmark to measure VAD multidimensional emotions, while the orthogonalized regression network could act as the baseline for VAD facial expression recognition. The newly labeled VAD dataset and prediction baseline code is publicly available on Github: https://github.com/ YeeHoran/ VAD-Net.
参考文献:
正在载入数据...