DR-FER : Discriminative and Robust Representation Learning for Facial Expression Recognition

Li, Ming and Fu, Huazhu and He, Shengfeng and Fan, Hehe and Liu, Jun and Keppo, Jussi and Shou, Mike Zheng (2024) DR-FER : Discriminative and Robust Representation Learning for Facial Expression Recognition. IEEE Transactions on Multimedia, 26. pp. 6297-6309. ISSN 1520-9210

Full text not available from this repository.

Abstract

Learning discriminative and robust representations is important for facial expression recognition (FER) due to subtly different emotional faces and their subjective annotations. Previous works usually address one representation solely because these two goals seem to be contradictory for optimization. Their performances inevitably suffer from challenges from the other representation. In this article, by considering this problem from two novel perspectives, we demonstrate that discriminative and robust representations can be learned in a unified approach, i.e., DR-FER, and mutually benefit each other. Moreover, we make it with the supervision from only original annotations. Specifically, to learn discriminative representations, we propose performing masked image modeling (MIM) as an auxiliary task to force our network to discover expression-related facial areas. This is the first attempt to employ MIM to explore discriminative patterns in a self-supervised manner. To extract robust representations, we present a category-aware self-paced learning schedule to mine high-quality annotated ( easy ) expressions and incorrectly annotated ( hard ) counterparts. We further introduce a retrieval similarity-based relabeling strategy to correct hard expression annotations, exploiting them more effectively. By enhancing the discrimination ability of the FER classifier as a bridge, these two learning goals significantly strengthen each other. Extensive experiments on several popular benchmarks demonstrate the superior performance of our DR-FER. Moreover, thorough visualizations and extra experiments on manually annotation-corrupted datasets show that our approach successfully accomplishes learning both discriminative and robust representations simultaneously.

Item Type:
Journal Article
Journal or Publication Title:
IEEE Transactions on Multimedia
Uncontrolled Keywords:
/dk/atira/pure/subjectarea/asjc/2200/2214
Subjects:
?? media technologysignal processingcomputer science applicationselectrical and electronic engineering ??
ID Code:
223091
Deposited By:
Deposited On:
15 Aug 2024 13:50
Refereed?:
Yes
Published?:
Published
Last Modified:
15 Aug 2024 13:50