Deep neural network augmentation: generating faces for affect analysis

Article


Kollias, D., Cheng, S., Ververas, E., Kotsia, I. and Zafeiriou, S. 2020. Deep neural network augmentation: generating faces for affect analysis. International Journal of Computer Vision. 128 (5), pp. 1455-1484. https://doi.org/10.1007/s11263-020-01304-3
TypeArticle
TitleDeep neural network augmentation: generating faces for affect analysis
AuthorsKollias, D., Cheng, S., Ververas, E., Kotsia, I. and Zafeiriou, S.
Abstract

This paper presents a novel approach for synthesizing facial affect; either in terms of the six basic expressions (i.e., anger, disgust, fear, joy, sadness and surprise), or in terms of valence (i.e., how positive or negative is an emotion) and arousal (i.e., power of the emotion activation). The proposed approach accepts the following inputs:(i) a neutral 2D image of a person; (ii) a basic facial expression or a pair of valence-arousal (VA) emotional state descriptors to be generated, or a path of affect in the 2D VA space to be generated as an image sequence. In order to synthesize affect in terms of VA, for this person, 600,000 frames from the 4DFAB database were annotated. The affect synthesis is implemented by fitting a 3D Morphable Model on the neutral image, then deforming the reconstructed face and adding the inputted affect, and blending the new face with the given affect into the original image. Qualitative experiments illustrate the generation of realistic images, when the neutral image is sampled from fifteen well known lab-controlled or in-the-wild databases, including Aff-Wild, AffectNet, RAF-DB; comparisons with generative adversarial networks (GANs) show the higher quality achieved by the proposed approach. Then, quantitative experiments are conducted, in which the synthesized images are used for data augmentation in training deep neural networks to perform affect recognition over all databases; greatly improved performances are achieved when compared with state-of-the-art methods, as well as with GAN-based data augmentation, in all cases.

KeywordsArticle, Special Issue on Generating Realistic Visual Data of Human Behavior, Dimensional, Categorical affect, Valence, Arousal, Basic emotions, Facial affect synthesis, 4DFAB, Blendshape models, 3DMM fitting, DNNs, StarGAN, GANimation, Data augmentation, Affect recognition, Facial expression transfer
LanguageEnglish
PublisherSpringer
JournalInternational Journal of Computer Vision
ISSN0920-5691
Electronic1573-1405
Publication dates
Online22 Feb 2020
Print31 May 2020
Publication process dates
Deposited08 Jun 2020
Submitted31 Oct 2018
Accepted05 Feb 2020
Output statusPublished
Publisher's version
License
Copyright Statement

© The Author(s) 2020.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Digital Object Identifier (DOI)https://doi.org/10.1007/s11263-020-01304-3
Permalink -

https://repository.mdx.ac.uk/item/88z87

  • 24
    total views
  • 7
    total downloads
  • 0
    views this month
  • 0
    downloads this month

Export as