TY - GEN
T1 - Examining Subject-Dependent and Subject-Independent Human Affect Inference from Limited Video Data
AU - Parameshwara, Ravikiran
AU - Radwan, Ibrahim
AU - Subramanian, Ramanathan
AU - Goecke, Roland
N1 - Funding Information:
This research was partially supported by the Australian Government through the Australian Research Council s Discovery Projects funding scheme (project DP190101294)
Publisher Copyright:
© 2023 IEEE.
PY - 2023/1/5
Y1 - 2023/1/5
N2 - Continuous human affect estimation from video data entails modelling the dynamic emotional state from a sequence of facial images. Though multiple affective video databases exist, they are limited in terms of data and dynamic annotations, as assigning continuous affective labels to video data is subjective, onerous and tedious. While studies have established the existence of signature facial expressions corresponding to the basic categorical emotions, individual differences in emoting facial expressions nevertheless exist; factoring out these idiosyncrasies is critical for effective emotion inference. This work explores continuous human affect recognition using AFEW-VA, an 'in-the-wild' video dataset with limited data, employing subject-independent (SI) and subject-dependent (SD) settings. The SI setting involves the use of training and test sets with mutually exclusive subjects, while training and test samples corresponding to the same subject can occur in the SD setting. A novel, dynamically-weighted loss function is employed with a Convolutional Neural Network (CNN)-Long Short- Term Memory (LSTM) architecture to optimise dynamic affect prediction. Superior prediction is achieved in the SD setting, as compared to the SI counterpart.
AB - Continuous human affect estimation from video data entails modelling the dynamic emotional state from a sequence of facial images. Though multiple affective video databases exist, they are limited in terms of data and dynamic annotations, as assigning continuous affective labels to video data is subjective, onerous and tedious. While studies have established the existence of signature facial expressions corresponding to the basic categorical emotions, individual differences in emoting facial expressions nevertheless exist; factoring out these idiosyncrasies is critical for effective emotion inference. This work explores continuous human affect recognition using AFEW-VA, an 'in-the-wild' video dataset with limited data, employing subject-independent (SI) and subject-dependent (SD) settings. The SI setting involves the use of training and test sets with mutually exclusive subjects, while training and test samples corresponding to the same subject can occur in the SD setting. A novel, dynamically-weighted loss function is employed with a Convolutional Neural Network (CNN)-Long Short- Term Memory (LSTM) architecture to optimise dynamic affect prediction. Superior prediction is achieved in the SD setting, as compared to the SI counterpart.
KW - Affect Recognition
KW - Emotion Recognition
KW - Subject-dependent
KW - Subject-independent
UR - http://www.scopus.com/inward/record.url?scp=85149264475&partnerID=8YFLogxK
UR - https://ieeexplore.ieee.org/document/10042798
U2 - 10.1109/FG57933.2023.10042798
DO - 10.1109/FG57933.2023.10042798
M3 - Conference contribution
AN - SCOPUS:85149264475
T3 - 2023 IEEE 17th International Conference on Automatic Face and Gesture Recognition, FG 2023
SP - 1
EP - 6
BT - 2023 IEEE 17th International Conference on Automatic Face and Gesture Recognition, FG 2023
PB - IEEE, Institute of Electrical and Electronics Engineers
T2 - 17th IEEE International Conference on Automatic Face and Gesture Recognition
Y2 - 5 January 2023 through 8 January 2023
ER -