We propose a novel landmarks-assisted collaborative end-to-end deep framework for 4D facial expression recognition (FER). Using 4D face scan data, we calculate its various geometrical images, and afterwards use rank pooling to generate their dynamic images encapsulating important facial muscle movements over time. As well, the given 3D landmarks are projected on a 2D plane as binary images and convolutional layers are used to extract sequences of feature vectors for every landmark video. During the training stage, the dynamic images are used to train an end-to-end deep network, while the feature vectors of landmark images are used train a long short-term memory (LSTM) network. The finally improved set of expression predictions are obtained when the dynamic and landmark images collaborate over multi-views using the proposed deep framework. Performance results obtained from extensive experimentation on the widely-adopted BU-4DFE database under globally used settings prove that our proposed collaborative framework outperforms the state-of-the-art 4D FER methods and reach a promising classification accuracy of 96.7% demonstrating its effectiveness.
Behzad Muzammil, Vo Nhat, Li Xiaobai, Zhao Guoying
A4 Article in conference proceedings
Place of publication:
2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020)
M. Behzad, N. Vo, X. Li and G. Zhao, “Landmarks-assisted Collaborative Deep Framework for Automatic 4D Facial Expression Recognition,” 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), Buenos Aires, Argentina, 2020, pp. 1-5, doi: 10.1109/FG47880.2020.00023
Read the publication here: