Black-box Machine Learning (ML) models are widely used in Affective Computing applications, including in high-stake tasks, such as autonomous driving or mental health applications. Despite their state-of-the-art performance, blackbox models are usually criticized as non-interpretable as they only produce a final prediction without providing insight into the decision-making process. Previous research on explainable affective computing applications mainly focuses on post-hoc techniques or reverting to traditional ML approaches, which may compromise either the explainability or the performance. Recently, concept-based models have demonstrated great success in general object classification tasks by training the ML model to learn both the classification label as well as the underlying concepts that contribute to the prediction.
However, the experimental scenario is different in affective computing applications, for example, facial expression recognition or emotion detection, where it is challenging to select meaningful concepts because of the inherent uncertainty of the tasks. Our work is the first to present a Concept-based FER approach and explore concept selection for this application area leveraging concept-based models, including Concept Bottleneck Models (CBMs) and Concept Embedding Models (CEMs). We believe the proposed framework paves the way for future interpretable investigations into downstream applications in human behavior understanding.
Xinyu Li and Marwa Mahmoud
[proper citation]