Personalized Feature Translation for Expression Recognition: An Efficient Source-Free Domain Adaptation Method
About
Facial expression recognition (FER) models are widely used in video-based affective computing applications, such as human-computer interaction and healthcare monitoring. However, deep FER models often struggle with subtle expressions and high inter-subject variability, limiting performance in real-world settings. Source-free domain adaptation (SFDA) has been proposed to personalize a pretrained source model using only unlabeled target data, avoiding privacy, storage, and transmission constraints. We address a particularly challenging setting where source data is unavailable and the target data contains only neutral expressions. Existing SFDA methods are not designed for adaptation from a single target class, while generating non-neutral facial images is often unstable and expensive. To address this, we propose Source-Free Domain Adaptation with Personalized Feature Translation (SFDA-PFT), a lightweight latent-space approach. A translator is first pretrained on source data to map subject-specific style features between subjects while preserving expression information through expression-consistency and style-aware objectives. It is then adapted to neutral target data without source data or image synthesis. By operating in the latent space, SFDA-PFT avoids noisy facial image generation, reduces computation, and learns discriminative embeddings for classification. Experiments on BioVid, StressID, BAH, and Aff-Wild2 show that SFDA-PFT consistently outperforms state-of-the-art SFDA methods in privacy-sensitive FER scenarios. Our code is publicly available at: \href{https://github.com/MasoumehSharafi/SFDA-PFT}{GitHub}.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Facial Expression Recognition | Aff-Wild2 (10 target subjects) | Accuracy (Subject 1)60.83 | 18 | |
| Ambivalence/hesitancy recognition | BAH 10 target subjects, 214 source subjects | Subject 1 Performance69.46 | 9 | |
| Facial Expression Recognition | BioVid (target subjects (10)) | Accuracy (Sub-1)80.65 | 9 | |
| Facial Expression Recognition | BAH 214 source subjects (10 target subjects) | Accuracy (Sub-1)61.52 | 9 | |
| Pain Recognition | BioVid 77 source subjects 10 target subjects | Subject 1 Accuracy84.93 | 9 | |
| Stress Recognition | StressID | Sub-1 Score78.33 | 9 | |
| Pain Recognition | BioVid | Accuracy82.46 | 5 |