Progressive Multi-Source Domain Adaptation for Personalized Facial Expression Recognition (P-MSDA)
2025
- IEEE Transactions on Affective Computing, 2025
Bringing Personalization to Facial Expression Recognition
Based on our publication in *IEEE Transactions on Affective Computing
π Problem Overview
Multi-source domain adaptation (MSDA) traditionally assumes using all available subjects as sources improves performance.
Figure 1 (a) illustrates:
- Large domain gaps between many source subjects and the target.
- Combining all sources introduces noise and negative transfer.
- Misaligned subjects confuse the model rather than helping it adapt.
Many source subjects lie far from the target in feature space β hurting adaptation.
This motivates our approach Figure 1 (b): β‘ Start with the most similar subjects β gradually introduce more challenging ones.
π‘ Our Proposed Method: Progressive MSDA (P-MSDA)
Our method consists of two major components:
1οΈβ£ Source Selection Using Similarity Ranking
We compute cosine similarity between the target subject and each source subject.
Sources are then ranked from closest β farthest.
This process is illustrated below:
Demonstrated how the model identifies subjects with the most similar expression distributions and facial structures, selecting them as the initial steps in the adaptation process.
These closest subjects form the basis for our curriculum:
- Step 1 β adapt to the closest subject
- Step 2 β adapt to the second closest
- β¦
- Step n β adapt to more diverse subjects
This progression greatly reduces early domain shift.
2οΈβ£ Progressive Domain Adaptation Pipeline
The complete workflow of P-MSDA:
Key components:
β Progressive Curriculum
Subjects are introduced one at a time (easy β hard).
β Density-Based Replay Memory
At each stage, representative samples are stored based on cluster density β preventing forgetting.
β Multi-domain Alignment
Using MMD-based discrepancy loss, the model aligns:
- Current source
- Replay samples
- Target domain
This ensures stable learning at every stage.
π Results Across Benchmark Datasets
Our approach outperforms all MSDA and UDA baselines on:
- BioVid (Heat Pain database)
- UNBC-McMaster
- Aff-Wild2
- Behavioural Ambivalence/Hesitancy (BAH) dataset
Highlights:
- BioVid: 88% accuracy
- UNBC-McMaster: 0.88 accuracy
- Aff-Wild2: 0.46
- BAH: 0.71
- Cross-Dataset (UNBC β BioVid): 0.78
π― Why Progressive Learning Works
Ablations confirm:
- Random subject ordering performs significantly worse
- Density-based replay selection boosts stability
- Progressive alignment yields compact target-specific clusters
- Training efficiency stays high (only 3 domains used at once)
β¨ Takeaway
P-MSDA enables robust, label-free personalization for FER systems:
- π Personalized FER without labeled target data
- π Scalable multi-source adaptation
- β Avoids negative transfer
- π Achieves state-of-the-art results
π Full Paper
Title: Progressive Multi-Source Domain Adaptation for Personalized Facial Expression Recognition
Authors: Muhammad Osama Zeeshan, Marco Pedersoli, Alessandro L. Koerich, Eric Granger
Venue: IEEE Transactions on Affective Computing
π Read the full paper here: IEEE Xplore Link
π» Code: Github P-MSDA
π¬ Contact
Feel free to reach out for discussion or collaboration!