Papers
Topics
Authors
Recent
Search
2000 character limit reached

Cross-Age Expression Transfer

Updated 29 December 2025
  • Cross-age expression transfer is a set of techniques that decouple facial expression from age and identity, enabling precise synthesis across diverse age groups.
  • Key methodologies include geometry-disentangled embeddings, contrastive loss, and adversarial generation to robustly align heterogeneous facial features.
  • Advanced models like GC-GAN and FACE-BE-SELF demonstrate enhanced SSIM, balanced F1 scores, and effective cross-domain performance on adult and child datasets.

Cross-age expression transfer refers to the set of methodologies and systems designed to recognize, synthesize, or adapt facial expressions across individuals of different ages, particularly in the presence of age-induced facial morphology changes. Central challenges include disentangling expression information from age and identity, robustly aligning heterogeneous facial representations, and ensuring semantic continuity of expression manifolds despite significant age variation. State-of-the-art approaches draw on generative adversarial networks, landmark-based geometric embeddings, and cross-domain adaptation techniques to enable precise transfer of affective signals between child and adult faces or across age-progressed imagery.

1. Foundations of Cross-Age Expression Transfer

Cross-age expression transfer is grounded in the observation that facial shape, musculature, and appearance cues evolve substantially from childhood through adulthood, resulting in distributional shift for both landmark-based and texture features. These discrepancies undermine the efficacy of conventional facial expression analysis models trained on a single age group. For example, FEA models trained on adult benchmark datasets such as CK+ are ineffective on child images, and vice versa, due to the "age-domain shift" driven by morphological and psychophysical developmental differences (Witherow et al., 2022).

To address these challenges, modern systems focus on:

  • Geometry-driven latent semantic embeddings.
  • Explicit age-invariant or age-disentangled representation learning.
  • Domain adaptation to harmonize feature distributions across disparate age groups.

2. Geometry-Contrastive Generative Adversarial Network (GC-GAN)

The Geometry-Contrastive Generative Adversarial Network (GC-GAN) aims to decouple expression from both identity and age-driven facial variations for the purpose of expression transfer (Qiao et al., 2018).

Key Components:

  • Landmark Representation: 68 2D facial landmarks are concatenated into a 136-dimensional vector gR136g \in \mathbb{R}^{136}, normalized to [1,1][-1,1].
  • Semantic Embedding: An auto-encoder E=(Eenc,Edec)E=(E_{\text{enc}},\,E_{\text{dec}}) compresses gg into a 32-D embedding zgz_g.
  • Contrastive Loss: The contrastive loss

Lcontr=α12max(0,mEenc(gkv)Eenc(gref)2)+(1α)12Eenc(gkv)Eenc(gref)2L_{\text{contr}} = \alpha \cdot \frac{1}{2} \max(0, m - \|E_{\text{enc}}(g_k^v) - E_{\text{enc}}(g_{\text{ref}})\|^2) + (1 - \alpha)\cdot \frac{1}{2}\|E_{\text{enc}}(g_k^v) - E_{\text{enc}}(g_{\text{ref}})\|^2

clusters embeddings by expression and separates different expressions, removing subject-specific geometry.

  • Adversarial Generation: The generator G=(Genc,Gdec)G=(G_{\text{enc}},\,G_{\text{dec}}) fuses appearance encoding ziz_i and geometric encoding zgz_g to synthesize an expression-transferred image I~jv\tilde{I}_j^v.
  • Training Loss: Combines adversarial, image reconstruction, and landmark reconstruction losses; pretrained embedding network is fixed during GAN training.

Handling Cross-Age Variation:

Although the core GC-GAN is subject-agnostic, cross-age expression transfer is enabled by further modifications:

  • Incorporating age-sensitive shape descriptors to gg.
  • Augmenting the embedding process with an adversarial age classifier to enforce age-invariance in zgz_g.
  • Disentangling age and expression via orthogonal embedding branches, enabling z=(zid,zg,zage)z = (z_{\text{id}}, z_g, z_{\text{age}}) to control both factors independently.

Quantitative Evaluation

Model Variant Multi-PIE SSIM CK+ SSIM BU-4DFE SSIM
Full GC-GAN 0.687 0.769 0.725
w/o LcontrL_{\text{contr}} 0.675 0.763 0.705
CDAAE (one-hot) 0.669 0.765 0.710

GC-GAN maintains high identity preservation (face-ID accuracy 0.977, expression sim 0.896) even as target expression geometry is drawn from subjects with markedly different shapes or ages. Embedding visualizations show that contrastive supervision yields semantically separated expression clusters, ensuring robust transfer across shape, pose, and implied age domains (Qiao et al., 2018).

3. Deep Adaptation Across Age: FACE-BE-SELF

The FACE-BE-SELF architecture extends cross-age expression analysis by introducing domain adaptation tailored explicitly for adult–child expression alignment (Witherow et al., 2022).

Architectural Features:

  • Dual-Stream Embedding: Both adult and child streams use a unified feature extractor M()M(\cdot), combining:
    • CNN-based appearance feature (zGR512z_G \in \mathbb{R}^{512}).
    • MLP-based geometric feature (zHR512z_H \in \mathbb{R}^{512}) using a selected landmark subset.
  • Landmark Feature Decomposition:
    • Extract all Euclidean landmark distances and Delaunay triangle features.
    • Characterize feature-factor associations using pairwise sample correlations.
    • Employ a Beta-mixture model (learned via EM) to select a maximal subset of features (Fexpr\mathcal{F}_{\text{expr}}) highly correlated with expression but not with age domain or subject identity.
  • Contrastive Domain-Alignment Loss:

    LDA=i,j[1(yi=yj)ziszjt22+1(yiyj)max(0,mziszjt2)2]L_{\text{DA}} = \sum_{i,j} \left[ \mathbf{1}(y_i = y_j) \|z_i^s - z_j^t\|_2^2 + \mathbf{1}(y_i \neq y_j) \max(0, m - \|z_i^s - z_j^t\|_2)^2 \right]

    aligns adult and child latent distributions for samples of the same expression class, ensuring cross-domain coherence.

Key Results

Model Adult F1 Child F1 Δ\Delta
Source-only CNN 0.707±0.064 0.322
Mix fine-tune 0.818 0.439 0.379
Best prior DA [38] 0.753 0.065
FACE-BE-SELF 0.8443 0.8303 0.014

FACE-BE-SELF yields near-equal adult/child performance on posed data and superior robustness to domain shift compared to transfer learning or prior DA baselines. On spontaneous expressions, it provides balanced F1 and AUC, highlighting the efficacy of data-driven, age-aware geometric feature selection and fusion with deep appearance cues (Witherow et al., 2022).

4. Evaluation Protocols and Datasets

State-of-the-art evaluation involves comprehensive within- and cross-domain protocols:

  • Data Splits: Nested k-fold cross-validation (e.g., 5×2), with subject-exclusive splits to avoid overlap.
  • Datasets:
    • CK+ (adult, posed)
    • CAFE (children 2–8 yrs, posed)
    • Aff-Wild2 (adult, spontaneous)
    • ChildEFES (children 4–6 yrs, spontaneous)
  • Metrics: F1 score and ROC-AUC for expression recognition per age group; SSIM, PSNR, and face-ID accuracy for generative models.
  • Unpaired and Cycle-Consistency Contexts: For datasets lacking true cross-age expression pairs, unpaired image-to-image methods or cycle losses may be leveraged to ensure semantic reversibility and transfer fidelity (Qiao et al., 2018).

5. Methodological Extensions for Cross-Age Transfer

Recent architectures propose several enhancement strategies for robust cross-age transfer:

  • Expanded Landmark Sets: Inclusion of 3D landmarks or age-sensitive facial ratios (e.g., nasolabial depth, brow–eye distance).
  • Disentanglement Branches: Orthogonal latent spaces for age and expression, maintained via covariance penalties or adversarial age-removal heads.
  • Multi-Scale Conditional Synthesis: Conditional normalization layers such as AdaIN modulated by both expression (zgz_g) and age (zagez_{\text{age}}) embeddings.
  • Fine-Scale Texture Synthesis: Additional "detail networks" to reconstruct age-related features (wrinkles, tone) at higher resolutions (Qiao et al., 2018).
  • Feature Selection via Mixture Modelling: Posterior-based selection of expression-relevant geometric features via Beta-mixture EM substrates, minimizing redundancy and confounding factors (Witherow et al., 2022).

6. Limitations and Prospects

Despite substantial progress, two main challenges persist:

  • Disentanglement Complexity: Achieving strict age-invariance in expression manifolds is nontrivial, and residual age cues may still influence embedding geometry. Adversarial training and explicit cycle-consistency terms are promising, but their optimal configuration remains active research.
  • Data Scarcity for Intermediate Ages: Both GC-GAN and FACE-BE-SELF focus on adults and young children; comprehensive datasets for teenagers or elderly faces are rare, impeding robust transfer across the full age spectrum.

Future extensions include adversarial alignment with maximum-mean-discrepancy losses, explicit temporal modeling in video sequences, and expansion to multi-modal (audio-visual) cross-age affect synthesis (Witherow et al., 2022).


Cross-age expression transfer thus comprises a fast-evolving intersection of face analysis, generative modeling, and domain adaptation, yielding architectures that factor and recombine geometry, appearance, and demographic cues to enable robust, semantically valid facial affect transfer across diverse age domains (Qiao et al., 2018, Witherow et al., 2022).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (2)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Cross-Age Expression Transfer.