DEEP GENERATIVE MODELS FOR REAL-TIME SYNTHESIS OF FACIAL MICRO-EXPRESSIONS

loading.default
thumbnail.default.alt

item.page.date

item.page.journal-title

item.page.journal-issn

item.page.volume-title

item.page.publisher

Journals Park Publishing

item.page.abstract

The article provides a comprehensive review of state-of-the-art deep generative models capable of synthesizing photorealistic facial micro-expressions in real time (≥60 fps) on consumer-grade and mobile hardware. Four major research directions from 2021–2025 are examined: (1) two-stream and hierarchical GANs augmented with perceptual losses from micro-expression detectors, (2) diffusion models with fine-grained Action Unit (AU) and temporal control, (3) hybrid parametric 3D face models (FLAME/DECA) combined with neural rendering techniques (3D Gaussian Splatting, NeuS2), and (4) long-sequence Video Transformers and Mamba-based architectures. Achieved quality metrics (FID, LPIPS, MERA-F1), inference speed, anatomical plausibility, and temporal consistency are thoroughly analysed. Particular attention is devoted to remaining challenges: cross-identity transfer and personalization, the scarcity of large-scale 4D datasets, and ethical risks posed by next-generation deepfakes. The technologies are shown to be ready for widespread commercial deployment, with a forecast that the gap between macro- and micro-expression synthesis quality will be fully closed between 2026 and 2028.

item.page.description

item.page.citation

item.page.collections

item.page.endorsement

item.page.review

item.page.supplemented

item.page.referenced