DEEP GENERATIVE MODELS FOR REAL-TIME SYNTHESIS OF FACIAL MICRO-EXPRESSIONS
loading.default
item.page.date
item.page.authors
item.page.journal-title
item.page.journal-issn
item.page.volume-title
item.page.publisher
Journals Park Publishing
item.page.abstract
The article provides a comprehensive review of state-of-the-art deep generative models capable of synthesizing photorealistic facial micro-expressions in real time (≥60 fps) on consumer-grade and mobile hardware. Four major research directions from 2021–2025 are examined: (1) two-stream and hierarchical GANs augmented with perceptual losses from micro-expression detectors, (2) diffusion models with fine-grained Action Unit (AU) and temporal control, (3) hybrid parametric 3D face models (FLAME/DECA) combined with neural rendering techniques (3D Gaussian Splatting, NeuS2), and (4) long-sequence Video Transformers and Mamba-based architectures. Achieved quality metrics (FID, LPIPS, MERA-F1), inference speed, anatomical plausibility, and temporal consistency are thoroughly analysed. Particular attention is devoted to remaining challenges: cross-identity transfer and personalization, the scarcity of large-scale 4D datasets, and ethical risks posed by next-generation deepfakes. The technologies are shown to be ready for widespread commercial deployment, with a forecast that the gap between macro- and micro-expression synthesis quality will be fully closed between 2026 and 2028.