Bridging the gap between mimics and muscles: Our method EIFER utilizes neural unpaired image-to-image translation to decouple facial geometry and appearance for muscle-activity-based expression synthesis and electrode-free facial electromyography.
Bridging the gap between mimics and muscles: Our method EIFER utilizes neural unpaired image-to-image translation to decouple facial geometry and appearance for muscle-activity-based expression synthesis and electrode-free facial electromyography.
EIFER (Electromyography-Informed Facial Expression Reconstruction) introduces a novel approach for facial expression analysis that addresses the challenges posed by surface electromyography (sEMG) electrode occlusion. Our method is grounded in the principle of decoupling facial geometry and appearance, achieved through a neural unpaired image-to-image translation framework, see the above figure. EIFER leverages 3D Morphable Models (3DMMs), specifically FLAME , to provide a parametric representation of facial shape and expression. We see the 3DMM expression space and muscle activity as the missing link between computer animation and physiologically-grounded facial expression analysis.
For monocular 3D face reconstruction, EIFER employs neural differential rendering and utilizes pre-trained SMIRK encoder networks. These encoders, consisting of sub-encoders based on MobileNetV3 backbones, estimate 3DMM parameters (shape, expression, pose) from the sEMG electrode occluded input facial images. The core of EIFER's occlusion handling lies in a CycleGAN-like adversarial architecture for unpaired image-to-image translation. This architecture comprises generator and discriminator networks trained adversarially using unpaired sets of sEMG-occluded and occlusion-free (reference) facial images. Crucially, EIFER learns a bidirectional mapping between the 3DMM expression parameter space and measured sEMG muscle activity through multi-layer perceptron (MLPs). This bidirectional mapping enables both the synthesis of facial expressions from muscle activity and the prediction of muscle activity from observed facial expressions, effectively achieving electrode-free facial electromyography.
EIFER combines many ideas from recent works in computer vision, graphics, and machine learning. Here are some related works that you might find interesting:
This publication is a joint effort of the Computer Vision Group at Friedrich Schiller University Jenaand the Department of Otorhinolaryngology at Jena University Hospital. The project Bridging the gap: Mimics and Muscles was funded by the German Science Foundation under the grant number DFG DE-735/15-1 and DFG GU-463/12-1. We would like to thank all participants who volunteered for the data collection and the reviewers for their valuable feedback.
We thank Niklas Penzel, Gideon Stein, Sai Vemuri, Sven Sickert, and Maha Shadaydeh for their manuscript feedback and advice throughout the project. Additionally, we would like to thank Nadiya Müller, Vanessa Tretzsch, Martin Heinrich, Anna-Maria Kuttenreich, Christian Dobel, Gerd Fabian Volk, Roland Graßme, Paul Funk, and Richard Schneider for their support. Furthermore, we thank Bernhard Egger highlighting the potential of our data for facial expression synthesis.
@article{buchner2025electromyography,
title={Electromyography-Informed Facial Expression Reconstruction for Physiological-Based Synthesis and Analysis},
author={B{\"u}chner, Tim and Anders, Christoph and Guntinas-Lichius, Orlando and Denzler, Joachim},
journal={arXiv preprint arXiv:2503.09556},
year={2025}
}