We release a new benchmark, named MultiREX (Multiface Region-based Expression evaluation), that evaluates the estimated geometry of monocular face capture systems considering complex expression sequences under multiple camera views. In particular, the protocol evaluates mesh deformations related to expression alone, treating the identity as a given.
The benchmark is based on the Multiface dataset, and includes 8 identities captured simultaneously from five viewpoints: Frontal, two Angled views
(yaw rotation around 40 degrees), and two Profile views (yaw rotation around 60 degrees). Each subject performs a range-of-motion
sequence covering a wide range of expressions, including extreme and asymmetrical motions.
The benchmark comprises 10k ground truth meshes and 49k images.
We obtain the ground truth identity (i.e., neutral mesh) by manually selecting a neutral frame for each subject and retopologizing the corresponding mesh to the
FLAME topology using the
Wrap 3D commercial software. From these two meshes, we compute a per-subject sparse conversion matrix that enables fast conversion from the FLAME to the Multiface topology and later quantitative evaluation.
The benchmark is being made public (coming soon), and we release: (i) the code to download assets,
(ii) neutral meshes in the FLAME topology alongside the code to convert between FLAME and Multiface topologies,
(iii) code to run the benchmark and compute the metrics.