Temporally Consistent Wide Baseline Facial Performance Capture via Image Warping
Date
2015Metadata
Show full item recordAbstract
In this paper, we present a method for detailed temporally consistent facial performance capture that supports any number of arbitrarily placed video cameras. Using a suitable 3D model as reference geometry, our method tracks facial movement and deformation as well as photometric changes due to illumination and shadows. In an analysis-by-synthesis framework, we warp one single reference image per camera to all frames of the sequence thereby drastically reducing temporal drift which is a serious problem for many state-of-the-art approaches. Temporal appearance variations are handled by a photometric estimation component modeling local intensity changes between the reference image and each individual frame. All parameters of the problem are estimated jointly so that we do not require separate estimation steps that might interfere with one another.
BibTeX
@inproceedings {10.2312:vmv.20151263,
booktitle = {Vision, Modeling & Visualization},
editor = {David Bommes and Tobias Ritschel and Thomas Schultz},
title = {{Temporally Consistent Wide Baseline Facial Performance Capture via Image Warping}},
author = {Kettern, Markus and Hilsmann, Anna and Eisert, Peter},
year = {2015},
publisher = {The Eurographics Association},
ISBN = {978-3-905674-95-8},
DOI = {10.2312/vmv.20151263}
}
booktitle = {Vision, Modeling & Visualization},
editor = {David Bommes and Tobias Ritschel and Thomas Schultz},
title = {{Temporally Consistent Wide Baseline Facial Performance Capture via Image Warping}},
author = {Kettern, Markus and Hilsmann, Anna and Eisert, Peter},
year = {2015},
publisher = {The Eurographics Association},
ISBN = {978-3-905674-95-8},
DOI = {10.2312/vmv.20151263}
}