Contact and Human Dynamics from Monocular Video
Date
2020Metadata
Show full item recordAbstract
Existing methods for human motion from video predict 2D and 3D poses that are approximately accurate, but contain visible errors that violate physical constraints, such as feet penetrating the ground and bodies leaning at extreme angles. We present a physics-based method for inferring 3D human motion from video sequences that takes initial 2D and 3D pose estimates as input. We first estimate ground contact timings with a neural network which is trained without hand-labeled data. A physicsbased trajectory optimization then solves for a physically-plausible motion, based on the inputs. We show this process produces motions that are more realistic than those from purely kinematic methods for character animation from dynamic videos. A detailed report that fully describes our method is available at geometry.stanford.edu/projects/human-dynamics-eccv-2020.
BibTeX
@inproceedings {10.2312:sca.20201218,
booktitle = {Eurographics/ ACM SIGGRAPH Symposium on Computer Animation - Showcases},
editor = {Holden, Daniel},
title = {{Contact and Human Dynamics from Monocular Video}},
author = {Rempe, Davis and Guibas, Leonidas J. and Hertzmann, Aaron and Russell, Bryan and Villegas, Ruben and Yang, Jimei},
year = {2020},
publisher = {The Eurographics Association},
ISSN = {1727-5288},
ISBN = {978-3-03868-129-8},
DOI = {10.2312/sca.20201218}
}
booktitle = {Eurographics/ ACM SIGGRAPH Symposium on Computer Animation - Showcases},
editor = {Holden, Daniel},
title = {{Contact and Human Dynamics from Monocular Video}},
author = {Rempe, Davis and Guibas, Leonidas J. and Hertzmann, Aaron and Russell, Bryan and Villegas, Ruben and Yang, Jimei},
year = {2020},
publisher = {The Eurographics Association},
ISSN = {1727-5288},
ISBN = {978-3-03868-129-8},
DOI = {10.2312/sca.20201218}
}