dc.contributor.author | Yang, Dongseok | en_US |
dc.contributor.author | Kang, Jiho | en_US |
dc.contributor.author | Ma, Lingni | en_US |
dc.contributor.author | Greer, Joseph | en_US |
dc.contributor.author | Ye, Yuting | en_US |
dc.contributor.author | Lee, Sung-Hee | en_US |
dc.contributor.editor | Bermano, Amit H. | en_US |
dc.contributor.editor | Kalogerakis, Evangelos | en_US |
dc.date.accessioned | 2024-04-16T14:43:26Z | |
dc.date.available | 2024-04-16T14:43:26Z | |
dc.date.issued | 2024 | |
dc.identifier.issn | 1467-8659 | |
dc.identifier.uri | https://doi.org/10.1111/cgf.15057 | |
dc.identifier.uri | https://diglib.eg.org:443/handle/10.1111/cgf15057 | |
dc.description.abstract | Full-body avatar presence is important for immersive social and environmental interactions in digital reality. However, current devices only provide three six degrees of freedom (DOF) poses from the headset and two controllers (i.e. three-point trackers). Because it is a highly under-constrained problem, inferring full-body pose from these inputs is challenging, especially when supporting the full range of body proportions and use cases represented by the general population. In this paper, we propose a deep learning framework, DivaTrack, which outperforms existing methods when applied to diverse body sizes and activities. We augment the sparse three-point inputs with linear accelerations from Inertial Measurement Units (IMU) to improve foot contact prediction. We then condition the otherwise ambiguous lower-body pose with the predictions of foot contact and upper-body pose in a two-stage model. We further stabilize the inferred full-body pose in a wide range of configurations by learning to blend predictions that are computed in two reference frames, each of which is designed for different types of motions. We demonstrate the effectiveness of our design on a large dataset that captures 22 subjects performing challenging locomotion for three-point tracking, including lunges, hula-hooping, and sitting. As shown in a live demo using the Meta VR headset and Xsens IMUs, our method runs in real-time while accurately tracking a user's motion when they perform a diverse set of movements. | en_US |
dc.publisher | The Eurographics Association and John Wiley & Sons Ltd. | en_US |
dc.rights | Attribution 4.0 International License | |
dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | |
dc.subject | CCS Concepts: Computing methodologies -> Motion capture | |
dc.subject | Computing methodologies | |
dc.subject | Motion capture | |
dc.title | DivaTrack: Diverse Bodies and Motions from Acceleration-Enhanced 3-Point Trackers | en_US |
dc.description.seriesinformation | Computer Graphics Forum | |
dc.description.sectionheaders | Camera Paths and Motion Tracking | |
dc.description.volume | 43 | |
dc.description.number | 2 | |
dc.identifier.doi | 10.1111/cgf.15057 | |
dc.identifier.pages | 13 pages | |