Reconstructing Animated Meshes from Time-Varying Point Clouds
Abstract
In this paper, we describe a novel approach for the reconstruction of animated meshes from a series of time-deforming point clouds. Given a set of unordered point clouds that have been captured by a fast 3-D scanner, our algorithm is able to compute coherent meshes which approximate the input data at arbitrary time instances. Our method is based on the computation of an implicit function in R?4 that approximates the time-space surface of the time-varying point cloud. We then use the four-dimensional implicit function to reconstruct a polygonal model for the first time-step. By sliding this template mesh along the time-space surface in an as-rigid-as-possible manner, we obtain reconstructions for further time-steps which have the same connectivity as the previously extracted mesh while recovering rigid motion exactly.The resulting animated meshes allow accurate motion tracking of arbitrary points and are well suited for animation compression. We demonstrate the qualities of the proposed method by applying it to several data sets acquired by real-time 3-D scanners.
BibTeX
@article {10.1111:j.1467-8659.2008.01287.x,
journal = {Computer Graphics Forum},
title = {{Reconstructing Animated Meshes from Time-Varying Point Clouds}},
author = {Suessmuth, Jochen and Winter, Marco and Greiner, Guenther},
year = {2008},
publisher = {The Eurographics Association and Blackwell Publishing Ltd},
ISSN = {1467-8659},
DOI = {10.1111/j.1467-8659.2008.01287.x}
}
journal = {Computer Graphics Forum},
title = {{Reconstructing Animated Meshes from Time-Varying Point Clouds}},
author = {Suessmuth, Jochen and Winter, Marco and Greiner, Guenther},
year = {2008},
publisher = {The Eurographics Association and Blackwell Publishing Ltd},
ISSN = {1467-8659},
DOI = {10.1111/j.1467-8659.2008.01287.x}
}