Show simple item record

dc.contributor.authorHenzler, Philipp
dc.date.accessioned2023-11-22T08:02:49Z
dc.date.available2023-11-22T08:02:49Z
dc.date.issued2023
dc.identifier.urihttps://diglib.eg.org:443/handle/10.2312/3543899
dc.description.abstractInferring 3D shape and appearance from natural images is a fundamental challenge in computer vision. Despite recent progress using deep learning methods, a key limitation is the availability of annotated training data, as acquisition is often very challenging and expensive, especially at a large scale. This thesis proposes to incorporate physical priors into neural networks that allow for self-supervised learning. As a result, easy-to-access unlabeled data can be used for model training. In particular, novel algorithms in the context of 3D reconstruction and texture/material synthesis are introduced, where only image data is available as supervisory signal. First, a method that learns to reason about 3D shape and appearance solely from unstructured 2D images, achieved via differentiable rendering in an adversarial fashion, is proposed. As shown next, learning from videos significantly improves 3D reconstruction quality. To this end, a novel ray-conditioned warp embedding is proposed that aggregates pixel-wise features from multiple source images. Addressing the challenging task of disentangling shape and appearance, first a method that enables 3D texture synthesis independent of shape or resolution is presented. For this purpose, 3D noise fields of different scales are transformed into stationary textures. The method is able to produce 3D textures, despite only requiring 2D textures for training. Lastly, the surface characteristics of textures under different illumination conditions are modeled in the form of material parameters. Therefore, a self-supervised approach is proposed that has no access to material parameters but only flash images. Similar to the previous method, random noise fields are reshaped to material parameters, which are conditioned to replicate the visual appearance of the input under matching light.en_US
dc.language.isoenen_US
dc.subjectComputer Graphicsen_US
dc.subjectComputer Visionen_US
dc.subjectDeep Learningen_US
dc.subjectMaterial modelingen_US
dc.subjectTexture synthesisen_US
dc.subject3D Reconstructionen_US
dc.titleSelf-Supervised Shape and Appearance Modeling via Neural Differentiable Graphicsen_US
dc.typeThesisen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record