Show simple item record

dc.contributor.authorMallick, Arijiten_US
dc.contributor.authorEngelhardt, Andreasen_US
dc.contributor.authorBraun, Raphaelen_US
dc.contributor.authorLensch, Hendrik P. A.en_US
dc.contributor.editorBender, Janen_US
dc.contributor.editorBotsch, Marioen_US
dc.contributor.editorKeim, Daniel A.en_US
dc.date.accessioned2022-09-26T09:28:37Z
dc.date.available2022-09-26T09:28:37Z
dc.date.issued2022
dc.identifier.isbn978-3-03868-189-2
dc.identifier.urihttps://doi.org/10.2312/vmv.20221197
dc.identifier.urihttps://diglib.eg.org:443/handle/10.2312/vmv20221197
dc.description.abstractImage super resolution is a classical computer vision problem. A branch of super resolution tasks deals with guided depth super resolution as objective. Here, the goal is to accurately upsample a given low resolution depth map with the help of features aggregated from the high resolution color image of that particular scene. Recently, the development of transformers has improved performance for general image processing tasks credited to self-attention. Unlike previous methods for guided joint depth upsampling which rely mostly on CNNs, we efficiently compute self-attention with the help of local image attention which avoids the quadratic growth typically found in self-attention layers. Our work combines CNNs and transformers to analyze the two input modalities and employs a cross-modal fusion network in order to predict both a weighted per-pixel filter kernel and a residual for the depth estimation. To further enhance the final output, we integrate a differentiable and a trainable deep guided filtering network which provides an additional depth prior. An ablation study and empirical trials demonstrate the importance of each proposed module. Our method shows comparable as well as state-of-the-art performance on the guided depth upsampling task.en_US
dc.publisherThe Eurographics Associationen_US
dc.rightsAttribution 4.0 International License
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/
dc.subjectCCS Concepts: Computing methodologies --> Computer vision; Image representations; Reconstruction
dc.subjectComputing methodologies
dc.subjectComputer vision
dc.subjectImage representations
dc.subjectReconstruction
dc.titleLocal Attention Guided Joint Depth Upsamplingen_US
dc.description.seriesinformationVision, Modeling, and Visualization
dc.description.sectionheadersJoint Session
dc.identifier.doi10.2312/vmv.20221197
dc.identifier.pages1-8
dc.identifier.pages8 pages


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

Attribution 4.0 International License
Except where otherwise noted, this item's license is described as Attribution 4.0 International License