Optimal Depth Buffer for Low-Cost Graphics Hardware
Abstract
3D applications using hardware depth buffers for visibility testing are confronted with multiple choices of buffer types, sizes and formats. Some of the options are not exposed through 3D API or may be used by the driver without application s knowledge. As a result, it becomes increasingly difficult to select depth buffer optimal for desired balance between performance and precision. In this paper we provide comparative evaluation of depth precision for main depth buffer types with different size and format combinations. Results indicate that integer storage is preferred for some buffer types, while others achieve maximal depth resolution with floating-point format optimized for known scene parameters. We propose to give 3D applications full control of the depth buffer optimization by supporting multiple storage formats with the same buffer size and exposing them in 3D API. In the search for a unified depth buffer solution, we describe new type of the depth buffer and compare it with other options. Complementary floating-point Z buffer is a combination of a reversed-direction Z buffer and an optimal floating-point storage format. Non-linear mapping and storage format compensate each other s effect on the depth precision; as a result, depth errors become significantly less dependent on the eye-space distance, improving depth resolution by the orders of magnitude in comparison with standard Z buffer. Results show that complementary Z buffer is also superior to inverse W buffer at any storage size. At 16 and 24 bits/pixel, average depth errors of complementary Z buffer remain 2 times larger than for true W buffer utilizing expensive high-precision per-pixel division. However, it provides absolutely best precision at 32 bits/pixel, when errors are limited by floating-point per-vertex input. Results suggest that complementary floating-point Z buffer can be considered as a candidate for replacement of both screen Z and inverse W buffers, at the same time making hardware investment in the true W buffer support less attractive.
BibTeX
@inproceedings {10.2312:EGGH:EGGH99:067-074,
booktitle = {SIGGRAPH/Eurographics Workshop on Graphics Hardware},
editor = {A. Kaufmann and W. Strasser and S. Molnar and B.- O. Schneider},
title = {{Optimal Depth Buffer for Low-Cost Graphics Hardware}},
author = {Lapidous, Eugene and Jiao, Guofang},
year = {1999},
publisher = {The Eurographics Association},
ISSN = {1727-3471},
ISBN = {1-58113-170-4},
DOI = {10.2312/EGGH/EGGH99/067-074}
}
booktitle = {SIGGRAPH/Eurographics Workshop on Graphics Hardware},
editor = {A. Kaufmann and W. Strasser and S. Molnar and B.- O. Schneider},
title = {{Optimal Depth Buffer for Low-Cost Graphics Hardware}},
author = {Lapidous, Eugene and Jiao, Guofang},
year = {1999},
publisher = {The Eurographics Association},
ISSN = {1727-3471},
ISBN = {1-58113-170-4},
DOI = {10.2312/EGGH/EGGH99/067-074}
}
Related items
Showing items related by title, author, creator and subject.
-
Time-constrained Animation Rendering on Desktop Grids
Aggarwal, Vibhor; Debattista, Kurt; Bashford-Rogers, Thomas; Chalmers, Alan (The Eurographics Association, 2012)The computationally intensive nature of high-fidelity rendering has led to a dependence on parallel infrastructures for generating animations. However, such an infrastructure is expensive thereby restricting easy access ... -
Texturing and Hypertexturing of Volumetric Objects
Miller, Chris M.; Jones, Mark W. (The Eurographics Association, 2005)Texture mapping is an extremely powerful and flexible tool for adding complex surface detail to an object. This paper introduces a method of surface texturing and hypertexturing complex volumetric objects in real-time. We ... -
Ray Tracing Dynamic Scenes with Shadows on the GPU
Guntury, Sashidhar; Narayanan, P. J. (The Eurographics Association, 2010)We present fast ray tracing of dynamic scenes in this paper with primary and shadow rays. We present a GPUfriendly strategy to bring coherency to shadow rays, based on previous work on grids as acceleration structures. We ...