Show simple item record

dc.contributor.authorYang, Jinfanen_US
dc.contributor.authorVining, Nicholasen_US
dc.contributor.authorKheradmand, Shakibaen_US
dc.contributor.authorCarr, Nathanen_US
dc.contributor.authorSigal, Leoniden_US
dc.contributor.authorSheffer, Allaen_US
dc.contributor.editorMyszkowski, Karolen_US
dc.contributor.editorNiessner, Matthiasen_US
dc.date.accessioned2023-05-03T06:09:30Z
dc.date.available2023-05-03T06:09:30Z
dc.date.issued2023
dc.identifier.issn1467-8659
dc.identifier.urihttps://doi.org/10.1111/cgf.14744
dc.identifier.urihttps://diglib.eg.org:443/handle/10.1111/cgf14744
dc.description.abstractArtist generated clip-art images typically consist of a small number of distinct, uniformly colored regions with clear boundaries. Legacy artist created images are often stored in low-resolution (100x100px or less) anti-aliased raster form. Compared to anti-aliasing free rasterization, anti-aliasing blurs inter-region boundaries and obscures the artist's intended region topology and color palette; at the same time, it better preserves subpixel details. Recovering the underlying artist-intended images from their low-resolution anti-aliased rasterizations can facilitate resolution independent rendering, lossless vectorization, and other image processing applications. Unfortunately, while human observers can mentally deblur these low-resolution images and reconstruct region topology, color and subpixel details, existing algorithms applicable to this task fail to produce outputs consistent with human expectations when presented with such images. We recover these viewer perceived blur-free images at subpixel resolution, producing outputs where each input pixel is replaced by four corresponding (sub)pixels. Performing this task requires computing the size of the output image color palette, generating the palette itself, and associating each pixel in the output with one of the colors in the palette. We obtain these desired output components by leveraging a combination of perceptual and domain priors, and real world data. We use readily available data to train a network that predicts, for each antialiased image, a low-blur approximation of the blur-free double-resolution outputs we seek. The images obtained at this stage are perceptually closer to the desired outputs but typically still have hundreds of redundant differently colored regions with fuzzy boundaries. We convert these low-blur intermediate images into blur-free outputs consistent with viewer expectations using a discrete partitioning procedure guided by the characteristic properties of clip-art images, observations about the antialiasing process, and human perception of anti-aliased clip-art. This step dramatically reduces the size of the output color palettes, and the region counts bringing them in line with viewer expectations and enabling the image processing applications we target. We demonstrate the utility of our method by using our outputs for a number of image processing tasks, and validate it via extensive comparisons to prior art. In our comparative study, participants preferred our deblurred outputs over those produced by the best-performing alternative by a ratio of 75 to 8.5.en_US
dc.publisherThe Eurographics Association and John Wiley & Sons Ltd.en_US
dc.rightsAttribution 4.0 International License
dc.rights.urihttps://creativecommons.org/licenses/by-nc-nd/4.0/
dc.titleSubpixel Deblurring of Anti-Aliased Raster Clip-Arten_US
dc.description.seriesinformationComputer Graphics Forum
dc.description.sectionheadersLogos and Clip-Art
dc.description.volume42
dc.description.number2
dc.identifier.doi10.1111/cgf.14744
dc.identifier.pages61-76
dc.identifier.pages16 pages


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record

Attribution 4.0 International License
Except where otherwise noted, this item's license is described as Attribution 4.0 International License