Transferring Pose and Augmenting Background Variation for Deep Human Image Parsing
Abstract
Human parsing is a fundamental task to estimate semantic parts in a human image such as face, arm, leg, hat, and dress. Recent deep-learning based methods have achieved significant improvements, but collecting training datasets of pixel-wise annotations is labor-intensive. In this paper, we propose two solutions to cope with limited dataset. First, to handle various poses, we incorporate a pose estimation network into an end-to-end human parsing network in order to transfer common features across the domains. The pose estimation network can be trained using rich datasets and feed valuable features to the human parsing network. Second, to handle complicated backgrounds, we increase the variations of background images automatically by replacing the original backgrounds of human images with those obtained from large-scale scenery image datasets. While each of the two solutions is versatile and beneficial to human parsing, their combination yields further improvement.
BibTeX
@inproceedings {10.2312:pg.20171317,
booktitle = {Pacific Graphics Short Papers},
editor = {Jernej Barbic and Wen-Chieh Lin and Olga Sorkine-Hornung},
title = {{Transferring Pose and Augmenting Background Variation for Deep Human Image Parsing}},
author = {Kikuchi, Takazumi and Endo, Yuki and Kanamori, Yoshihiro and Hashimoto, Taisuke and Mitani, Jun},
year = {2017},
publisher = {The Eurographics Association},
ISBN = {978-3-03868-051-2},
DOI = {10.2312/pg.20171317}
}
booktitle = {Pacific Graphics Short Papers},
editor = {Jernej Barbic and Wen-Chieh Lin and Olga Sorkine-Hornung},
title = {{Transferring Pose and Augmenting Background Variation for Deep Human Image Parsing}},
author = {Kikuchi, Takazumi and Endo, Yuki and Kanamori, Yoshihiro and Hashimoto, Taisuke and Mitani, Jun},
year = {2017},
publisher = {The Eurographics Association},
ISBN = {978-3-03868-051-2},
DOI = {10.2312/pg.20171317}
}