Layered-Garment Net: Generating Multiple Implicit Garment Layers from a Single Image

Published in 16th Asian Conference on Computer Vision (ACCV2022), 2022

Recommended citation: Aggarwal, A., Wang, J., Hogue, S., Ni, S., Budagavi, M., & Guo, X. (2022). Layered-Garment Net: Generating Multiple Implicit Garment Layers from a Single Image. In Proceedings of the Asian Conference on Computer Vision (pp. 3000-3017). https://openaccess.thecvf.com/content/ACCV2022/papers/Aggarwal_Layered-Garment_Net_Generating_Multiple_Implicit_Garment_Layers_from_a_Single_ACCV_2022_paper.pdf

Abstract

Recent research works have focused on generating human models and garments from their 2D images. However, state-of-the-art researches focus either on only a single layer of the garment on a human model or on generating multiple garment layers without any guarantee of the intersection-free geometric relationship between them. In reality, people wear multiple layers of garments in their daily life, where an inner layer of garment could be partially covered by an outer one. In this paper, we try to address this multi-layer modeling problem and propose the Layered-Garment Net (LGN) that is capable of generating intersection-free multiple layers of garments defined by implicit function fields over the body surface, given the person’s near front-view image. With a special design of garment indication fields (GIF), we can enforce an implicit covering relationship between the signed distance fields (SDF) of different layers to avoid self-intersections among different garment surfaces and the human body. Experiments demonstrate the strength of our proposed LGN framework in generating multi-layer garments as compared to state-of-the-art methods. To the best of our knowledge, LGN is the first research work to generate intersection-free multiple layers of garments on the human body from a single image.

Direct Link