|
|
|
|
LEADER |
01000naa a22002652 4500 |
001 |
NLM372693962 |
003 |
DE-627 |
005 |
20240524233207.0 |
007 |
cr uuu---uuuuu |
008 |
240524s2024 xx |||||o 00| ||eng c |
024 |
7 |
|
|a 10.1109/TVCG.2024.3403160
|2 doi
|
028 |
5 |
2 |
|a pubmed24n1417.xml
|
035 |
|
|
|a (DE-627)NLM372693962
|
035 |
|
|
|a (NLM)38781062
|
040 |
|
|
|a DE-627
|b ger
|c DE-627
|e rakwb
|
041 |
|
|
|a eng
|
100 |
1 |
|
|a Qu, Linzi
|e verfasserin
|4 aut
|
245 |
1 |
0 |
|a Sketch2Human
|b Deep Human Generation with Disentangled Geometry and Appearance Constraints
|
264 |
|
1 |
|c 2024
|
336 |
|
|
|a Text
|b txt
|2 rdacontent
|
337 |
|
|
|a ƒaComputermedien
|b c
|2 rdamedia
|
338 |
|
|
|a ƒa Online-Ressource
|b cr
|2 rdacarrier
|
500 |
|
|
|a Date Revised 24.05.2024
|
500 |
|
|
|a published: Print-Electronic
|
500 |
|
|
|a Citation Status Publisher
|
520 |
|
|
|a Geometry- and appearance-controlled full-body human image generation is an interesting but challenging task. Existing solutions are either unconditional or dependent on coarse conditions (e.g., pose, text), thus lacking explicit geometry and appearance control of body and garment. Sketching offers such editing ability and has been adopted in various sketch-based face generation and editing solutions. However, directly adapting sketch-based face generation to full-body generation often fails to produce high-fidelity and diverse results due to the high complexity and diversity in the pose, body shape, and garment shape and texture. Recent geometrically controllable diffusion-based methods mainly rely on prompts to generate appearance. It is hard to balance the realism and the faithfulness of their results to the sketch when the input is coarse. This work presents Sketch2Human, the first system for controllable full-body human image generation guided by a semantic sketch (for geometry control) and a reference image (for appearance control). Our solution is based on the latent space of StyleGAN-Human with inverted geometry and appearance latent codes as input. Specifically, we present a sketch encoder trained with a large synthetic dataset sampled from StyleGAN-Human's latent space and directly supervised by sketches rather than real images. Considering the entangled information of partial geometry and texture in StyleGAN-Human and the absence of disentangled datasets, we design a novel training scheme that creates geometry-preserved and appearance-transferred training data to tune a generator to achieve disentangled geometry and appearance control. Although our method is trained with synthetic data, it can also handle hand-drawn sketches. Qualitative and quantitative evaluations demonstrate the superior performance of our method to state-of-the-art methods
|
650 |
|
4 |
|a Journal Article
|
700 |
1 |
|
|a Shang, Jiaxiang
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Ye, Hui
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Han, Xiaoguang
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Fu, Hongbo
|e verfasserin
|4 aut
|
773 |
0 |
8 |
|i Enthalten in
|t IEEE transactions on visualization and computer graphics
|d 1996
|g PP(2024) vom: 23. Mai
|w (DE-627)NLM098269445
|x 1941-0506
|7 nnns
|
773 |
1 |
8 |
|g volume:PP
|g year:2024
|g day:23
|g month:05
|
856 |
4 |
0 |
|u http://dx.doi.org/10.1109/TVCG.2024.3403160
|3 Volltext
|
912 |
|
|
|a GBV_USEFLAG_A
|
912 |
|
|
|a SYSFLAG_A
|
912 |
|
|
|a GBV_NLM
|
912 |
|
|
|a GBV_ILN_350
|
951 |
|
|
|a AR
|
952 |
|
|
|d PP
|j 2024
|b 23
|c 05
|