|
|
|
|
LEADER |
01000naa a22002652 4500 |
001 |
NLM377232513 |
003 |
DE-627 |
005 |
20240906233543.0 |
007 |
cr uuu---uuuuu |
008 |
240906s2024 xx |||||o 00| ||eng c |
024 |
7 |
|
|a 10.1109/TVCG.2024.3454467
|2 doi
|
028 |
5 |
2 |
|a pubmed24n1525.xml
|
035 |
|
|
|a (DE-627)NLM377232513
|
035 |
|
|
|a (NLM)39236131
|
040 |
|
|
|a DE-627
|b ger
|c DE-627
|e rakwb
|
041 |
|
|
|a eng
|
100 |
1 |
|
|a Li, Wensheng
|e verfasserin
|4 aut
|
245 |
1 |
0 |
|a Efficient Integration of Neural Representations for Dynamic Humans
|
264 |
|
1 |
|c 2024
|
336 |
|
|
|a Text
|b txt
|2 rdacontent
|
337 |
|
|
|a ƒaComputermedien
|b c
|2 rdamedia
|
338 |
|
|
|a ƒa Online-Ressource
|b cr
|2 rdacarrier
|
500 |
|
|
|a Date Revised 06.09.2024
|
500 |
|
|
|a published: Print-Electronic
|
500 |
|
|
|a Citation Status Publisher
|
520 |
|
|
|a While numerous studies have explored NeRF-based novel view synthesis for dynamic humans, they often require training that exceeds several hours, limiting their practicality. Efforts to improve training efficiency have also encountered challenges because it is hard to optimize non-rigid transformations, thus leading to coarse renderings. In this work, we introduce an innovative approach for efficiently learning and integrating neural human representations. To achieve this, we propose a comprehensive utilization of the features stored in both canonical and observational spaces, facilitated through a collaborative refinement process that integrates canonical representations with observational details. Specifically, we initially propose decomposing high-dimensional multi-space feature volume into several feature planes, subsequently utilizing matrix multiplication to explicitly establish the correlations between different planes. This enables the simultaneous optimization of their counterparts across all dimensions by optimizing interpolated features, efficiently integrating associated details, and accelerating the rate of convergence. Additionally, we use the proposed collaborative refinement process to iteratively enhance the canonical representation. By integrating multi-space representations, we further facilitate the co-optimization of multiple frames' time-dependent observations. Experiments demonstrate that our method can achieve high-quality free-viewpoint renderings within nearly 5 minutes of optimization. Compared to state-of-the-art approaches, our results show more realistic rendering details, marking a significant advancement in both performance and efficiency
|
650 |
|
4 |
|a Journal Article
|
700 |
1 |
|
|a Zeng, Lingzhe
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Gao, Chengying
|e verfasserin
|4 aut
|
700 |
1 |
|
|a Liu, Ning
|e verfasserin
|4 aut
|
773 |
0 |
8 |
|i Enthalten in
|t IEEE transactions on visualization and computer graphics
|d 1996
|g PP(2024) vom: 05. Sept.
|w (DE-627)NLM098269445
|x 1941-0506
|7 nnns
|
773 |
1 |
8 |
|g volume:PP
|g year:2024
|g day:05
|g month:09
|
856 |
4 |
0 |
|u http://dx.doi.org/10.1109/TVCG.2024.3454467
|3 Volltext
|
912 |
|
|
|a GBV_USEFLAG_A
|
912 |
|
|
|a SYSFLAG_A
|
912 |
|
|
|a GBV_NLM
|
912 |
|
|
|a GBV_ILN_350
|
951 |
|
|
|a AR
|
952 |
|
|
|d PP
|j 2024
|b 05
|c 09
|