OAK

CanonicalFusion: Generating Drivable 3D Human Avatars from Multiple Images

Metadata Downloads
Author(s)
Shin, JisuLee, JunmyeongLee, SeongminPark, Min-gyuKang, Ju-miYoon, JuhongJeon, Hae Gon
Type
Conference Paper
Citation
18th European Conference on Computer Vision, ECCV 2024, pp.38 - 56
Issued Date
2025-10-04
Abstract
We present a novel framework for reconstructing animatable human avatars from multiple images, termed CanonicalFusion. Our central concept involves integrating individual reconstruction results into the canonical space. To be specific, we first predict Linear Blend Skinning (LBS) weight maps and depth maps using a shared-encoder-dual-decoder network, enabling direct canonicalization of the 3D mesh from the predicted depth maps. Here, instead of predicting high-dimensional skinning weights, we infer compressed skinning weights, i.e., 3-dimensional vector, with the aid of pre-trained MLP networks. We also introduce a forward skinning-based differentiable rendering scheme to merge the reconstructed results from multiple images. This scheme refines the initial mesh by reposing the canonical mesh via the forward skinning and by minimizing photometric and geometric errors between the rendered and the predicted results. Our optimization scheme considers the position and color of vertices as well as the joint angles for each image, thereby mitigating the negative effects of pose errors. We conduct extensive experiments to demonstrate the effectiveness of our method and compare our CanonicalFusion with state-of-the-art methods. Our source codes are available at https://github.com/jsshin98/CanonicalFusion. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2025.
Publisher
Springer Science and Business Media Deutschland GmbH
Conference Place
IT
Milan
URI
https://scholar.gist.ac.kr/handle/local/32426
공개 및 라이선스
  • 공개 구분공개
파일 목록
  • 관련 파일이 존재하지 않습니다.

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.