AvatarWild:Fully controllable head avatars in the wild  

在线阅读下载全文

作  者:Shaoxu Meng Tong Wu Fang-Lue Zhang Shu-Yu Chen Yuewen Ma Wenbo Hu Lin Gao 

机构地区:[1]Beijing Key Laboratory of Mobile Computing and Pervasive Device,Institute of Computing Technology,Chinese Academy of Sciences,Beijing,China [2]University of Chinese Academy of Sciences,Beijing,China [3]Victoria University of Wellington,Wellington,New Zealand [4]ByteDance Ltd.,Beijing,China

出  处:《Visual Informatics》2024年第3期96-106,共11页可视信息学(英文)

基  金:supported by National Natural Science Foundation of China(No.6247075018 and No.62322210);the Innovation Funding of ICT,CAS(No.E461020);Beijing Munici-pal Natural Science Foundation for Distinguished Young Scholars(No.JQ21013);Beijing Municipal Science and Technology Commission(No.Z231100005923031).

摘  要:Recent advancements in the field have resulted in significant progress in achieving realistic head reconstruction and manipulation using neural radiance fields(NeRF).Despite these advances,capturing intricate facial details remains a persistent challenge.Moreover,casually captured input,involving both head poses and camera movements,introduces additional difficulties to existing methods of head avatar reconstruction.To address the challenge posed by video data captured with camera motion,we propose a novel method,AvatarWild,for reconstructing head avatars from monocular videos taken by consumer devices.Notably,our approach decouples the camera pose and head pose,allowing reconstructed avatars to be visualized with different poses and expressions from novel viewpoints.To enhance the visual quality of the reconstructed facial avatar,we introduce a view-dependent detail enhancement module designed to augment local facial details without compromising viewpoint consistency.Our method demonstrates superior performance compared to existing approaches,as evidenced by reconstruction and animation results on both multi-view and single-view datasets.Remarkably,our approach stands out by exclusively relying on video data captured by portable devices,such as smartphones.This not only underscores the practicality of our method but also extends its applicability to real-world scenarios where accessibility and ease of data capture are crucial.

关 键 词:Neural radiance fields Head avatar synthesis Face reconstruction Face reenactment Facial animation 

分 类 号:TP39[自动化与计算机技术—计算机应用技术]

 

参考文献:

正在载入数据...

 

二级参考文献:

正在载入数据...

 

耦合文献:

正在载入数据...

 

引证文献:

正在载入数据...

 

二级引证文献:

正在载入数据...

 

同被引文献:

正在载入数据...

 

相关期刊文献:

正在载入数据...

相关的主题
相关的作者对象
相关的机构对象