Vid2Avatar: 3D Avatar Reconstruction from Videos in the Wild via Self-supervised Scene Decomposition

02/22/2023
by   Chen Guo, et al.
0

We present Vid2Avatar, a method to learn human avatars from monocular in-the-wild videos. Reconstructing humans that move naturally from monocular in-the-wild videos is difficult. Solving it requires accurately separating humans from arbitrary backgrounds. Moreover, it requires reconstructing detailed 3D surface from short video sequences, making it even more challenging. Despite these challenges, our method does not require any groundtruth supervision or priors extracted from large datasets of clothed human scans, nor do we rely on any external segmentation modules. Instead, it solves the tasks of scene decomposition and surface reconstruction directly in 3D by modeling both the human and the background in the scene jointly, parameterized via two separate neural fields. Specifically, we define a temporally consistent human representation in canonical space and formulate a global optimization over the background model, the canonical human shape and texture, and per-frame human pose parameters. A coarse-to-fine sampling strategy for volume rendering and novel objectives are introduced for a clean separation of dynamic human and static background, yielding detailed and robust 3D human geometry reconstructions. We evaluate our methods on publicly available datasets and show improvements over prior art.

READ FULL TEXT

page 1

page 6

page 8

research
06/25/2021

Animatable Neural Radiance Fields from Monocular RGB Video

We present animatable neural radiance fields for detailed human avatar c...
research
05/23/2023

REC-MV: REconstructing 3D Dynamic Cloth from Monocular Videos

Reconstructing dynamic 3D garment surfaces with open boundaries from mon...
research
05/10/2023

Reconstructing Animatable Categories from Videos

Building animatable 3D models is challenging due to the need for 3D scan...
research
09/30/2022

MonoNeuralFusion: Online Monocular Neural 3D Reconstruction with Geometric Priors

High-fidelity 3D scene reconstruction from monocular videos continues to...
research
03/03/2022

PINA: Learning a Personalized Implicit Neural Avatar from a Single RGB-D Video Sequence

We present a novel method to learn Personalized Implicit Neural Avatars ...
research
10/10/2019

MetaPix: Few-Shot Video Retargeting

We address the task of unsupervised retargeting of human actions from on...
research
12/23/2021

BANMo: Building Animatable 3D Neural Models from Many Casual Videos

Prior work for articulated 3D shape reconstruction often relies on speci...

Please sign up or login with your details

Forgot password? Click here to reset