Publication

4DPV: 4D pet from videos by coarse-to-fine non-rigid radiance fields

Conference Article

Conference

Asian Conference on Computer Vision (ACCV)

Edition

17th

Pages

141–157

Doc link

https://doi.org/10.1007/978-981-96-0969-7_9

File

Download the digital copy of the doc pdf document

Abstract

We present a coarse-to-fine neural deformation model to simultaneously recover the camera pose and the 4D reconstruction of an unknown object from multiple RGB sequences in the wild. To that end, our approach does not consider any pre-built 3D template nor 3D training data as well as controlled illumination conditions, and can sort out the problem in a self-supervised manner. Our model exploits canonical and image-variant spaces where both coarse and fine components are considered. We introduce a neural local quadratic model with spatio-temporal consistency to encode fine details that is combined with canonical embeddings in order to establish correspondences across sequences. We thoroughly validate the method on challenging scenarios with complex and real-world deformations, providing both quantitative and qualitative evaluations, an ablation study and a comparison with respect to competing approaches.

Categories

artificial intelligence, computer vision, optimisation.

Author keywords

Neural Rendering, deformable bodies, novel view synthesis.

Scientific reference

S. Montoya and A. Agudo. 4DPV: 4D pet from videos by coarse-to-fine non-rigid radiance fields, 17th Asian Conference on Computer Vision, 2024, Hanoi, in Computer Vision – ACCV 2024, Vol 15472 of Lecture Notes in Computer Science, pp. 141–157, 2024.