Autonomous Vision Conference Paper 2020

Differentiable Volumetric Rendering: Learning Implicit 3D Representations without 3D Supervision

Michalecvpr

Learning-based 3D reconstruction methods have shown impressive results. However, most methods require 3D supervision which is often hard to obtain for real-world datasets. Recently, several works have proposed differentiable rendering techniques to train reconstruction models from RGB images. Unfortunately, these approaches are currently restricted to voxel- and mesh-based representations, suffering from discretization or low resolution. In this work, we propose a differentiable rendering formulation for implicit shape and texture representations. Implicit representations have recently gained popularity as they represent shape and texture continuously. Our key insight is that depth gradients can be derived analytically using the concept of implicit differentiation. This allows us to learn implicit shape and texture representations directly from RGB images. We experimentally show that our single-view reconstructions rival those learned with full 3D supervision. Moreover, we find that our method can be used for multi-view 3D reconstruction, directly resulting in watertight meshes.

Author(s): Michael Niemeyer and Lars Mescheder and Michael Oechsle and Andreas Geiger
Book Title: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020)
Pages: 3501 -- 3512
Year: 2020
Publisher: IEEE
Bibtex Type: Conference Paper (inproceedings)
Address: Piscataway, NJ
DOI: 10.1109/CVPR42600.2020.00356
Event Name: IEEE/CVF International Conference on Computer Vision and Pattern Recognition (CVPR 2020)
Event Place: Seattle, USA
State: Published
Electronic Archiving: grant_archive
ISBN: 9781728171692
Links:

BibTex

@inproceedings{Niemeyer2020CVPR,
  title = {Differentiable Volumetric Rendering: Learning Implicit 3D Representations without 3D Supervision},
  booktitle = {2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020)},
  abstract = {Learning-based 3D reconstruction methods have shown impressive results. However, most methods require 3D supervision which is often hard to obtain for real-world datasets. Recently, several works have proposed differentiable rendering techniques to train reconstruction models from RGB images. Unfortunately, these approaches are currently restricted to voxel- and mesh-based representations, suffering from discretization or low resolution. In this work, we propose a differentiable rendering formulation for implicit shape and texture representations. Implicit representations have recently gained popularity as they represent shape and texture continuously. Our key insight is that depth gradients can be derived analytically using the concept of implicit differentiation. This allows us to learn implicit shape and texture representations directly from RGB images. We experimentally show that our single-view reconstructions rival those learned with full 3D supervision. Moreover, we find that our method can be used for multi-view 3D reconstruction, directly resulting in watertight meshes.},
  pages = {3501 -- 3512},
  publisher = {IEEE},
  address = {Piscataway, NJ},
  year = {2020},
  slug = {niemeyer2020cvpr},
  author = {Niemeyer, Michael and Mescheder, Lars and Oechsle, Michael and Geiger, Andreas}
}