Autonomous Vision Conference Paper 2020

Convolutional Occupancy Networks

Screenshot from 2020 08 06 14 35 32

Recently, implicit neural representations have gained popularity for learning-based 3D reconstruction. While demonstrating promising results, most implicit approaches are limited to comparably simple geometry of single objects and do not scale to more complicated or large-scale scenes. The key limiting factor of implicit methods is their simple fully-connected network architecture which does not allow for integrating local information in the observations or incorporating inductive biases such as translational equivariance. In this paper, we propose Convolutional Occupancy Networks, a more flexible implicit representation for detailed reconstruction of objects and 3D scenes. By combining convolutional encoders with implicit occupancy decoders, our model incorporates inductive biases, enabling structured reasoning in 3D space. We investigate the effectiveness of the proposed representation by reconstructing complex geometry from noisy point clouds and low-resolution voxel representations. We empirically find that our method enables the fine-grained implicit 3D reconstruction of single objects, scales to large indoor scenes, and generalizes well from synthetic to real data.

Author(s): Songyou Peng and Michael Niemeyer and Lars Mescheder and Marc Pollefeys and Andreas Geiger
Book Title: Computer Vision – ECCV 2020
Volume: 3
Pages: 523--540
Year: 2020
Month: August
Series: Lecture Notes in Computer Science, 12348
Editors: Vedaldi, Andrea and Bischof, Horst and Brox, Thomas and Frahm, Jan-Michael
Publisher: Springer
Bibtex Type: Conference Paper (inproceedings)
Address: Cham
DOI: 10.1007/978-3-030-58580-8_31
Event Name: 16th European Conference on Computer Vision (ECCV 2020)
Event Place: Glasgow
State: Published
Electronic Archiving: grant_archive
ISBN: 978-3-030-58579-2
Links:

BibTex

@inproceedings{Peng2020ECCV,
  title = {Convolutional Occupancy Networks},
  booktitle = {Computer Vision – ECCV 2020},
  abstract = {Recently, implicit neural representations have gained popularity for learning-based 3D reconstruction. While demonstrating promising results, most implicit approaches are limited to comparably simple geometry of single objects and do not scale to more complicated or large-scale scenes. The key limiting factor of implicit methods is their simple fully-connected network architecture which does not allow for integrating local information in the observations or incorporating inductive biases such as translational equivariance. In this paper, we propose Convolutional Occupancy Networks, a more flexible implicit representation for detailed reconstruction of objects and 3D scenes. By combining convolutional encoders with implicit occupancy decoders, our model incorporates inductive biases, enabling structured reasoning in 3D space. We investigate the effectiveness of the proposed representation by reconstructing complex geometry from noisy point clouds and low-resolution voxel representations. We empirically find that our method enables the fine-grained implicit 3D reconstruction of single objects, scales to large indoor scenes, and generalizes well from synthetic to real data.},
  volume = {3},
  pages = {523--540},
  series = {Lecture Notes in Computer Science, 12348},
  editors = {Vedaldi, Andrea and Bischof, Horst and Brox, Thomas and Frahm, Jan-Michael},
  publisher = {Springer},
  address = {Cham},
  month = aug,
  year = {2020},
  slug = {peng2020eccv},
  author = {Peng, Songyou and Niemeyer, Michael and Mescheder, Lars and Pollefeys, Marc and Geiger, Andreas},
  month_numeric = {8}
}