Unsupervised Monocular Depth Reconstruction of Non-Rigid Scenes

Open Access
Authors
  • A. Takmaz
  • D.P. Paudel
  • T. Probst
  • A. Chhatkuli
Publication date 2021
Book title 2021 International Conference on 3D Vision
Book subtitle proceedings : 3DV 2021 : virtual conference, 1-3 December 2021
ISBN
  • 9781665426893
ISBN (electronic)
  • 9781665426886
Event 2021 International Conference on 3D Vision
Pages (from-to) 825-836
Publisher Piscataway, NJ: Conference Publishing Services, IEEE Computer Society
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Monocular depth reconstruction of complex and dynamic scenes is a highly challenging problem. While for rigid scenes learning-based methods have been offering promising results even in unsupervised cases, there exists little to no literature addressing the same for dynamic and deformable scenes. In this work, we present an unsupervised monocular framework for dense depth estimation of dynamic scenes, which jointly reconstructs rigid and nonrigid parts without explicitly modelling the camera motion. Using dense correspondences, we derive a training objective that aims to opportunistically preserve pairwise distances between reconstructed 3D points. In this process, the dense depth map is learned implicitly using the as-rigid-as-possible hypothesis. Our method provides promising results, demonstrating its capability of reconstructing 3D from challenging videos of non-rigid scenes. Furthermore, the proposed method also provides unsupervised motion segmentation results as an auxiliary output.
Document type Conference contribution
Language English
Published at https://doi.org/10.48550/arXiv.2012.15680 https://doi.org/10.1109/3DV53792.2021.00091
Published at https://www.computer.org/csdl/proceedings-article/3dv/2021/268800a825/1zWEmrlTmsE
Other links https://www.proceedings.com/62174.html
Downloads
2012.15680 (Accepted author manuscript)
Permalink to this page
Back