Published October 22, 2022 | Version 1.0.0
Dataset Open

GigaDepth: Learning Depth from Structured Light with Branching Neural Networks

  • 1. TU Wien, Vienna, Austria

Description

Structured light-based depth sensors provide accurate depth information independently of the scene appearance by extracting pattern positions from the captured pixel intensities.

Spatial neighborhood encoding, in particular, is a popular structured light approach for off-the-shelf hardware. However, it suffers from the distortion and fragmentation of the projected pattern by the scene's geometry in the vicinity of a pixel. This forces algorithms to find a delicate balance between depth prediction accuracy and robustness to pattern fragmentation or appearance change. While stereo matching provides more robustness at the expense of accuracy, we show that learning to regress a pixel's position within the projected pattern is not only more accurate when combined with classification but can be made equally robust. We propose to split the regression problem into smaller classification sub-problems in a coarse-to-fine manner with the use of a weight-adaptive layer that efficiently implements branching per-pixel Multilayer Perceptrons applied to features extracted by a Convolutional Neural Network.

As our approach requires full supervision, we train our algorithm on a rendered dataset sufficiently close to the real-world domain. On a separately captured real-world dataset, we show that our network outperforms state-of-the-art and is significantly more robust than other regression-based approaches.

Technical info (English)

Dataset used for training and evaluation. IR images are compressed via jpg.

Files

captured_test.zip

Files (316.5 MiB)

Name Size
md5:759e7d252df52b1bc5e11f25e1815054
316.5 MiB Preview Download

Additional details

Dates

Accepted
2022-10-22