Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture [ICCV 15]

Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture [ICCV 15]

2017, Dec 01    

One Line Summary

Depth estimation, surface normal estimaiton and semantic labeling are done using a single basic architecture without any super pixels or low level segmentation, state-of-the-art results on all three tasks are shown.

Motivation

Use of a single architecture helps simplify the implementation of systems that require multiple modalities, e.g. robotics or augmented reality, which in turn can help enable research progress in these areas. Especially, in the case of depth and normals, much of the computation can be shared between modalities, making the system more efficient.

Detailed Summary

Starting with coarse level predictions in first scale and then doing finer level predictions in the next scales, effectively exploting the correaltion among the tasks of depth prediction, surface normal estimaiton and semantic segmentation by sharing the first two scales for all three tasks and only third scale network being individual to each task.

Novelty and Contributions

  • Use of the Multi-scale architecture where first and second scale are shared among different tasks
  • Showing the state-of-the-art results on all three tasks hence emprically proving the correlation between the three scene understanding tasks.

Network Details

Network figure

  • First two scales of the network are trained jointly from the start, somewhat simplifying the training procedure and yielding performance gains.
  • Predictions from each scale are fed into the next scale.

Results

Depth Estimation

Results Results

Surface Normal Estimation

Results Results

Semantic Segmentation

Results Results

Similar Works

Authours

David Eigen, Rob Fergus

Sources

Paper

Code