Michael Schelling, long term member of the research group Visual Computing has brilliantly defended his phd project of the title Viewpoint Guided Learning for 3D Scene Understanding and Representations. The jury consisted of Prof. Dr Timo Ropinski (Institute of Media Informatics, Ulm University), Prof. Dr. Andreas Kolb (University of Siegen) , completed by apl. Prof. Dr Rainer Michalzik and Prof. Dr Heiko Neumann (both Ulm University), and Prof. Dr Matthias Tichy (head of the commission and minutes).
Abstract: Deep Learning has become the dominating technique in many domains, including visual computing. The visual perception, of both humans and machines, is rooted in the 2D space, as visual sensors are ultimately only able to capture a two-dimensional representation from their specific viewpoint. Even 3D sensors, such as indirect Time-of- Flight (iToF) cameras, can only perceive information which is visible from their respective viewpoint, and thus can only model the observed 3D space as a 2D manifold, which is commonly referred to as a 2.5D image.
Consequently, the visual observation is highly dependent on both the observed 3D geometry and the chosen viewpoint. This interdependency of the viewpoint and the observed 3D structure can be exploited to improve the performance of neural networks for visual tasks. To this end, I present algorithmic optimizations for neural network training, which allow for the use of viewpoint information when addressing 3D problems and, vice versa, the use of 3D information when addressing viewpoint related problems.
First, a dynamic labeling strategy is presented, which enables 3D point networks to identify informative viewpoints for 3D models. Second, it is shown how the integration of information about the view direction into 3D point convolutional networks can improve the error correction rates for iToF cameras. Lastly, a training method for compensating motion artefacts in iToF images, e.g., through changes of the viewpoint, is derived, which allows for the supervision of 2D networks via the reconstructed 2.5D depth image.
We wish him all the best having succeeded in laying a solid foundation to further adventures.