1 Introduction
With the wide availability of low cost commodity 3D full frame sensors, such as the Microsoft Kinect and Intel Real-Sense, it is becoming easier and easier to capture 3D data. Starting from the robust system of KinectFusion [1], [2], large scale scanning [3], [4], to the more recent work of fusion of dynamic objects over space and time [5], [6], good quality 3D models can be obtained even with a single hand-held depth camera. These approaches typically use a volumetric representation of 3D geometry, the final model is mostly presented as a white (texture-less) model or colored with simple per-vertex color. Given the high cost associated with a dense voxel grid, per-vertex colored model appears to be blurry and low resolution, the final model's appearance captured by these methods left something to be desired.