I. Introduction
To better understand 3D shapes, objects, and scenes captured by LIDAR sensors and RGB-D cameras, rapidly increasing attention has been paid to 3D point cloud processing algorithms in the community of computer vision, robotics, and mixed realities. Significant progress has been made in recent works, which mainly focus on high-level tasks, such as 3D object classification, segmentation, and detection [1]–[9]. However, most of these works still face considerable challenges since the point clouds captured in the real scenarios are usually noisy, sparse and non-uniform.