I. Introduction
With vast potentials toward real-world applications, e.g., human-robot collaboration [1], medical robotics [2], assisted living, and so on, robotic manipulators integrated with vision systems have attracted significant research attention. For effective task execution in rapidly evolving environments, e.g., kitchens, hospital storage rooms, and flexible manufacturing workcells, a manipulator must be able to quickly understand its immediate surrounding environment, which may change on a daily basis even if the robot itself remains stationary. At the hardware level, this is best achieved by coupling vision to the end-effector of the manipulator, avoiding the inflexibility of fixed cameras, which have limited field-of-view (FoV) and their views may also be blocked by the robot and changing obstacles in the environment. Our study focuses on developing exploration algorithms for enabling high-DoF eye-on-hand manipulation systems to rapidly and accurately map out their immediate surroundings, ensuring their safe on-demand deployment, as shown in Fig. 1.