I. Introduction
It is increasingly common for physical locations to be surrounded and monitored by multiple cameras with overlapping fields of view (hereinafter ‘overlapping cameras’), e.g., intersections, shopping malls, public transport, construction sites and airports, as shown in Fig. 1. Such multiple overlapping cameras offer exciting opportunities to observe a scene from different angles, enabling enriched, comprehensive and robust analysis. For example, our analysis of the CityFlowV2 dataset [4] (5 cameras deployed to monitor vehicles on the road intersection) shows that each camera separately detects only 3.7 vehicles per frame on average, while five cameras detect a total of 12.0 vehicles altogether. Since a target vehicle can be captured by multiple cameras from different distances and angles, we can also observe objects of interest with a holistic view. Such view diversity can make the analytics more enriched and robust, e.g., a vehicle’s license plate may be occluded in one camera’s view due to its position or occlusion, but not in the other cameras.