Loading [MathJax]/extensions/MathMenu.js
Fast Motion Object Detection Algorithm Using Complementary Depth Image on an RGB-D Camera | IEEE Journals & Magazine | IEEE Xplore

Fast Motion Object Detection Algorithm Using Complementary Depth Image on an RGB-D Camera


Abstract:

Stereo vision has become a popular topic in recent years, especially in-depth images from stereo vision. Depth information can be extracted either from a dual camera or R...Show More

Abstract:

Stereo vision has become a popular topic in recent years, especially in-depth images from stereo vision. Depth information can be extracted either from a dual camera or RGB-D camera. In image processing, the realization of object detection is only based on the color information or depth images separately; however, both have advantages and disadvantages. Therefore, many researchers have combined them together to achieve better results. A new fast motion object-detection algorithm is presented based on the complementary depth images and color information, which is able to detect motion objects without background noise. The experiment results show that the proposed fast object detection algorithm can achieve 84.4% of the segmentation accuracy rate on average with a 45 FPS computation speed on an embedded platform.
Published in: IEEE Sensors Journal ( Volume: 17, Issue: 17, 01 September 2017)
Page(s): 5728 - 5734
Date of Publication: 05 July 2017

ISSN Information:

References is not available for this document.

I. Introduction

Recently, stereo vision has been used in surveillance systems to improve object recognition and tracking efficiency performance [1], [2]. Stereo vision can provide depth and color information for advanced image processing. Intrinsically different than RGB information, pixels in a depth map essentially represent 3-D space using stereo vision; hence, depth video represents the variation of the space information. Advances and developments in computer vision now can obtain stereo vision from either a dual camera system [3]–[10] or an RGB-D camera [11]–[14]. A dual camera system uses two video camera inputs to capture images through stereo matching then calculate the distance between the object and the camera. Calibration in a dual camera system is necessary before calculating the disparity map; however, the computational complexity of disparity match is typically difficult. Still, a major advantage is that the depth range is adjustable. On the other hand, an RGB-D system is composed of an RGB camera and a depth sensor, such as a Kinect or Asus Xtion Pro [15]. The biggest advantage of this system is that it is more convenient for acquiring depth information from a real-time video stream compared to the dual camera system. However, the depth information is noisy due to hardware drawbacks of a time-of-flight infrared-based depth sensor, and it can’t be used in outdoor environments.

Select All
1.
D. Stepanov and I. Tishchenko, "The concept of video surveillance system based on the principles of stereo vision", Proc. 18th Conf. Open Innov. Assoc. Seminar Inf. Secur. Protection Inf. Technol., pp. 328-334, Apr. 2016.
2.
U. A. Khan, M. Quaritsch and B. Rinner, "Design of a heterogeneous energy-aware stereo-vision based sensing platform for traffic surveillance", Proc. 9th Workshop Intell. Solutions Embedded Syst., pp. 47-52, Jul. 2011.
3.
M.-C. Chuang, J.-N. Hwang, K. Williams and R. Towler, "Tracking live fish from low-contrast and low-frame-rate stereo videos", IEEE Trans. Circuits Syst. Video Technol., vol. 25, no. 1, pp. 167-179, Jan. 2015.
4.
S. J. Krotosky and M. M. Trivedi, "On color- infrared- and multimodal-stereo approaches to pedestrian detection", IEEE Trans. Intell. Transp. Syst., vol. 8, no. 4, pp. 619-629, Dec. 2007.
5.
Y. Fang, I. Masaki and B. Horn, "Depth-based target segmentation for intelligent vehicles: Fusion of radar and binocular stereo", IEEE Trans. Intell. Transp. Syst., vol. 3, no. 3, pp. 196-202, Sep. 2002.
6.
G. Toulminet, M. Bertozzi, S. Mousset, A. Bensrhair and A. Broggi, "Vehicle detection by means of stereo vision-based obstacles features extraction and monocular pattern analysis", IEEE Trans. Image Process., vol. 15, no. 8, pp. 2364-2375, Aug. 2006.
7.
Q. Yang, D. Li, L. Wang and M. Zhang, "Full-image guided filtering for fast stereo matching", IEEE Signal Process. Lett., vol. 20, no. 3, pp. 237-240, Mar. 2013.
8.
A. Hosni, C. Rhemann, M. Bleyer, C. Rother and M. Gelautz, "Fast cost-volume filtering for visual correspondence and beyond", IEEE Trans. Pattern Anal. Mach. Intell., vol. 35, no. 2, pp. 504-511, Feb. 2013.
9.
R. K. Gupta and S.-Y. Cho, "Window-based approach for fast stereo correspondence", IET Comput. Vis., vol. 7, no. 2, pp. 123-134, Apr. 2013.
10.
L. Nalpantidis, A. Amanatiadis, G. C. Sirakoulis and A. Gasteratos, "Efficient hierarchical matching algorithm for processing uncalibrated stereo vision images and its hardware architecture", IET Image Process., vol. 5, no. 5, pp. 481-492, Aug. 2011.
11.
J. Han, E. J. Pauwels, P. M. de Zeeuw and P. H. N. de With, "Employing a RGB-D sensor for real-time tracking of humans across multiple re-entries in a smart environment", IEEE Trans. Consum. Electron., vol. 58, no. 2, pp. 255-263, May 2012.
12.
B. Ni, Y. Pei, P. Moulin and S. Yan, "Multilevel depth and image fusion for human activity detection", IEEE Trans. Cybern., vol. 43, no. 5, pp. 1383-1394, Oct. 2013.
13.
C. Wang, H. Liu and L. Ma, "Depth motion detection—A novel RS-trigger temporal logic based method", IEEE Signal Process. Lett., vol. 21, no. 6, pp. 717-721, Jun. 2014.
14.
H. Zhang, C. Reardon and L. E. Parker, "Real-time multiple human perception with color-depth cameras on a mobile robot", IEEE Trans. Cybern., vol. 43, no. 5, pp. 1429-1441, Oct. 2013.
15.
Z. Zhang, "Microsoft Kinect sensor and its effect", IEEE MultiMedia, vol. 19, no. 2, pp. 4-10, Feb. 2012.
16.
J. S. Kim, D. H. Yeom and Y. H. Joo, "Fast and robust algorithm of tracking multiple moving objects for intelligent video surveillance systems", IEEE Trans. Consum. Electron., vol. 57, no. 3, pp. 1165-1170, Aug. 2011.

Contact IEEE to Subscribe

References

References is not available for this document.