Object Pose Estimation in Monocular Image Using Modified FDCM
DOI:
https://doi.org/10.7494/csci.2020.21.1.3426Keywords:
Pose estimation, FDCM, Monocular Image, Voronoi diagram, Line Based Matching, LSDAbstract
In this paper, a new method for object detection and pose estimation in a monocular image is proposed based on FDCM method. it can detect object with high speed running time, even if the object was under the partial occlusion or in bad illumination. In addition, It requires only single template without any training process. The Modied FDCM based on FDCM with improvments, the LSD method was used in MFDCM instead of the line tting method, besides the integral distance transform was replaced with a distance transform image, and using an angular Voronoi diagram. In addition, the search process depends on Line segments based search instead of the sliding window search in FDCM. The MFDCM was evaluated by comparing it with FDCM in dierent scenarios and with other four methods: COF, HALCON, LINE2D, and BOLD using D-textureless dataset. The comparison results show that MFDCM was at least 14 times faster than FDCM in tested scenarios. Furthermore, it has the highest correct detection rate among all tested method with small advantage from COF and BLOD methods, while it was a little slower than LINE2D which was the fasted method among compared methods. The results proves that MFDCM able to detect and pose estimation of the objects in the clear or clustered background from a monocular image with high speed running time, even if the object was under the partial occlusion which makes it robust and reliable for real-time applications.
Downloads
References
M. Y. Liu, O. Tuzel, A. Veeraraghavan, and R. Chellappa, “Fast directional chamfer matching,” Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., pp. 1696–1703, 2010.
Y. Konishi, Y. Ijiri, M. Suwa, and M. Kawade, “Textureless object detection using cumulative orientation feature,” Proc. - Int. Conf. Image Process. ICIP, vol. 2015–Decem, pp. 1310–1313, 2015.
C. Steger, “Occlusion, clutter, and illumination invariant object recognition,” Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci., vol. 34, no. 3/A, pp. 345–350, 2002.
A. Eccv, “Gradient Response Maps for Real-Time Detection of Texture-Less Objects,” pp. 1–14, 2012.
F. Tombari, A. Franchi, and L. Di, “BOLD features to detect texture-less objects,” Proc. IEEE Int. Conf. Comput. Vis., pp. 1265–1272, 2013.
C. J. Phillips, M. Lecce, and K. Daniilidis, “Seeing Glassware : from Edge Detection to Pose Estimation and Shape Recovery,” Robot. Sci. Syst., 2016.
E. Muñoz, Y. Konishi, C. Beltran, V. Murino, and A. Del Bue, “Fast 6D pose from a single RGB image using cascaded forests templates,” IEEE Int. Conf. Intell. Robot. Syst., vol. 2016–Novem, pp. 4062–4069, 2016.
P. Han and G. Zhao, “CAD-based 3D objects recognition in monocular images for mobile augmented reality,” Comput. Graph., vol. 50, pp. 36–46, 2015.
C. Jaynes and J. Hou, “Temporal Registration using a Kalman Filter for Augmented Reality Applications,” Int. J., no. June 2003, 2014.
M. Imperoli and A. Pretto, “Active Detection and Localization of Textureless Objects in Cluttered Environments,” 2016.
P. Tsarouchi, S. A. Matthaiakis, G. Michalos, S. Makris, and G. Chryssolouris, “A method for detection of randomly placed objects for robotic handling,” CIRP J. Manuf. Sci. Technol., vol. 14, pp. 20–27, 2016.
P. Tsarouchi, G. Michalos, S. Makris, and G. Chryssolouris, “Vision system for robotic handling of randomly placed objects,” Procedia CIRP, vol. 9, pp. 61–66, 2013.
M.-Y. Liu, O. Tuzel, A. Veeraraghavan, Y. Taguchi, T. K. Marks, and R. Chellappa, “Fast object localization and pose estimation in heavy clutter for robotic bin picking,” Int. J. Rob. Res., vol. 31, no. 8, pp. 951–973, 2012.
M. Ulrich, C. Wiedemann, and C. Steger, “Combining scale-space and similarity-based aspect graphs for fast 3D object recognition,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 34, no. 10, pp. 1902–1914, 2012.
S. Hinterstoisser et al., “Gradient Response Maps and their Efficient Computation for Real-Time Detection of Texture-Less Objects,” IEEE Trans. Pattern Anal. Mach. Intell., pp. 1–14, 2011.
S. Hinterstoisser, V. Lepetit, S. Ilic, P. Fua, and N. Navab, “Dominant orientation templates for real-time detection of texture-less objects,” Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., pp. 2257–2264, 2010.
A. Tejani, D. Tang, R. Kouskouridas, and T. K. Kim, “Latent-class Hough forests for 3D object detection and pose estimation,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 8694 LNCS, no. PART 6, pp. 462–477, 2014.
S. Hinterstoisser et al., “Multimodal templates for real-time detection of texture-less objects in heavily cluttered scenes,” Proc. IEEE Int. Conf. Comput. Vis., pp. 858–865, 2011.
T. Mörwald, M. Zillich, and M. Vincze, “Edge tracking of textured objects with a recursive particle filter,” 19th Int. Conf. Comput. Graph. Vis., pp. 96–103, 2009.
R. Hernorrhagic and S. Zuiderzee, “Model Based Training, Detection and Pose Estimation of Texture-Less 3D Objects in Heavily Cluttered Scenes Stefan,” pp. 1–14, 2016.
X. Peng, “Combine color and shape in real-time detection of texture-less objects,” Comput. Vis. Image Underst., vol. 135, no. February, pp. 31–48, 2015.
B. Bratanič, F. Pernuš, B. Likar, and D. Tomaževič, “Real-time pose estimation of rigid objects in heavily cluttered environments,” Comput. Vis. Image Underst., vol. 141, pp. 38–51, 2015.
B. Fan, F. Wu, and Z. Hu, “Aggregating gradient distributions into intensity orders: A novel local image descriptor,” Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., pp. 2377–2384, 2011.
B. Fan, F. Wu, and Z. Hu, “Rotationally invariant descriptors using intensity order pooling,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 34, no. 10, pp. 2031–2045, 2012.
R. Rios-Cabrera and T. Tuytelaars, “Boosting masked dominant orientation templates for efficient object detection,” Comput. Vis. Image Underst., vol. 120, pp. 103–116, 2014.
S. James and J. Collomosse, “Interactive video asset retrieval using sketched queries,” no. November, pp. 1–8, 2014.
X. Xu, L. Tian, J. Feng, and J. Zhou, “OSRI: A rotationally invariant binary descriptor,” IEEE Trans. Image Process., vol. 23, no. 7, pp. 2983–2995, 2014.
H. Wei, C. Yang, and Q. Yu, “Efficient graph-based search for object detection,” Inf. Sci. (Ny)., vol. 385–386, no. 825, pp. 395–414, 2017.
H. Wei, C. Yang, and Q. Yu, “Contour segment grouping for object detection,” J. Vis. Commun. Image Represent., vol. 48, no. July, pp. 292–309, 2017.
B. Nagarajan and B. Rajathilagam, “Object Shapes from Regular Curves through Sparse Representations,” Procedia Comput. Sci., vol. 133, pp. 635–642, 2018.
G. Pavlakos, X. Zhou, A. Chan, K. G. Derpanis, and K. Daniilidis, “6-DoF Object Pose from Semantic Keypoints,” 2017.
J. Wu et al., “Single image 3D interpreter network,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 9910 LNCS, pp. 365–382, 2016.
D. Damen, P. Bunnun, A. Calway, and W. Mayol-cuevas, “Real-time Learning and Detection of 3D Texture-less Objects: A Scalable Approach,” Procedings Br. Mach. Vis. Conf. 2012, p. 23.1-23.12, 2012.
B. Barrois and C. Wöhler, “3D Pose Estimation Based on Multiple Monocular Cues,” 2007.
M. Shi, H. Caesar, and V. Ferrari, “Fast 6D Pose Estimation from a Monocular Image Using Hierarchical Pose Trees,” vol. 1, pp. 398–413, 2017.
M. Imperoli and A. Pretto, “D2co: Fast and robust registration of 3d textureless objects using the directional chamfer distance,” Lect. Notes Comput. Sci. (including Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinformatics), vol. 9163, pp. 316–328, 2015.
M. Zhu et al., “Single image 3D object detection and pose estimation for grasping,” 2014 IEEE Int. Conf. Robot. Autom., pp. 3936–3943, 2014.
H. Cai, T. Werner, and J. Matas, “Fast detection of multiple textureless 3-D objects,” Comput. Vis. Syst., 2013.
D. Gálvez-López, M. Salas, J. D. Tardós, and J. M. M. Montiel, “Real-time monocular object SLAM,” Rob. Auton. Syst., vol. 75, pp. 435–449, 2016.
R. Grompone Von Gioi, J. Jakubowicz, J.-M. Morel, and G. Randall, “LSD: a Line Segment Detector Rafael,” Image Process. Line, vol. 2, pp. 35–55, 2012.