Learning to separate: detecting heavily-occluded objects in urban scenes
Files
Published version
Date
2020-12-04
Authors
Yang, Chenhongyi
Ablavsky, Vitaly
Wang, Kaihong
Feng, Qi
Betke, Margrit
Version
OA Version
Citation
C. Yang, V. Ablavsky, K. Wang, Q. Feng, M. Betke. 2020. "Learning to Separate: Detecting Heavily-Occluded Objects in Urban Scenes.." Vedaldi A., Bischof H., Brox T., Frahm JM. (eds) . ECCV 2020. Lecture Notes in Computer Science, vol 12363. Computer Vision – ECCV 2020. https://doi.org/10.1007/978-3-030-58523-5_31
Abstract
While visual object detection with deep learning has received much attention in the past decade, cases when heavy intra-class occlusions occur have not been studied thoroughly. In this work, we propose a Non-Maximum-Suppression (NMS) algorithm that dramatically improves the detection recall while maintaining high precision in scenes with heavy occlusions. Our NMS algorithm is derived from a novel embedding mechanism, in which the semantic and geometric features of the detected boxes are jointly exploited. The embedding makes it possible to determine whether two heavily-overlapping boxes belong to the same object in the physical world. Our approach is particularly useful for car detection and pedestrian detection in urban scenes where occlusions often happen. We show the effectiveness of our approach by creating a model called SG-Det (short for Semantics and Geometry Detection) and testing SG-Det on two widely-adopted datasets, KITTI and CityPersons for which it achieves state-of-the-art performance. Our code is available at https://github.com/ChenhongyiYang/SG-NMS.