Many approaches have been suggested for automatic pedestrian and car detection to cope with the large variability regarding size, occlusion, background variability, and aspect. Current deep learning-based frameworks rely either on a proposal generation mechanism (e.g., “faster R-CNN”) or on inspection of image quadrants/octants (e.g., “YOLO”), which are then further processed with deep convolutional neural networks (CNN). We analyze the discriminative generalized Hough transform (DGHT), which operates on edge images, for pedestrian and car detection. The analysis motivates one to use the DGHT as an efficient proposal generation mechanism, followed by a proposal (bounding box) refinement and proposal acceptance or rejection based on a deep CNN. We analyze in detail the different components of our pipeline. Due to the low false negative rate and the low number of candidates of the DGHT as well as the high accuracy of the CNN, we obtain competitive performance to the state of the art in pedestrian and car detection on the IAIR database with much less generated proposals than other proposal-generating algorithms, being outperformed only by YOLOv2 fine-tuned to IAIR cars. By evaluations on further databases (without retraining or adaptation), we show the generalization capability of our pipeline.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.