Publication Type
Journal Article
Version
publishedVersion
Publication Date
3-2019
Abstract
Training a generic objectness measure to produce object proposals has recently become of significant interest. We observe that generic objects with well-defined closed boundaries can be detected by looking at the norm of gradients, with a suitable resizing of their corresponding image windows to a small fixed size. Based on this observation and computational reasons, we propose to resize the window to 8 × 8 and use the norm of the gradients as a simple 64D feature to describe it, for explicitly training a generic objectness measure. We further show how the binarized version of this feature, namely binarized normed gradients (BING), can be used for efficient objectness estimation, which requires only a few atomic operations (e.g., add, bitwise shift, etc.). To improve localization quality of the proposals while maintaining efficiency, we propose a novel fast segmentation method and demonstrate its effectiveness for improving BING’s localization performance, when used in multi-thresholding straddling expansion (MTSE) post-processing. On the challenging PASCAL VOC2007 dataset, using 1000 proposals per image and intersection-over-union threshold of 0.5, our proposal method achieves a 95.6% object detection rate and 78.6% mean average best overlap in less than 0.005 second per image.
Keywords
visual attention, category agnostic proposals, object proposals, objectness
Discipline
Databases and Information Systems
Research Areas
Data Science and Engineering
Publication
Computational Visual Media
Volume
5
Issue
1
First Page
3
Last Page
20
ISSN
2096-0433
Identifier
10.1007/s41095-018-0120-1
Publisher
SpringerOpen (part of Springer Nature)
Citation
CHENG, Ming-Ming; LIU, Yun; LIN, Wen-yan; ZHANG, Ziming; ROSIN, Paul L.; and TORR, Philip H. S..
BING: Binarized normed gradients for objectness estimation at 300fps. (2019). Computational Visual Media. 5, (1), 3-20.
Available at: https://ink.library.smu.edu.sg/sis_research/4716
Copyright Owner and License
Authors
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.1007/s41095-018-0120-1