File Download

There are no files associated with this item.

Supplementary

Conference Paper: Sparse R-CNN: End-to-End Object Detection With Learnable Proposals

TitleSparse R-CNN: End-to-End Object Detection With Learnable Proposals
Authors
Issue Date2021
Citation
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Virtual Conference, 19-25 June 2021, p. 14454-14463 How to Cite?
AbstractWe present Sparse R-CNN, a purely sparse method for object detection in images. Existing works on object detection heavily rely on dense object candidates, such as k anchor boxes pre-defined on all grids of image feature map of size H × W. In our method, however, a fixed sparse set of learned object proposals, total length of N, are provided to object recognition head to perform classification and location. By eliminating HW k (up to hundreds of thousands) hand-designed object candidates to N (e.g. 100) learnable proposals, Sparse R-CNN completely avoids all efforts related to object candidates design and many-toone label assignment. More importantly, final predictions are directly output without non-maximum suppression postprocedure. Sparse R-CNN demonstrates accuracy, run-time and training convergence performance on par with the wellestablished detector baselines on the challenging COCO dataset, e.g., achieving 45.0 AP in standard 3× training schedule and running at 22 fps using ResNet-50 FPN model. We hope our work could inspire re-thinking the convention of dense prior in object detectors. The code is available at: https://github.com/PeizeSun/SparseR-CNN.
DescriptionPaper Session Eleven: Paper ID 2829
Persistent Identifierhttp://hdl.handle.net/10722/301313

 

DC FieldValueLanguage
dc.contributor.authorSun, P-
dc.contributor.authorZhang, R-
dc.contributor.authorJiang, Y-
dc.contributor.authorKong, T-
dc.contributor.authorXu, C-
dc.contributor.authorZhan, W-
dc.contributor.authorTomizuka, M-
dc.contributor.authorLi, L-
dc.contributor.authorYuan, Z-
dc.contributor.authorWang, C-
dc.contributor.authorLuo, P-
dc.date.accessioned2021-07-27T08:09:16Z-
dc.date.available2021-07-27T08:09:16Z-
dc.date.issued2021-
dc.identifier.citationProceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Virtual Conference, 19-25 June 2021, p. 14454-14463-
dc.identifier.urihttp://hdl.handle.net/10722/301313-
dc.descriptionPaper Session Eleven: Paper ID 2829-
dc.description.abstractWe present Sparse R-CNN, a purely sparse method for object detection in images. Existing works on object detection heavily rely on dense object candidates, such as k anchor boxes pre-defined on all grids of image feature map of size H × W. In our method, however, a fixed sparse set of learned object proposals, total length of N, are provided to object recognition head to perform classification and location. By eliminating HW k (up to hundreds of thousands) hand-designed object candidates to N (e.g. 100) learnable proposals, Sparse R-CNN completely avoids all efforts related to object candidates design and many-toone label assignment. More importantly, final predictions are directly output without non-maximum suppression postprocedure. Sparse R-CNN demonstrates accuracy, run-time and training convergence performance on par with the wellestablished detector baselines on the challenging COCO dataset, e.g., achieving 45.0 AP in standard 3× training schedule and running at 22 fps using ResNet-50 FPN model. We hope our work could inspire re-thinking the convention of dense prior in object detectors. The code is available at: https://github.com/PeizeSun/SparseR-CNN.-
dc.languageeng-
dc.relation.ispartofIEEE Computer Vision and Pattern Recognition (CVPR) Proceedings-
dc.titleSparse R-CNN: End-to-End Object Detection With Learnable Proposals-
dc.typeConference_Paper-
dc.identifier.emailLuo, P: pluo@hku.hk-
dc.identifier.authorityLuo, P=rp02575-
dc.identifier.hkuros323748-
dc.identifier.spage14454-
dc.identifier.epage14463-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats