File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: Scale-equivalent distillation for semi-supervised object detection
Title | Scale-equivalent distillation for semi-supervised object detection |
---|---|
Authors | |
Issue Date | 2022 |
Publisher | IEEE. |
Citation | IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (Hybrid), New Orleans, Louisiana, USA, 19-24, 2022. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, p. 14522-14531 How to Cite? |
Abstract | Recent Semi-Supervised Object Detection (SS-OD) methods are mainly based on self-training, ie, generating hard pseudo-labels by a teacher model on unlabeled data as supervisory signals. Although they achieved certain success, the limited labeled data in semi-supervised learning scales up the challenges of object detection. We analyze the challenges these methods meet with the empirical experiment results. We find that the massive False Negative samples and inferior localization precision lack consideration. Besides, the large variance of object sizes and class imbalance (ie, the extreme ratio between background and object) hinder the performance of prior arts. Further, we overcome these challenges by introducing a novel approach, Scale-Equivalent Distillation (SED), which is a simple yet effective end-to-end knowledge distillation framework robust to large object size variance and class imbalance. SED has several appealing benefits compared to the previous works.(1) SED imposes a consistency regularization to handle the large scale variance problem.(2) SED alleviates the noise problem from the False Negative samples and inferior localization precision.(3) A re-weighting strategy can implicitly screen the potential foreground regions of the unlabeled data to reduce the effect of class imbalance. Extensive experiments show that SED consistently outperforms the recent state-of-the-art methods on different datasets with significant margins. For example, it surpasses the supervised counterpart by more than 10 mAP when using 5% and 10% labeled data on MS-COCO. |
Persistent Identifier | http://hdl.handle.net/10722/315845 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Guo, Q | - |
dc.contributor.author | Mu, Y | - |
dc.contributor.author | Chen, J | - |
dc.contributor.author | Wang, T | - |
dc.contributor.author | Yu, Y | - |
dc.contributor.author | Luo, P | - |
dc.date.accessioned | 2022-08-19T09:05:29Z | - |
dc.date.available | 2022-08-19T09:05:29Z | - |
dc.date.issued | 2022 | - |
dc.identifier.citation | IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (Hybrid), New Orleans, Louisiana, USA, 19-24, 2022. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, p. 14522-14531 | - |
dc.identifier.uri | http://hdl.handle.net/10722/315845 | - |
dc.description.abstract | Recent Semi-Supervised Object Detection (SS-OD) methods are mainly based on self-training, ie, generating hard pseudo-labels by a teacher model on unlabeled data as supervisory signals. Although they achieved certain success, the limited labeled data in semi-supervised learning scales up the challenges of object detection. We analyze the challenges these methods meet with the empirical experiment results. We find that the massive False Negative samples and inferior localization precision lack consideration. Besides, the large variance of object sizes and class imbalance (ie, the extreme ratio between background and object) hinder the performance of prior arts. Further, we overcome these challenges by introducing a novel approach, Scale-Equivalent Distillation (SED), which is a simple yet effective end-to-end knowledge distillation framework robust to large object size variance and class imbalance. SED has several appealing benefits compared to the previous works.(1) SED imposes a consistency regularization to handle the large scale variance problem.(2) SED alleviates the noise problem from the False Negative samples and inferior localization precision.(3) A re-weighting strategy can implicitly screen the potential foreground regions of the unlabeled data to reduce the effect of class imbalance. Extensive experiments show that SED consistently outperforms the recent state-of-the-art methods on different datasets with significant margins. For example, it surpasses the supervised counterpart by more than 10 mAP when using 5% and 10% labeled data on MS-COCO. | - |
dc.language | eng | - |
dc.publisher | IEEE. | - |
dc.relation.ispartof | Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022 | - |
dc.rights | Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022. Copyright © IEEE. | - |
dc.rights | ©20xx IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. | - |
dc.title | Scale-equivalent distillation for semi-supervised object detection | - |
dc.type | Conference_Paper | - |
dc.identifier.email | Yu, Y: yzyu@cs.hku.hk | - |
dc.identifier.email | Luo, P: pluo@hku.hk | - |
dc.identifier.authority | Yu, Y=rp01415 | - |
dc.identifier.authority | Luo, P=rp02575 | - |
dc.identifier.hkuros | 335569 | - |
dc.identifier.spage | 14522 | - |
dc.identifier.epage | 14531 | - |
dc.publisher.place | United States | - |