File Download

There are no files associated with this item.

Supplementary

Conference Paper: Scale-equivalent distillation for semi-supervised object detection

TitleScale-equivalent distillation for semi-supervised object detection
Authors
Issue Date2022
PublisherIEEE.
Citation
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (Hybrid), New Orleans, Louisiana, USA, 19-24, 2022. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, p. 14522-14531 How to Cite?
AbstractRecent Semi-Supervised Object Detection (SS-OD) methods are mainly based on self-training, ie, generating hard pseudo-labels by a teacher model on unlabeled data as supervisory signals. Although they achieved certain success, the limited labeled data in semi-supervised learning scales up the challenges of object detection. We analyze the challenges these methods meet with the empirical experiment results. We find that the massive False Negative samples and inferior localization precision lack consideration. Besides, the large variance of object sizes and class imbalance (ie, the extreme ratio between background and object) hinder the performance of prior arts. Further, we overcome these challenges by introducing a novel approach, Scale-Equivalent Distillation (SED), which is a simple yet effective end-to-end knowledge distillation framework robust to large object size variance and class imbalance. SED has several appealing benefits compared to the previous works.(1) SED imposes a consistency regularization to handle the large scale variance problem.(2) SED alleviates the noise problem from the False Negative samples and inferior localization precision.(3) A re-weighting strategy can implicitly screen the potential foreground regions of the unlabeled data to reduce the effect of class imbalance. Extensive experiments show that SED consistently outperforms the recent state-of-the-art methods on different datasets with significant margins. For example, it surpasses the supervised counterpart by more than 10 mAP when using 5% and 10% labeled data on MS-COCO.
Persistent Identifierhttp://hdl.handle.net/10722/315845

 

DC FieldValueLanguage
dc.contributor.authorGuo, Q-
dc.contributor.authorMu, Y-
dc.contributor.authorChen, J-
dc.contributor.authorWang, T-
dc.contributor.authorYu, Y-
dc.contributor.authorLuo, P-
dc.date.accessioned2022-08-19T09:05:29Z-
dc.date.available2022-08-19T09:05:29Z-
dc.date.issued2022-
dc.identifier.citationIEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (Hybrid), New Orleans, Louisiana, USA, 19-24, 2022. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, p. 14522-14531-
dc.identifier.urihttp://hdl.handle.net/10722/315845-
dc.description.abstractRecent Semi-Supervised Object Detection (SS-OD) methods are mainly based on self-training, ie, generating hard pseudo-labels by a teacher model on unlabeled data as supervisory signals. Although they achieved certain success, the limited labeled data in semi-supervised learning scales up the challenges of object detection. We analyze the challenges these methods meet with the empirical experiment results. We find that the massive False Negative samples and inferior localization precision lack consideration. Besides, the large variance of object sizes and class imbalance (ie, the extreme ratio between background and object) hinder the performance of prior arts. Further, we overcome these challenges by introducing a novel approach, Scale-Equivalent Distillation (SED), which is a simple yet effective end-to-end knowledge distillation framework robust to large object size variance and class imbalance. SED has several appealing benefits compared to the previous works.(1) SED imposes a consistency regularization to handle the large scale variance problem.(2) SED alleviates the noise problem from the False Negative samples and inferior localization precision.(3) A re-weighting strategy can implicitly screen the potential foreground regions of the unlabeled data to reduce the effect of class imbalance. Extensive experiments show that SED consistently outperforms the recent state-of-the-art methods on different datasets with significant margins. For example, it surpasses the supervised counterpart by more than 10 mAP when using 5% and 10% labeled data on MS-COCO.-
dc.languageeng-
dc.publisherIEEE.-
dc.relation.ispartofProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022-
dc.rightsProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022. Copyright © IEEE.-
dc.rights©20xx IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.-
dc.titleScale-equivalent distillation for semi-supervised object detection-
dc.typeConference_Paper-
dc.identifier.emailYu, Y: yzyu@cs.hku.hk-
dc.identifier.emailLuo, P: pluo@hku.hk-
dc.identifier.authorityYu, Y=rp01415-
dc.identifier.authorityLuo, P=rp02575-
dc.identifier.hkuros335569-
dc.identifier.spage14522-
dc.identifier.epage14531-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats