File Download

There are no files associated with this item.

Supplementary

Conference Paper: Panoptic segformer: Delving deeper into panoptic segmentation with transformers

TitlePanoptic segformer: Delving deeper into panoptic segmentation with transformers
Authors
Issue Date2022
PublisherIEEE.
Citation
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (Hybrid), New Orleans, Louisiana, USA, 19-24, 2022. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, p. 1280-1289 How to Cite?
AbstractPanoptic segmentation involves a combination of joint semantic segmentation and instance segmentation, where image contents are divided into two types: things and stuff. We present Panoptic SegFormer, a general framework for panoptic segmentation with transformers. It contains three innovative components: an efficient deeply-supervised mask decoder, a query decoupling strategy, and an improved post-processing method. We also use Deformable DETR to efficiently process multi-scale features, which is a fast and efficient version of DETR. Specifically, we supervise the attention modules in the mask decoder in a layer-wise manner. This deep supervision strategy lets the attention modules quickly focus on meaningful semantic regions. It improves performance and reduces the number of required training epochs by half compared to Deformable DETR. Our query decoupling strategy decouples the responsibilities of the query set and avoids mutual interference between things and stuff. In addition, our post-processing strategy improves performance without additional costs by jointly considering classification and segmentation qualities to resolve conflicting mask overlaps. Our approach increases the accuracy 6.2\% PQ over the baseline DETR model. Panoptic SegFormer achieves state-of-the-art results on COCO test-dev with 56.2\% PQ. It also shows stronger zero-shot robustness over existing methods.
Persistent Identifierhttp://hdl.handle.net/10722/315795

 

DC FieldValueLanguage
dc.contributor.authorLi, Z-
dc.contributor.authorWang, W-
dc.contributor.authorXie, E-
dc.contributor.authorYu, Z-
dc.contributor.authorAnandkumar, A-
dc.contributor.authorAlvarez, JM-
dc.contributor.authorLuo, P-
dc.contributor.authorLu, T-
dc.date.accessioned2022-08-19T09:04:34Z-
dc.date.available2022-08-19T09:04:34Z-
dc.date.issued2022-
dc.identifier.citationIEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (Hybrid), New Orleans, Louisiana, USA, 19-24, 2022. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, p. 1280-1289-
dc.identifier.urihttp://hdl.handle.net/10722/315795-
dc.description.abstractPanoptic segmentation involves a combination of joint semantic segmentation and instance segmentation, where image contents are divided into two types: things and stuff. We present Panoptic SegFormer, a general framework for panoptic segmentation with transformers. It contains three innovative components: an efficient deeply-supervised mask decoder, a query decoupling strategy, and an improved post-processing method. We also use Deformable DETR to efficiently process multi-scale features, which is a fast and efficient version of DETR. Specifically, we supervise the attention modules in the mask decoder in a layer-wise manner. This deep supervision strategy lets the attention modules quickly focus on meaningful semantic regions. It improves performance and reduces the number of required training epochs by half compared to Deformable DETR. Our query decoupling strategy decouples the responsibilities of the query set and avoids mutual interference between things and stuff. In addition, our post-processing strategy improves performance without additional costs by jointly considering classification and segmentation qualities to resolve conflicting mask overlaps. Our approach increases the accuracy 6.2\% PQ over the baseline DETR model. Panoptic SegFormer achieves state-of-the-art results on COCO test-dev with 56.2\% PQ. It also shows stronger zero-shot robustness over existing methods.-
dc.languageeng-
dc.publisherIEEE.-
dc.relation.ispartofProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022-
dc.rightsProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022. Copyright © IEEE.-
dc.titlePanoptic segformer: Delving deeper into panoptic segmentation with transformers-
dc.typeConference_Paper-
dc.identifier.emailLi, Z: lzq@smail.nju.edu.cn-
dc.identifier.emailWang, W: wangwenhai@pjlab.org.cn-
dc.identifier.emailYu, Z: zhidingy@nvidia.com-
dc.identifier.emailAnandkumar, A: aanandkumar@nvidia.com-
dc.identifier.emailAlvarez, JM: josea@nvidia.com-
dc.identifier.emailLuo, P: pluo@hku.hk-
dc.identifier.emailLu, T: lutong@nju.edu.cn-
dc.identifier.authorityLuo, P=rp02575-
dc.identifier.hkuros335571-
dc.identifier.spage1280-
dc.identifier.epage1289-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats