File Download

There are no files associated with this item.

Supplementary

Conference Paper: Dynamic token normalization improves vision transformer

TitleDynamic token normalization improves vision transformer
Authors
KeywordsClassification
Normalization
Transformer
Issue Date2022
PublisherICLR.
Citation
10th International Conference on Learning Representation (ICLR) (Virtual), April 25-29, 2022 How to Cite?
AbstractVision Transformer (ViT) and its variants (e.g., Swin, PVT) have achieved great success in various computer vision tasks, owing to their capability to learn long-range contextual information. Layer Normalization (LN) is an essential ingredient in these models. However, we found that the ordinary LN makes tokens at different positions similar in magnitude because it normalizes embeddings within each token. It is difficult for Transformers to capture inductive bias such as the positional context in an image with LN. We tackle this problem by proposing a new normalizer, termed Dynamic Token Normalization (DTN), where normalization is performed both within each token (intra-token) and across different tokens (inter-token). DTN has several merits. Firstly, it is built on a unified formulation and thus can represent various existing normalization methods. Secondly, DTN learns to normalize tokens in both intra-token and inter-token manners, enabling Transformers to capture both the global contextual information and the local positional context. {Thirdly, by simply replacing LN layers, DTN can be readily plugged into various vision transformers, such as ViT, Swin, PVT, LeViT, T2T-ViT, BigBird and Reformer. Extensive experiments show that the transformer equipped with DTN consistently outperforms baseline model with minimal extra parameters and computational overhead. For example, DTN outperforms LN by - top-1 accuracy on ImageNet, by - box AP in object detection on COCO benchmark, by - mCE in robustness experiments on ImageNet-C, and by - accuracy in Long ListOps on Long-Range Arena.} Codes will be made public at url{https://github.com/wqshao126/DTN}
DescriptionPoster
Persistent Identifierhttp://hdl.handle.net/10722/315683

 

DC FieldValueLanguage
dc.contributor.authorShao, W-
dc.contributor.authorGe, Y-
dc.contributor.authorZhang, Z-
dc.contributor.authorXu, X-
dc.contributor.authorWang, X-
dc.contributor.authorShan, Y-
dc.contributor.authorLuo, P-
dc.date.accessioned2022-08-19T09:02:29Z-
dc.date.available2022-08-19T09:02:29Z-
dc.date.issued2022-
dc.identifier.citation10th International Conference on Learning Representation (ICLR) (Virtual), April 25-29, 2022-
dc.identifier.urihttp://hdl.handle.net/10722/315683-
dc.descriptionPoster-
dc.description.abstractVision Transformer (ViT) and its variants (e.g., Swin, PVT) have achieved great success in various computer vision tasks, owing to their capability to learn long-range contextual information. Layer Normalization (LN) is an essential ingredient in these models. However, we found that the ordinary LN makes tokens at different positions similar in magnitude because it normalizes embeddings within each token. It is difficult for Transformers to capture inductive bias such as the positional context in an image with LN. We tackle this problem by proposing a new normalizer, termed Dynamic Token Normalization (DTN), where normalization is performed both within each token (intra-token) and across different tokens (inter-token). DTN has several merits. Firstly, it is built on a unified formulation and thus can represent various existing normalization methods. Secondly, DTN learns to normalize tokens in both intra-token and inter-token manners, enabling Transformers to capture both the global contextual information and the local positional context. {Thirdly, by simply replacing LN layers, DTN can be readily plugged into various vision transformers, such as ViT, Swin, PVT, LeViT, T2T-ViT, BigBird and Reformer. Extensive experiments show that the transformer equipped with DTN consistently outperforms baseline model with minimal extra parameters and computational overhead. For example, DTN outperforms LN by - top-1 accuracy on ImageNet, by - box AP in object detection on COCO benchmark, by - mCE in robustness experiments on ImageNet-C, and by - accuracy in Long ListOps on Long-Range Arena.} Codes will be made public at url{https://github.com/wqshao126/DTN}-
dc.languageeng-
dc.publisherICLR.-
dc.relation.ispartofAdvances in Neural Information Processing Systems (NeurlPS)-
dc.subjectClassification-
dc.subjectNormalization-
dc.subjectTransformer-
dc.titleDynamic token normalization improves vision transformer-
dc.typeConference_Paper-
dc.identifier.emailLuo, P: pluo@hku.hk-
dc.identifier.authorityLuo, P=rp02575-
dc.identifier.hkuros335599-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats