File Download
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: Cross-Layer Retrospective Retrieving via Layer Attention
Title | Cross-Layer Retrospective Retrieving via Layer Attention |
---|---|
Authors | |
Issue Date | 1-May-2023 |
Abstract | More and more evidence has shown that strengthening layer interactions can enhance the representation power of a deep neural network, while self-attention excels at learning interdependencies by retrieving query-activated information. Motivated by this, we devise a cross-layer attention mechanism, called multi-head recurrent layer attention (MRLA), that sends a query representation of the current layer to all previous layers to retrieve query-related information from different levels of receptive fields. A light-weighted version of MRLA is also proposed to reduce the quadratic computation cost. The proposed layer attention mechanism can enrich the representation power of many state-of-the-art vision networks, including CNNs and vision transformers. Its effectiveness has been extensively evaluated in image classification, object detection and instance segmentation tasks, where improvements can be consistently observed. For example, our MRLA can improve 1.6% Top-1 accuracy on ResNet-50, while only introducing 0.16M parameters and 0.07B FLOPs. Surprisingly, it can boost the performances by a large margin of 3-4% box AP and mask AP in dense prediction tasks. Our code is available at https://github.com/joyfang1106/MRLA. |
Persistent Identifier | http://hdl.handle.net/10722/338246 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Fang, Yanwen | - |
dc.contributor.author | CAI, Yuxi | - |
dc.contributor.author | Chen, Jintai | - |
dc.contributor.author | Zhao, Jingyu | - |
dc.contributor.author | Tian, Guangjian | - |
dc.contributor.author | Li, Guodong | - |
dc.date.accessioned | 2024-03-11T10:27:22Z | - |
dc.date.available | 2024-03-11T10:27:22Z | - |
dc.date.issued | 2023-05-01 | - |
dc.identifier.uri | http://hdl.handle.net/10722/338246 | - |
dc.description.abstract | <p>More and more evidence has shown that strengthening layer interactions can enhance the representation power of a deep neural network, while self-attention excels at learning interdependencies by retrieving query-activated information. Motivated by this, we devise a cross-layer attention mechanism, called multi-head recurrent layer attention (MRLA), that sends a query representation of the current layer to all previous layers to retrieve query-related information from different levels of receptive fields. A light-weighted version of MRLA is also proposed to reduce the quadratic computation cost. The proposed layer attention mechanism can enrich the representation power of many state-of-the-art vision networks, including CNNs and vision transformers. Its effectiveness has been extensively evaluated in image classification, object detection and instance segmentation tasks, where improvements can be consistently observed. For example, our MRLA can improve 1.6% Top-1 accuracy on ResNet-50, while only introducing 0.16M parameters and 0.07B FLOPs. Surprisingly, it can boost the performances by a large margin of 3-4% box AP and mask AP in dense prediction tasks. Our code is available at <a href="https://github.com/joyfang1106/MRLA">https://github.com/joyfang1106/MRLA</a>.<br></p> | - |
dc.language | eng | - |
dc.relation.ispartof | International Conference on Learning Representations (ICLR 2023) (01/05/2023-05/05/2023, Kigali, Rwanda) | - |
dc.title | Cross-Layer Retrospective Retrieving via Layer Attention | - |
dc.type | Conference_Paper | - |
dc.description.nature | preprint | - |