File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: Channel equilibrium networks for learning deep representation
Title | Channel equilibrium networks for learning deep representation |
---|---|
Authors | |
Issue Date | 2020 |
Publisher | ML Research Press. The Journal's web site is located at http://proceedings.mlr.press/ |
Citation | Thirty-seventh International Conference on Machine Learning (ICML 2020), Vienna, Austria, 12-18 July 2020. In Proceedings of Machine Learning Research (PMLR), v. 119: Proceedings of ICML 2020, p. 8645-8654 How to Cite? |
Abstract | Convolutional Neural Networks (CNNs) are typically constructed by stacking multiple building blocks, each of which contains a normalization layer such as batch normalization (BN) and a rectified linear function such as ReLU. However, this work shows that the combination of normalization and rectified linear function leads to inhibited channels, which have small magnitude and contribute little to the learned feature representation, impeding the generalization ability of CNNs. Unlike prior arts that simply removed the inhibited channels, we propose to wake them up'' during training by designing a novel neural building block, termed Channel Equilibrium (CE) block, which enables channels at the same layer to contribute equally to the learned representation. We show that CE is able to prevent inhibited channels both empirically and theoretically. CE has several appealing benefits. (1) It can be integrated into many advanced CNN architectures such as ResNet and MobileNet, outperforming their original networks. (2) CE has an interesting connection with the Nash Equilibrium, a well-known solution of a non-cooperative game. (3) Extensive experiments show that CE achieves state-of-the-art performance on various challenging benchmarks such as ImageNet and COCO. |
Description | ICML 2020 held virtually due to COVID-19 |
Persistent Identifier | http://hdl.handle.net/10722/284167 |
ISSN |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Shao, W | - |
dc.contributor.author | Tang, S | - |
dc.contributor.author | Pan, X | - |
dc.contributor.author | Tan, P | - |
dc.contributor.author | Wang, X | - |
dc.contributor.author | Luo, P | - |
dc.date.accessioned | 2020-07-20T05:56:37Z | - |
dc.date.available | 2020-07-20T05:56:37Z | - |
dc.date.issued | 2020 | - |
dc.identifier.citation | Thirty-seventh International Conference on Machine Learning (ICML 2020), Vienna, Austria, 12-18 July 2020. In Proceedings of Machine Learning Research (PMLR), v. 119: Proceedings of ICML 2020, p. 8645-8654 | - |
dc.identifier.issn | 2640-3498 | - |
dc.identifier.uri | http://hdl.handle.net/10722/284167 | - |
dc.description | ICML 2020 held virtually due to COVID-19 | - |
dc.description.abstract | Convolutional Neural Networks (CNNs) are typically constructed by stacking multiple building blocks, each of which contains a normalization layer such as batch normalization (BN) and a rectified linear function such as ReLU. However, this work shows that the combination of normalization and rectified linear function leads to inhibited channels, which have small magnitude and contribute little to the learned feature representation, impeding the generalization ability of CNNs. Unlike prior arts that simply removed the inhibited channels, we propose to wake them up'' during training by designing a novel neural building block, termed Channel Equilibrium (CE) block, which enables channels at the same layer to contribute equally to the learned representation. We show that CE is able to prevent inhibited channels both empirically and theoretically. CE has several appealing benefits. (1) It can be integrated into many advanced CNN architectures such as ResNet and MobileNet, outperforming their original networks. (2) CE has an interesting connection with the Nash Equilibrium, a well-known solution of a non-cooperative game. (3) Extensive experiments show that CE achieves state-of-the-art performance on various challenging benchmarks such as ImageNet and COCO. | - |
dc.language | eng | - |
dc.publisher | ML Research Press. The Journal's web site is located at http://proceedings.mlr.press/ | - |
dc.relation.ispartof | Proceedings of Machine Learning Research (PMLR) | - |
dc.relation.ispartof | The 37th International Conference on Machine Learning (ICML 2020) | - |
dc.title | Channel equilibrium networks for learning deep representation | - |
dc.type | Conference_Paper | - |
dc.identifier.email | Luo, P: pluo@hku.hk | - |
dc.identifier.authority | Luo, P=rp02575 | - |
dc.identifier.hkuros | 311028 | - |
dc.identifier.volume | 119: Proceedings of ICML 2020 | - |
dc.identifier.spage | 8645 | - |
dc.identifier.epage | 8654 | - |
dc.publisher.place | United States | - |