File Download

There are no files associated with this item.

Supplementary

Conference Paper: Learning versatile neural architectures by propagating network codes

TitleLearning versatile neural architectures by propagating network codes
Authors
KeywordsMultitask NAS
Task-Transferable Architecture
Neural Predictor
NAS Benchmark
Issue Date2022
PublisherICLR.
Citation
10th International Conference on Learning Representation (ICLR) (Virtual), April 25-29, 2022 How to Cite?
AbstractThis work explores how to design a single neural network capable of adapting to multiple heterogeneous vision tasks, such as image segmentation, 3D detection, and video recognition. This goal is challenging because both network architecture search (NAS) spaces and methods in different tasks are inconsistent. We solve this challenge from both sides. We first introduce a unified design space for multiple tasks and build a multitask NAS benchmark (NAS-Bench-MR) on many widely used datasets, including ImageNet, Cityscapes, KITTI, and HMDB51. We further propose Network Coding Propagation (NCP), which back-propagates gradients of neural predictors to directly update architecture codes along the desired gradient directions to solve various tasks. In this way, optimal architecture configurations can be found by NCP in our large search space in seconds. Unlike prior arts of NAS that typically focus on a single task, NCP has several unique benefits. (1) NCP transforms architecture optimization from data-driven to architecture-driven, enabling joint search an architecture among multitasks with different data distributions. (2) NCP learns from network codes but not original data, enabling it to update the architecture efficiently across datasets. (3) In addition to our NAS-Bench-MR, NCP performs well on other NAS benchmarks, such as NAS-Bench-201. (4) Thorough studies of NCP on inter-, cross-, and intra-tasks highlight the importance of cross-task neural architecture design, i.e., multitask neural architectures and architecture transferring between different tasks.
DescriptionPoster Session 9
Persistent Identifierhttp://hdl.handle.net/10722/315800

 

DC FieldValueLanguage
dc.contributor.authorDing, M-
dc.contributor.authorHuo, Y-
dc.contributor.authorLu, H-
dc.contributor.authorYang, L-
dc.contributor.authorWang, Z-
dc.contributor.authorLu, Z-
dc.contributor.authorWang, J-
dc.contributor.authorLuo, P-
dc.date.accessioned2022-08-19T09:04:40Z-
dc.date.available2022-08-19T09:04:40Z-
dc.date.issued2022-
dc.identifier.citation10th International Conference on Learning Representation (ICLR) (Virtual), April 25-29, 2022-
dc.identifier.urihttp://hdl.handle.net/10722/315800-
dc.descriptionPoster Session 9-
dc.description.abstractThis work explores how to design a single neural network capable of adapting to multiple heterogeneous vision tasks, such as image segmentation, 3D detection, and video recognition. This goal is challenging because both network architecture search (NAS) spaces and methods in different tasks are inconsistent. We solve this challenge from both sides. We first introduce a unified design space for multiple tasks and build a multitask NAS benchmark (NAS-Bench-MR) on many widely used datasets, including ImageNet, Cityscapes, KITTI, and HMDB51. We further propose Network Coding Propagation (NCP), which back-propagates gradients of neural predictors to directly update architecture codes along the desired gradient directions to solve various tasks. In this way, optimal architecture configurations can be found by NCP in our large search space in seconds. Unlike prior arts of NAS that typically focus on a single task, NCP has several unique benefits. (1) NCP transforms architecture optimization from data-driven to architecture-driven, enabling joint search an architecture among multitasks with different data distributions. (2) NCP learns from network codes but not original data, enabling it to update the architecture efficiently across datasets. (3) In addition to our NAS-Bench-MR, NCP performs well on other NAS benchmarks, such as NAS-Bench-201. (4) Thorough studies of NCP on inter-, cross-, and intra-tasks highlight the importance of cross-task neural architecture design, i.e., multitask neural architectures and architecture transferring between different tasks.-
dc.languageeng-
dc.publisherICLR.-
dc.subjectMultitask NAS-
dc.subjectTask-Transferable Architecture-
dc.subjectNeural Predictor-
dc.subjectNAS Benchmark-
dc.titleLearning versatile neural architectures by propagating network codes-
dc.typeConference_Paper-
dc.identifier.emailLuo, P: pluo@hku.hk-
dc.identifier.authorityLuo, P=rp02575-
dc.identifier.hkuros335592-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats