File Download

There are no files associated with this item.

Supplementary

Conference Paper: Understanding Train-Validation Split in Meta-Learning with Neural Networks

TitleUnderstanding Train-Validation Split in Meta-Learning with Neural Networks
Authors
Issue Date5-May-2023
Abstract

The goal of meta-learning is to learn a good prior model from a collection of tasks such that the learned prior is able to adapt quickly to new tasks without accessing many data from the new tasks. A common practice in meta-learning is to perform a train-validation split on each task, where the training set is used for adapting the model parameter to that specific task and the validation set is used for learning a prior model that is shared across all tasks. Despite its success and popularity in multitask learning and few-shot learning, the understanding of the train-validation split is still limited, especially when the neural network models are used. In this paper, we study the benefit of train-validation split for classification problems with neural network models trained by gradient descent. We prove that the train-validation split is necessary to learn a good prior model when the noise in the training sample is large, while the train-train method fails. We validate our theory by conducting experiment on both synthetic and real datasets. To the best of our knowledge, this is the first work towards the theoretical understanding of train-validation split in meta-learning with neural networks.


Persistent Identifierhttp://hdl.handle.net/10722/338365

 

DC FieldValueLanguage
dc.contributor.authorZuo, Xinzhe-
dc.contributor.authorChen, Zixiang-
dc.contributor.authorYao, Huaxiu-
dc.contributor.authorCao, Yuan-
dc.contributor.authorGu, Quanquan-
dc.date.accessioned2024-03-11T10:28:19Z-
dc.date.available2024-03-11T10:28:19Z-
dc.date.issued2023-05-05-
dc.identifier.urihttp://hdl.handle.net/10722/338365-
dc.description.abstract<p>The goal of meta-learning is to learn a good prior model from a collection of tasks such that the learned prior is able to adapt quickly to new tasks without accessing many data from the new tasks. A common practice in meta-learning is to perform a train-validation split on each task, where the training set is used for adapting the model parameter to that specific task and the validation set is used for learning a prior model that is shared across all tasks. Despite its success and popularity in multitask learning and few-shot learning, the understanding of the train-validation split is still limited, especially when the neural network models are used. In this paper, we study the benefit of train-validation split for classification problems with neural network models trained by gradient descent. We prove that the train-validation split is necessary to learn a good prior model when the noise in the training sample is large, while the train-train method fails. We validate our theory by conducting experiment on both synthetic and real datasets. To the best of our knowledge, this is the first work towards the theoretical understanding of train-validation split in meta-learning with neural networks.</p>-
dc.languageeng-
dc.relation.ispartofThe 11th International Conference on Learning Representations (ICLR 2023) (01/05/2023-05/05/2023, Kigali, Rwanda)-
dc.titleUnderstanding Train-Validation Split in Meta-Learning with Neural Networks-
dc.typeConference_Paper-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats