File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: How Does Semi-supervised learning with Pseudo-labelers Work? A Case Study
Title | How Does Semi-supervised learning with Pseudo-labelers Work? A Case Study |
---|---|
Authors | |
Issue Date | 5-May-2023 |
Abstract | Semi-supervised learning is a popular machine learning paradigm that utilizes a large amount of unlabeled data as well as a small amount of labeled data to facilitate learning tasks. While semi-supervised learning has achieved great success in training neural networks, its theoretical understanding remains largely open. In this paper, we aim to theoretically understand a semi-supervised learning approach based on pre-training and linear probing. In particular, the semi-supervised learning approach we consider first trains a two-layer neural network based on the unlabeled data with the help of pseudo-labelers. Then it linearly probes the pre-trained network on a small amount of labeled data. We prove that, under a certain toy data generation model and two-layer convolutional neural network, the semisupervised learning approach can achieve nearly zero test loss, while a neural network directly trained by supervised learning on the same amount of labeled data can only achieve constant test loss. Through this case study, we demonstrate a separation between semi-supervised learning and supervised learning in terms of test loss provided the same amount of labeled data. |
Persistent Identifier | http://hdl.handle.net/10722/338363 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Kou, Yiwen | - |
dc.contributor.author | Chen, Zixiang | - |
dc.contributor.author | Cao, Yuan | - |
dc.contributor.author | Gu, Quanquan | - |
dc.date.accessioned | 2024-03-11T10:28:18Z | - |
dc.date.available | 2024-03-11T10:28:18Z | - |
dc.date.issued | 2023-05-05 | - |
dc.identifier.uri | http://hdl.handle.net/10722/338363 | - |
dc.description.abstract | <p>Semi-supervised learning is a popular machine learning paradigm that utilizes a large amount of unlabeled data as well as a small amount of labeled data to facilitate learning tasks. While semi-supervised learning has achieved great success in training neural networks, its theoretical understanding remains largely open. In this paper, we aim to theoretically understand a semi-supervised learning approach based on pre-training and linear probing. In particular, the semi-supervised learning approach we consider first trains a two-layer neural network based on the unlabeled data with the help of pseudo-labelers. Then it linearly probes the pre-trained network on a small amount of labeled data. We prove that, under a certain toy data generation model and two-layer convolutional neural network, the semisupervised learning approach can achieve nearly zero test loss, while a neural network directly trained by supervised learning on the same amount of labeled data can only achieve constant test loss. Through this case study, we demonstrate a separation between semi-supervised learning and supervised learning in terms of test loss provided the same amount of labeled data.<br></p> | - |
dc.language | eng | - |
dc.relation.ispartof | The 11th International Conference on Learning Representations (ICLR 2023) (01/05/2023-05/05/2023, Kigali, Rwanda) | - |
dc.title | How Does Semi-supervised learning with Pseudo-labelers Work? A Case Study | - |
dc.type | Conference_Paper | - |