@article{zhou-etal-2021-self-supervised,
title = "Self-supervised Regularization for Text Classification",
author = "Zhou, Meng and
Li, Zechen and
Xie, Pengtao",
editor = "Roark, Brian and
Nenkova, Ani",
journal = "Transactions of the Association for Computational Linguistics",
volume = "9",
year = "2021",
address = "Cambridge, MA",
publisher = "MIT Press",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2021.tacl-1.39/",
doi = "10.1162/tacl_a_00389",
pages = "641--656",
abstract = "Text classification is a widely studied problem and has broad applications. In many real-world problems, the number of texts for training classification models is limited, which renders these models prone to overfitting. To address this problem, we propose SSL-Reg, a data-dependent regularization approach based on self-supervised learning (SSL). SSL (Devlin et al., 2019a) is an unsupervised learning approach that defines auxiliary tasks on input data without using any human-provided labels and learns data representations by solving these auxiliary tasks. In SSL-Reg, a supervised classification task and an unsupervised SSL task are performed simultaneously. The SSL task is unsupervised, which is defined purely on input texts without using any human- provided labels. Training a model using an SSL task can prevent the model from being overfitted to a limited number of class labels in the classification task. Experiments on 17 text classification datasets demonstrate the effectiveness of our proposed method. Code is available at \url{https://github.com/UCSD-AI4H/SSReg}."
}
Markdown (Informal)
[Self-supervised Regularization for Text Classification](https://preview.aclanthology.org/jlcl-multiple-ingestion/2021.tacl-1.39/) (Zhou et al., TACL 2021)
ACL