Jianyong Wang
2022
Prompt Tuning for Discriminative Pre-trained Language Models
Yuan Yao
|
Bowen Dong
|
Ao Zhang
|
Zhengyan Zhang
|
Ruobing Xie
|
Zhiyuan Liu
|
Leyu Lin
|
Maosong Sun
|
Jianyong Wang
Findings of the Association for Computational Linguistics: ACL 2022
Recent works have shown promising results of prompt tuning in stimulating pre-trained language models (PLMs) for natural language processing (NLP) tasks. However, to the best of our knowledge, existing works focus on prompt-tuning generative PLMs that are pre-trained to generate target tokens, such as BERT. It is still unknown whether and how discriminative PLMs, e.g., ELECTRA, can be effectively prompt-tuned. In this work, we present DPT, the first prompt tuning framework for discriminative PLMs, which reformulates NLP tasks into a discriminative language modeling problem. Comprehensive experiments on text classification and question answering show that, compared with vanilla fine-tuning, DPT achieves significantly higher performance, and also prevents the unstable problem in tuning large PLMs in both full-set and low-resource settings.
Not Just Plain Text! Fuel Document-Level Relation Extraction with Explicit Syntax Refinement and Subsentence Modeling
Zhichao Duan
|
Xiuxing Li
|
Zhenyu Li
|
Zhuo Wang
|
Jianyong Wang
Findings of the Association for Computational Linguistics: EMNLP 2022
Document-level relation extraction (DocRE) aims to identify semantic labels among entities within a single document. One major challenge of DocRE is to dig decisive details regarding a specific entity pair from long text. However, in many cases, only a fraction of text carries required information, even in the manually labeled supporting evidence. To better capture and exploit instructive information, we propose a novel expLicit syntAx Refinement and Subsentence mOdeliNg based framework (LARSON). By introducing extra syntactic information, LARSON can model subsentences of arbitrary granularity and efficiently screen instructive ones. Moreover, we incorporate refined syntax into text representations which further improves the performance of LARSON. Experimental results on three benchmark datasets (DocRED, CDR, and GDA) demonstrate that LARSON significantly outperforms existing methods.
Search
Co-authors
- Yuan Yao 1
- Bowen Dong 1
- Ao Zhang 1
- Zhengyan Zhang 1
- Ruobing Xie 1
- show all...