@inproceedings{bhattarai-erk-2024-learn,
title = "To Learn or Not to Learn: Replaced Token Detection for Learning the Meaning of Negation",
author = "Bhattarai, Gunjan and
Erk, Katrin",
editor = "Calzolari, Nicoletta and
Kan, Min-Yen and
Hoste, Veronique and
Lenci, Alessandro and
Sakti, Sakriani and
Xue, Nianwen",
booktitle = "Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)",
month = may,
year = "2024",
address = "Torino, Italia",
publisher = "ELRA and ICCL",
url = "https://preview.aclanthology.org/fix-sig-urls/2024.lrec-main.1411/",
pages = "16237--16250",
abstract = "State-of-the-art language models perform well on a variety of language tasks, but they continue to struggle with understanding negation cues in tasks like natural language inference (NLI). Inspired by Hossain et al. (2020), who show under-representation of negation in language model pretraining datasets, we experiment with additional pretraining with negation data for which we introduce two new datasets. We also introduce a new learning strategy for negation building on ELECTRA{'}s (Clark et al., 2020) replaced token detection objective. We find that continuing to pretrain ELECTRA-Small{'}s discriminator leads to substantial gains on a variant of RTE (Recognizing Textual Entailment) with additional negation. On SNLI (Stanford NLI) (Bowman et al., 2015), there are no gains due to the extreme under-representation of negation in the data. Finally, on MNLI (Multi-NLI) (Williams et al., 2018), we find that performance on negation cues is primarily stymied by neutral-labeled examples."
}
Markdown (Informal)
[To Learn or Not to Learn: Replaced Token Detection for Learning the Meaning of Negation](https://preview.aclanthology.org/fix-sig-urls/2024.lrec-main.1411/) (Bhattarai & Erk, LREC-COLING 2024)
ACL