2023
pdf
abs
What Makes Pre-trained Language Models Better Zero-shot Learners?
Jinghui Lu
|
Dongsheng Zhu
|
Weidong Han
|
Rui Zhao
|
Brian Mac Namee
|
Fei Tan
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Current methods for prompt learning in zero-shot scenarios widely rely on a development set with sufficient human-annotated data to select the best-performing prompt template a posteriori. This is not ideal because in a real-world zero-shot scenario of practical relevance, no labelled data is available. Thus, we propose a simple yet effective method for screening reasonable prompt templates in zero-shot text classification: Perplexity Selection (Perplection). We hypothesize that language discrepancy can be used to measure the efficacy of prompt templates, and thereby develop a substantiated perplexity-based scheme allowing for forecasting the performance of prompt templates in advance. Experiments show that our method leads to improved prediction performance in a realistic zero-shot setting, eliminating the need for any labelled examples.
2020
pdf
abs
Diverging Divergences: Examining Variants of Jensen Shannon Divergence for Corpus Comparison Tasks
Jinghui Lu
|
Maeve Henchion
|
Brian Mac Namee
Proceedings of the Twelfth Language Resources and Evaluation Conference
Jensen-Shannon divergence (JSD) is a distribution similarity measurement widely used in natural language processing. In corpus comparison tasks, where keywords are extracted to reveal the divergence between different corpora (for example, social media posts from proponents of different views on a political issue), two variants of JSD have emerged in the literature. One of these uses a weighting based on the relative sizes of the corpora being compared. In this paper we argue that this weighting is unnecessary and, in fact, can lead to misleading results. We recommend that this weighted version is not used. We base this recommendation on an analysis of the JSD variants and experiments showing how they impact corpus comparison results as the relative sizes of the corpora being compared change.
2014
pdf
bib
The Effect of Sensor Errors in Situated Human-Computer Dialogue
Niels Schütte
|
John Kelleher
|
Brian Mac Namee
Proceedings of the Third Workshop on Vision and Language
2010
pdf
bib
Proceedings of the 6th International Natural Language Generation Conference
John Kelleher
|
Brian Mac Namee
|
Ielka van der Sluis
Proceedings of the 6th International Natural Language Generation Conference
2008
pdf
Referring Expression Generation Challenge 2008 DIT System Descriptions (DIT-FBI, DIT-TVAS, DIT-CBSR, DIT-RBR, DIT-FBI-CBSR, DIT-TVAS-RBR)
John D. Kelleher
|
Brian Mac Namee
Proceedings of the Fifth International Natural Language Generation Conference