@inproceedings{furniturewala-etal-2024-impact,
title = "Impact of Decoding Methods on Human Alignment of Conversational {LLM}s",
author = "Furniturewala, Shaz and
Jaidka, Kokil and
Sharma, Yashvardhan",
editor = "De Clercq, Orph{\'e}e and
Barriere, Valentin and
Barnes, Jeremy and
Klinger, Roman and
Sedoc, Jo{\~a}o and
Tafreshi, Shabnam",
booktitle = "Proceedings of the 14th Workshop on Computational Approaches to Subjectivity, Sentiment, {\&} Social Media Analysis",
month = aug,
year = "2024",
address = "Bangkok, Thailand",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.wassa-1.22/",
doi = "10.18653/v1/2024.wassa-1.22",
pages = "273--279",
abstract = "To be included into chatbot systems, Large language models (LLMs) must be aligned with human conversational conventions. However, being trained mainly on web-scraped data gives existing LLMs a voice closer to informational text than actual human speech. In this paper, we examine the effect of decoding methods on the alignment between LLM-generated and human conversations, including Beam Search, Top K Sampling, and Nucleus Sampling. We present new measures of alignment in substance, style, and psychometric orientation, and experiment with two conversation datasets. Our results provide subtle insights: better alignment is attributed to fewer beams in Beam Search and lower values of P in Nucleus Sampling. We also find that task-oriented and open-ended datasets perform differently in terms of alignment, indicating the significance of taking into account the context of the interaction."
}
Markdown (Informal)
[Impact of Decoding Methods on Human Alignment of Conversational LLMs](https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.wassa-1.22/) (Furniturewala et al., WASSA 2024)
ACL