DSpace Repository

Impact of Decoding Methods on Human Alignment of Conversational LLMs

Show simple item record

dc.contributor.author Sharma, Yashvardhan
dc.date.accessioned 2024-11-12T07:05:14Z
dc.date.available 2024-11-12T07:05:14Z
dc.date.issued 2024-07
dc.identifier.uri https://arxiv.org/abs/2407.19526
dc.identifier.uri http://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/16336
dc.description.abstract To be included into chatbot systems, Large language models (LLMs) must be aligned with human conversational conventions. However, being trained mainly on web-scraped data gives existing LLMs a voice closer to informational text than actual human speech. In this paper, we examine the effect of decoding methods on the alignment between LLM-generated and human conversations, including Beam Search, Top K Sampling, and Nucleus Sampling. We present new measures of alignment in substance, style, and psychometric orientation, and experiment with two conversation datasets. Our results provide subtle insights: better alignment is attributed to fewer beams in Beam Search and lower values of P in Nucleus Sampling. We also find that task-oriented and open-ended datasets perform differently in terms of alignment, indicating the significance of taking into account the context of the interaction. en_US
dc.language.iso en en_US
dc.subject Computer Science en_US
dc.subject Chatbot systems en_US
dc.subject Large Language Models (LLMs) en_US
dc.subject Human Alignment en_US
dc.title Impact of Decoding Methods on Human Alignment of Conversational LLMs en_US
dc.type Preprint en_US


Files in this item

Files Size Format View

There are no files associated with this item.

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account