Please use this identifier to cite or link to this item:
http://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/16336
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Sharma, Yashvardhan | - |
dc.date.accessioned | 2024-11-12T07:05:14Z | - |
dc.date.available | 2024-11-12T07:05:14Z | - |
dc.date.issued | 2024-07 | - |
dc.identifier.uri | https://arxiv.org/abs/2407.19526 | - |
dc.identifier.uri | http://dspace.bits-pilani.ac.in:8080/jspui/handle/123456789/16336 | - |
dc.description.abstract | To be included into chatbot systems, Large language models (LLMs) must be aligned with human conversational conventions. However, being trained mainly on web-scraped data gives existing LLMs a voice closer to informational text than actual human speech. In this paper, we examine the effect of decoding methods on the alignment between LLM-generated and human conversations, including Beam Search, Top K Sampling, and Nucleus Sampling. We present new measures of alignment in substance, style, and psychometric orientation, and experiment with two conversation datasets. Our results provide subtle insights: better alignment is attributed to fewer beams in Beam Search and lower values of P in Nucleus Sampling. We also find that task-oriented and open-ended datasets perform differently in terms of alignment, indicating the significance of taking into account the context of the interaction. | en_US |
dc.language.iso | en | en_US |
dc.subject | Computer Science | en_US |
dc.subject | Chatbot systems | en_US |
dc.subject | Large Language Models (LLMs) | en_US |
dc.subject | Human Alignment | en_US |
dc.title | Impact of Decoding Methods on Human Alignment of Conversational LLMs | en_US |
dc.type | Preprint | en_US |
Appears in Collections: | Department of Computer Science and Information Systems |
Files in This Item:
There are no files associated with this item.
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.