English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Conference Paper

Characterizing similarities and divergences in conversational tones in humans and LLMs by sampling with people

MPS-Authors

Huang,  Dun-Ming
Department of Electrical Engineering and Computer Sciences, University of California ;
Research Group Computational Auditory Perception, Max Planck Institute for Empirical Aesthetics, Max Planck Society;

/persons/resource/persons255681

van Rijn,  Pol       
Research Group Computational Auditory Perception, Max Planck Institute for Empirical Aesthetics, Max Planck Society;

/persons/resource/persons242173

Jacoby,  Nori       
Research Group Computational Auditory Perception, Max Planck Institute for Empirical Aesthetics, Max Planck Society;

External Resource
No external resources are shared
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)

cap-25-rij-01-characterizing.pdf
(Publisher version), 7MB

Supplementary Material (public)
There is no public supplementary material available
Citation

Huang, D.-M., van Rijn, P., Sucholutsky, I., Marjieh, R., & Jacoby, N. (2024). Characterizing similarities and divergences in conversational tones in humans and LLMs by sampling with people. In Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (pp. 10486-10512). Bangkok, Thailand: Association for Computational Linguistics. doi:10.18653/v1/2024.acl-long.565.


Cite as: https://hdl.handle.net/21.11116/0000-0011-0DEA-8
Abstract
Conversational tones — the manners and attitudes in which speakers communicate — are essential to effective communication. As Large Language Models (LLMs) become increasingly popular, it is necessary to characterize the divergences in their conversational tones relative to humans. Prior research relied on pre-existing taxonomies or text corpora, which suffer from experimenter bias and may not be representative of real-world distributions. Inspired by methods from cognitive science, we propose an iterative method for simultaneously eliciting conversational tones and sentences, where participants alternate between two tasks: (1) one participant identifies the tone of a given sentence and (2) a different participant generates a sentence based on that tone. We run 50 iterations of this process with both human participants and GPT-4 and obtain a dataset of sentences and frequent conversational tones. In an additional experiment, humans and GPT-4 annotated all sentences with all tones. With data from 1,339 participants, 33,370 human judgments, and 29,900 GPT-4 queries, we show how our approach can be used to create an interpretable geometric representation of relations between tones in humans and GPT-4. This work showcases how combining ideas from machine learning and cognitive science can address challenges in human-computer interactions.