English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT

Released

Conference Paper

Human Speech Perception in Noise: Can Large Language Models Paraphrase to Improve It?

MPS-Authors
/persons/resource/persons296430

Demberg,  Vera       
Multimodal Language Processing, MPI for Informatics, Max Planck Society;

External Resource
No external resources are shared
Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)

arXiv:2408.04029.pdf
(Preprint), 551KB

Supplementary Material (public)
There is no public supplementary material available
Citation

Chingacham, A., Zhang, M., Demberg, V., & Klakow, D. (2024). Human Speech Perception in Noise: Can Large Language Models Paraphrase to Improve It? In N. Soni, L. Flek, A. Sharma, D. Yang, S. Hooker, & H. A. Schwartz (Eds.), Proceedings of the 1st Human-Centered Large Language Modeling Workshop (pp. 1-15). Kerrville, TX: ACL. doi:10.18653/v1/2024.hucllm-1.1.


Cite as: https://hdl.handle.net/21.11116/0000-0010-43EE-7
Abstract
Large Language Models (LLMs) can generate text by transferring style
attributes like formality resulting in formal or informal text. However,
instructing LLMs to generate text that when spoken, is more intelligible in an
acoustically difficult environment, is an under-explored topic. We conduct the
first study to evaluate LLMs on a novel task of generating acoustically
intelligible paraphrases for better human speech perception in noise. Our
experiments in English demonstrated that with standard prompting, LLMs struggle
to control the non-textual attribute, i.e., acoustic intelligibility, while
efficiently capturing the desired textual attributes like semantic equivalence.
To remedy this issue, we propose a simple prompting approach,
prompt-and-select, which generates paraphrases by decoupling the desired
textual and non-textual attributes in the text generation pipeline. Our
approach resulted in a 40% relative improvement in human speech perception, by
paraphrasing utterances that are highly distorted in a listening condition with
babble noise at a signal-to-noise ratio (SNR) -5 dB. This study reveals the
limitation of LLMs in capturing non-textual attributes, and our proposed method
showcases the potential of using LLMs for better human speech perception in
noise.