Hello Valeriy!
Sharon Oviatt had a couple of papers looking at something like this (references below) – talkers go into a “hyperspeech” when a computer mis-recognizes what they
say.
Oviatt, S., Levow, G. A., Moreton, E., & MacEachern, M. (1998). Modeling global and focal hyperarticulation during human-computer error resolution.
Journal of the Acoustical Society of America, 104, 3080-3098. doi:10.1121/1.423888
Oviatt, S., MacEachern, M., & Levow, G. A. (1998). Predicting hyperarticulate speech during human-computer error resolution.
Speech Communication, 24, 87-110. doi:10.1016/S0167-6393(98)00005-3 I don’t know if the idea came from these papers or if I heard it somewhere else, but speech recognizers are (or used to be) trained up on citation-style speech,
so hyperarticulation should make speech recognition worse. I’ve been surprised by how well Siri does and will sometimes try to “mess with´ it to see how it behaves.
Sarah Hargus Ferguson, PhD, CCC-A Associate Professor Department of Communication Sciences and Disorders From: AUDITORY - Research in Auditory Perception <AUDITORY@xxxxxxxxxxxxxxx>
On Behalf Of Valeriy Shafiro Dear list, I am wondering if any one has any references or suggestions about this question. These days I hear more and more
people talking to machines, e.g. Siri, Google, Alexa, etc., and doing it in more and more places. Automatic speech recognition has improved tremendously, but still it seems to me that when people talk to machines they often switch into a different production
mode. At times it may sound like talking to a (large) dog and sometimes like talking to a customer service agent in a land far away who is diligently trying to follow a script rather than listen to what you are saying. And I wonder whether adjustments
that people make in their speech production when talking with machines in that mode are in fact optimal for improving recognition accuracy. Since machines are not processing speech in the same way as humans, I wonder if changes in speech production that make
speech more recognizable for other people (or even pets) are always the same as they are for machines. In other words, do people tend to make the most optimal adjustments to make their speech more recognizable to machines. Or is it more like falling back
on clear speech modes that work with other kinds of listeners (children, nonnative speakers, pets), or something in between? I realize there is a lot to this question, but perhaps people have started looking into it. I am happy to collate
references and replies and send to the list. Best, Valeriy |