Many core speech technologies are now highly tuned and rival or exceed human performance. However, the user experience is often lacking, in part because speech synthesizers, speech recognizers, dialog managers, and other components are still inflexible in their expectations of human behavior. Moreover, considerations of style are becoming more important as we seek to deploy variations of one basic system across domains and genres. Thus more attention to interaction style variation and speaking style variation may help us improve the value for users.
Recently work in these areas has been rapidly advancing. We now know more about many types of style, including expressive styles, styles relating to social roles, and styles relating to stance and interpersonal dynamics. There are now many techniques for style in speech synthesis and voice conversion, style transfer in text and speech. The connections of style to genre, priming, accommodation/entrainment, lexical choice, social identity, dialog activity modeling, and next dialog move selection, have been increasingly examined.
However, work on style has been fragmented. There has been little communication between those working on style as it relates to various core technologies, various phenomena, various applications needs, and various linguistic questions.
Accordingly, the aim of this interdisciplinary special session is to bring together researchers who work on different aspects of style for spoken dialog, who work on different tasks, or who approach style from different perspectives. We expect this will foster cross-fertilization, broader understanding of the goals and challenges, and identification of shared priorities for further research.
We welcome submissions on all aspects of style in speech and interaction. All papers must have an empirical component and must make a significant contribution to speech science or speech technology. We further encourage authors to also elaborate on how their work contributes to or relates to "the big picture" of style in spoken dialog. We also welcome papers whose motivations, contributions, or implications relate to more than one traditional Interspeech area, or which highlight issues not commonly addressed at Interspeech.
Questions of interest include:
Nigel Ward, University of Texas at El Paso
Kallirroi Georgila, University of Southern California
Yang Gao , Amazon AWS AI
Mark Hasegawa-Johnson, University of Illinois
Koji Inoue, Kyoto University
Simon King , University of Edinburgh
Rivka Levitan, City University of New York
Katherine Metcalf, Apple
Eva Szekely, KTH Royal Institute for Technology
Pol van Rijn, Max Planck Institute for Empirical Aesthetics
Rafael Valle , NVIDIA