TY - JOUR
T1 - DAVID
T2 - An open-source platform for real-time transformation of infra-segmental emotional cues in running speech
AU - Rachman, Laura
AU - Liuni, Marco
AU - Arias, Pablo
AU - Lind, Andreas
AU - Johansson, Petter
AU - Hall, Lars
AU - Richardson, Daniel
AU - Watanabe, Katsumi
AU - Dubal, Stéphanie
AU - Aucouturier, Jean Julien
N1 - Funding Information:
This research was funded by a European Research Council Grant StG-335536 CREAM to JJA and a Japan Science and Technology Agency CREST grant to KW. PJ was supported by the Bank of Sweden Tercentenary Foundation and Swedish Research Council Grant 2014-1371. LH was supported by the Swedish Research Council Grant 2011-1795.
Publisher Copyright:
© 2017, The Author(s).
PY - 2018/2/1
Y1 - 2018/2/1
N2 - We present an open-source software platform that transforms emotional cues expressed by speech signals using audio effects like pitch shifting, inflection, vibrato, and filtering. The emotional transformations can be applied to any audio file, but can also run in real time, using live input from a microphone, with less than 20-ms latency. We anticipate that this tool will be useful for the study of emotions in psychology and neuroscience, because it enables a high level of control over the acoustical and emotional content of experimental stimuli in a variety of laboratory situations, including real-time social situations. We present here results of a series of validation experiments aiming to position the tool against several methodological requirements: that transformed emotions be recognized at above-chance levels, valid in several languages (French, English, Swedish, and Japanese) and with a naturalness comparable to natural speech.
AB - We present an open-source software platform that transforms emotional cues expressed by speech signals using audio effects like pitch shifting, inflection, vibrato, and filtering. The emotional transformations can be applied to any audio file, but can also run in real time, using live input from a microphone, with less than 20-ms latency. We anticipate that this tool will be useful for the study of emotions in psychology and neuroscience, because it enables a high level of control over the acoustical and emotional content of experimental stimuli in a variety of laboratory situations, including real-time social situations. We present here results of a series of validation experiments aiming to position the tool against several methodological requirements: that transformed emotions be recognized at above-chance levels, valid in several languages (French, English, Swedish, and Japanese) and with a naturalness comparable to natural speech.
KW - Emotional transformations
KW - Infra-segmental cues
KW - Nonverbal behavior
KW - Real-time
KW - Software
KW - Voice
UR - http://www.scopus.com/inward/record.url?scp=85016981246&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85016981246&partnerID=8YFLogxK
U2 - 10.3758/s13428-017-0873-y
DO - 10.3758/s13428-017-0873-y
M3 - Article
C2 - 28374144
AN - SCOPUS:85016981246
SN - 1554-351X
VL - 50
SP - 323
EP - 343
JO - Behavior Research Methods
JF - Behavior Research Methods
IS - 1
ER -