TY - GEN
T1 - Utility of gestural cues in indexing semantic miscommunication
AU - Inoue, Masashi
AU - Ogihara, Mitsunori
AU - Hanada, Ryoko
AU - Furuyama, Nobuhiro
PY - 2010/7/14
Y1 - 2010/7/14
N2 - In multimedia data analysis, automated indexing of conversational video is an emerging topic. One challenging problem in this topic is the recognition of higher-level concepts, such as miscommunications in conversations. While detecting miscommunications is generally easy for speakers as well as observers, it is not currently understood which cues contribute to their detection and to what extent. To make use of the knowledge on gestural cues in multimedia systems, the applicability of machine learning is investigated as a means of detecting miscommunication from gestural patterns observed in psychotherapeutic face-to-face conversations. Various features are taken from gesture data, and both simple and complex classifiers are constructed using these features. Both short-term and long-term effects are tested using different time window sizes. Also, two types of gestures, communicative and non-communicative, are considered. The experimental results suggest that there is no single gestural feature that can explain the occurrence of semantic miscommunication. Another interesting finding is that gestural cues correlate more with long-term gestural patterns than with short-term ones.
AB - In multimedia data analysis, automated indexing of conversational video is an emerging topic. One challenging problem in this topic is the recognition of higher-level concepts, such as miscommunications in conversations. While detecting miscommunications is generally easy for speakers as well as observers, it is not currently understood which cues contribute to their detection and to what extent. To make use of the knowledge on gestural cues in multimedia systems, the applicability of machine learning is investigated as a means of detecting miscommunication from gestural patterns observed in psychotherapeutic face-to-face conversations. Various features are taken from gesture data, and both simple and complex classifiers are constructed using these features. Both short-term and long-term effects are tested using different time window sizes. Also, two types of gestures, communicative and non-communicative, are considered. The experimental results suggest that there is no single gestural feature that can explain the occurrence of semantic miscommunication. Another interesting finding is that gestural cues correlate more with long-term gestural patterns than with short-term ones.
KW - Face-toface
KW - Gesture
KW - Psychotherapy
KW - Semantic indexing
UR - http://www.scopus.com/inward/record.url?scp=77954400664&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=77954400664&partnerID=8YFLogxK
U2 - 10.1109/FUTURETECH.2010.5482653
DO - 10.1109/FUTURETECH.2010.5482653
M3 - Conference contribution
AN - SCOPUS:77954400664
SN - 9781424469505
T3 - 2010 5th International Conference on Future Information Technology, FutureTech 2010 - Proceedings
BT - 2010 5th International Conference on Future Information Technology, FutureTech 2010 - Proceedings
T2 - 5th International Conference on Future Information Technology, FutureTech 2010
Y2 - 20 May 2010 through 24 May 2010
ER -