Audio-visual interaction in sparse representation features for noise robust audio-visual speech recognition

Peng Shen, Satoshi Tamura, Satoru Hayamizu

研究成果: Paper査読

1 被引用数 (Scopus)

抄録

In this paper, we investigate audio-visual interaction in sparse representation to obtain robust features for audio-visual speech recognition. Firstly, we introduce our system which uses sparse representation method for noise robust audio-visual speech recognition. Then, we introduce the dictionary matrix used in this paper, and consider the construction of audio-visual dictionary. Finally, we reformulate audio and visual signals as a group sparse representation problem in a combined feature-space domain, and then we improve the joint sparsity feature fusion method with the group sparse representation features and audio sparse representation features. The proposed methods are evaluated using CENSREC-1-AV database with both audio noise and visual noise. From the experimental results, we showed the effectiveness of our proposed method comparing with traditional methods.

本文言語English
ページ43-48
ページ数6
出版ステータスPublished - 2013
外部発表はい
イベント2013 International Conference on Auditory-Visual Speech Processing, AVSP 2013 - Annecy, France
継続期間: 2013 8月 292013 9月 1

Conference

Conference2013 International Conference on Auditory-Visual Speech Processing, AVSP 2013
国/地域France
CityAnnecy
Period13/8/2913/9/1

ASJC Scopus subject areas

  • 言語および言語学
  • 言語聴覚療法
  • 耳鼻咽喉科学

フィンガープリント

「Audio-visual interaction in sparse representation features for noise robust audio-visual speech recognition」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル