Video semantic indexing using object detection-derived features

Kotaro Kikuchi, Kazuya Ueki, Tetsuji Ogawa, Tetsunori Kobayashi

研究成果: Conference contribution

3 被引用数 (Scopus)

抄録

A new feature extraction method based on object detection to achieve accurate and robust semantic indexing of videos is proposed. Local features (e.g., SIFT and HOG) and convolutional neural network (CNN)-derived features, which have been used in semantic indexing, in general are extracted from the entire image and do not explicitly represent the information of meaningful objects that contributes to the determination of semantic categories. In this case, the background region, which does not contain the meaningful objects, is unduly considered, exerting a harmful effect on the indexing performance. In the present study, an attempt was made to suppress the undesirable effects derived from the redundant background information by incorporating object detection technology into semantic indexing. In the proposed method, a combination of the meaningful objects detected in the video frame image is represented as a feature vector for verification of semantic categories. Experimental comparisons demonstrate that the proposed method facilitates the TRECVID semantic indexing task.

本文言語English
ホスト出版物のタイトル2016 24th European Signal Processing Conference, EUSIPCO 2016
出版社European Signal Processing Conference, EUSIPCO
ページ1288-1292
ページ数5
ISBN(電子版)9780992862657
DOI
出版ステータスPublished - 2016 11月 28
イベント24th European Signal Processing Conference, EUSIPCO 2016 - Budapest, Hungary
継続期間: 2016 8月 282016 9月 2

出版物シリーズ

名前European Signal Processing Conference
2016-November
ISSN(印刷版)2219-5491

Other

Other24th European Signal Processing Conference, EUSIPCO 2016
国/地域Hungary
CityBudapest
Period16/8/2816/9/2

ASJC Scopus subject areas

  • 信号処理
  • 電子工学および電気工学

フィンガープリント

「Video semantic indexing using object detection-derived features」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル