Stacked residual recurrent neural network with word weight for text classification

Wei Cao, Anping Song*, Jinglu Hu

*この研究の対応する著者

研究成果: Article査読

16 被引用数 (Scopus)

抄録

Neural networks, and in particular recurrent neural networks (RNNs) have recently been shown to give a state-ofthe- art performance on some text classification tasks. However, most existing methods assume that each word in a sentence contributes the same importance, it is different from the real world. For example, if we do sentiment analysis, the word "awesome" is much more important than any other words in the sentence "This movie is awesome". Motivated by this deficiency and in order to achieve a further performance, in this paper, a Stacked Residual RNN with Word Weight method is proposed, we extend the stacked RNN to a deep one with residual network architecture and introduce a word weight based network to consider the weight of each word. Our proposed method is able to learn high the hierarchical meaning of each word in a sentence and consider the weight of each word for text classification task. Experimental result indicates that our method achieves high performance compared with the state-of-the-art approaches.

本文言語English
ページ(範囲)277-284
ページ数8
ジャーナルIAENG International Journal of Computer Science
44
3
出版ステータスPublished - 2017 8月 1

ASJC Scopus subject areas

  • コンピュータ サイエンス(全般)

フィンガープリント

「Stacked residual recurrent neural network with word weight for text classification」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル