TY - CHAP
T1 - A convolutional attention model for text classification
AU - Du, Jiachen
AU - Gui, Lin
AU - Xu, Ruifeng
AU - He, Yulan
PY - 2018/1/5
Y1 - 2018/1/5
N2 - Neural network models with attention mechanism have shown their efficiencies on various tasks. However, there is little research work on attention mechanism for text classification and existing attention model for text classification lacks of cognitive intuition and mathematical explanation. In this paper, we propose a new architecture of neural network based on the attention model for text classification. In particular, we show that the convolutional neural network (CNN) is a reasonable model for extracting attentions from text sequences in mathematics. We then propose a novel attention model base on CNN and introduce a new network architecture which combines recurrent neural network with our CNN-based attention model. Experimental results on five datasets show that our proposed models can accurately capture the salient parts of sentences to improve the performance of text classification.
AB - Neural network models with attention mechanism have shown their efficiencies on various tasks. However, there is little research work on attention mechanism for text classification and existing attention model for text classification lacks of cognitive intuition and mathematical explanation. In this paper, we propose a new architecture of neural network based on the attention model for text classification. In particular, we show that the convolutional neural network (CNN) is a reasonable model for extracting attentions from text sequences in mathematics. We then propose a novel attention model base on CNN and introduce a new network architecture which combines recurrent neural network with our CNN-based attention model. Experimental results on five datasets show that our proposed models can accurately capture the salient parts of sentences to improve the performance of text classification.
UR - https://www.scopus.com/record/display.uri?eid=2-s2.0-85041173594&origin=SingleRecordEmailAlert&dgcid=raven_sc_search_en_us_email&txGid=1a57e5c3a195f3d54162cee2ae1d99a8
UR - http://link.springer.com/10.1007/978-3-319-73618-1_16
U2 - 10.1007/978-3-319-73618-1_16
DO - 10.1007/978-3-319-73618-1_16
M3 - Other chapter contribution
T3 - Natural Language Processing and Chinese Computing
SP - 183
EP - 195
BT - English
T2 - 6th CCF International Conference
Y2 - 8 November 2017 through 12 November 2017
ER -