Title
Hierarchical gated recurrent neural network with adversarial and virtual adversarial training on text classification.
Abstract
Document classification aims to assign one or more classes to a document for ease of management by understanding the content of a document. Hierarchical attention network (HAN) has been showed effective to classify documents that are ambiguous. HAN parses information-intense documents into slices (i.e., words and sentences) such that each slice can be learned separately and in parallel before assigning the classes. However, introducing hierarchical attention approach leads to the redundancy of training parameters which is prone to overfitting. To mitigate the concern of overfitting, we propose a variant of hierarchical attention network using adversarial and virtual adversarial perturbations in 1) word representation, 2) sentence representation and 3) both word and sentence representations. The proposed variant is tested on eight publicly available datasets. The results show that the proposed variant outperforms the hierarchical attention network with and without using random perturbation. More importantly, the proposed variant achieves state-of-the-art performance on multiple benchmark datasets. Visualizations and analysis are provided to show that perturbation can effectively alleviate the overfitting issue and improve the performance of hierarchical attention network.
Year
DOI
Venue
2019
10.1016/j.neunet.2019.08.017
Neural Networks
Keywords
Field
DocType
Machine learning,Adversarial training,Text classification,Small-scale datasets,Neural network
Document classification,Word representation,Recurrent neural network,Redundancy (engineering),Artificial intelligence,Overfitting,Sentence,Machine learning,Mathematics,Adversarial system
Journal
Volume
Issue
ISSN
119
1
0893-6080
Citations 
PageRank 
References 
1
0.35
0
Authors
5
Name
Order
Citations
PageRank
Hoon-Keng Poon110.35
Wun-She Yap210517.55
Yee-Kai Tee310.35
Wai-Kong Lee43713.00
Bok-Min Goi549862.02