Title
HiCAN: Hierarchical Convolutional Attention Network for Sequence Modeling
Abstract
Convolutional neural networks (CNN) are widely used on sequential data since it can capture local context dependencies and temporal order information inside sequences. Attention (ATT) mechanisms have also attracted enormous interests due to its capability of capturing the important parts of a sequence. These two neural networks can extract different features from sequences. In order to combine the advantages of CNN and ATT, we propose a convolutional attention network (CAN), which merges the structure of CNN and ATT into a single neural network and can serve as a new basic module in complex neural networks. Based on CAN, we then build a sequence encoding model with hierarchical structure, "hierarchical convolutional attention network (HiCAN)", to tackle sequence modeling problems. It can explicitly capture both the local and global context dependencies and temporal order information in sequences. Extensive experiments conducted on session-based recommendation (Recommender Systems) demonstrate that HiCAN is able to outperform state-of-the-art methods and show higher computational efficiency. Furthermore, we conduct extended experiments on text classification (Natural Language Processing). The results show that our model can also achieve competitive performance on NLP tasks.
Year
DOI
Venue
2019
10.1145/3357384.3357996
Proceedings of the 28th ACM International Conference on Information and Knowledge Management
Keywords
Field
DocType
attention mechanisms, sequential behavior, session-based recommendation
Data mining,Computer science,Sequence modeling
Conference
ISBN
Citations 
PageRank 
978-1-4503-6976-3
0
0.34
References 
Authors
0
4
Name
Order
Citations
PageRank
Yi Cao101.35
Weifeng Zhang242.15
Bo Song331.40
Congfu Xu413214.31