Title
A Comparative Study of Deep Neural Network Models on Multi-Label Text Classification in Finance
Abstract
Multi-Label Text Classification (MLTC) is a well-known NLP task that allows the classification of texts into multiple categories indicating their most relevant domains. However, training model tasks on texts from web user deal with redundancy or ambiguity of linguistic information. In this work, we propose a comparative study about different neural network models for a multi-label text categorisation task in finance domain. Our main contribution consists of presenting a new annotated dataset that contains <sup xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">~</sup> 26k posts from users associated to finance categories. To build that dataset, we defined 10 specific-domain categories that cover financial texts. To serve as a baseline, we present a comparative study analysing both the performance and training time of different learning models for the task of multilabel text categorisation on the new dataset. The results show that transformer-based language models outperformed RNN-based neural networks in all scenarios in terms of precision. However, transformers took much more time than RNN models to train an epoch model.
Year
DOI
Venue
2021
10.1109/ICSC50631.2021.00039
2021 IEEE 15th International Conference on Semantic Computing (ICSC)
Keywords
DocType
ISSN
Natural Language Processing,Neural Networks,Text Classification,Finances
Conference
2325-6516
ISBN
Citations 
PageRank 
978-1-7281-8900-0
0
0.34
References 
Authors
19
5
Name
Order
Citations
PageRank
Macedo Maia163.15
Juliano Efson Sales2254.72
André Freitas318031.90
Siegfried Handschuh41988181.71
Markus Endres512.38