Abstract | ||
---|---|---|
In this paper, we present the Polylingual Labeled Topic Model, a model which combines the characteristics of the existing Polylingual Topic Model and Labeled LDA. The model accounts for multiple languages with separate topic distributions for each language while restricting the permitted topics of a document to a set of predefined labels. We explore the properties of the model in a two-language setting on a dataset from the social science domain. Our experiments show that our model outperforms LDA and Labeled LDA in terms of their held-out perplexity and that it produces semantically coherent topics which are well interpretable by human subjects. |
Year | DOI | Venue |
---|---|---|
2015 | 10.1007/978-3-319-24489-1_26 | Lecture Notes in Artificial Intelligence |
Keywords | Field | DocType |
Thesauri,Classification,Probabilistic linking,Topic models | Perplexity,Information retrieval,Computer science,Artificial intelligence,Natural language processing,Topic model,Machine learning | Journal |
Volume | ISSN | Citations |
9324 | 0302-9743 | 4 |
PageRank | References | Authors |
0.53 | 6 | 4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Lisa Posch | 1 | 19 | 5.44 |
Arnim Bleier | 2 | 62 | 9.60 |
Philipp Schaer | 3 | 110 | 19.30 |
Markus Strohmaier | 4 | 1210 | 102.65 |