Abstract | ||
---|---|---|
Latent topic models have been successfully applied as an unsupervised topic discov- ery technique in large document collections. With the proliferation of hypertext document collection such as the Internet, there has also been great interest in extending these approaches to hypertext (6, 9). These ap- proaches typically model links in an analo- gous fashion to how they model words - the document-link co-occurrence matrix is mod- eled in the same way that the document-word co-occurrence matrix is modeled in standard topic models. In this paper we present a probabilistic gen- erative model for hypertext document collec- tions that explicitly models the generation of links. Specifically, links from a word w to a document d depend directly on how fre- quent the topic of w is in d, in addition to the in-degree of d. We show how to perform EM learning on this model efficiently. By not modeling links as analogous to words, we end up using far fewer free parameters and obtain better link prediction results. |
Year | Venue | Keywords |
---|---|---|
2008 | Uncertainty in Artificial Intelligence | co occurrence matrix |
DocType | Volume | Citations |
Conference | abs/1206.3254 | 19 |
PageRank | References | Authors |
1.33 | 14 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Amit Gruber | 1 | 234 | 13.91 |
Michal Rosen-Zvi | 2 | 1443 | 131.65 |
Yair Weiss | 3 | 10240 | 834.60 |