Abstract | ||
---|---|---|
Given a topic with title, narrative and description, we start by building a language model for the topic. The top 1000 tweets were retrieved from Twitter commercial search engine by applying the title of the topic as a query. We exploit pseudo relevance feedback technologies to estimate probability distributions of each term in the topic, then comparing these probabilities with a background distribution model. We select the highest dierent terms as our expanded query terms. We then generate a vector for each topic, the features of the vector are non-stop word title terms, selected narrative terms and query expansion terms. Dierent weights are assigned to the dierent types of terms. Since we are allowed to deliver at most 10 tweets every day, and the latency time can not exceed 100 minutes, we solve the tweet notication scenario as a multiple-choice secretary problem. Two dierent solutions were tested. |
Year | Venue | Field |
---|---|---|
2015 | TREC | Data mining,Relevance feedback,Computer science,Probability distribution,Natural language processing,Artificial intelligence,Language model,Social media,Information retrieval,Query expansion,Microblogging,Secretary problem,Exploit,Narrative |
DocType | Citations | PageRank |
Conference | 7 | 0.70 |
References | Authors | |
3 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Luchen Tan | 1 | 55 | 9.04 |
Adam Roegiest | 2 | 58 | 12.18 |
Charles L.A. Clarke | 3 | 3289 | 286.78 |