Abstract | ||
---|---|---|
For CLEF 2008 JHU conducted monolingual and bilingual experiments in the ad hoc TEL and Persian tasks. Additionally we performed several post hoc experiments using previous CLEF ad hoc tests sets in 13 languages. In all three tasks we explored alternative methods of tokenizing documents including plain words, stemmed words, automatically induced segments, a single selected n-gram from each word, and all n-grams from words (i.e., traditional character n-grams). Character n-grams demonstrated consistent gains over ordinary words in each of these three diverse sets of experiments. Using mean average precision, relative gains of of 50-200% on the TEL task, 5% on the Persian task, and 18% averaged over 13 languages from past CLEF evaluations, were observed. |
Year | DOI | Venue |
---|---|---|
2008 | 10.1007/978-3-642-04447-2_21 | CLEF (Working Notes) |
Keywords | DocType | Volume |
consistent gain,bilingual experiment,alternative method,persian task,past clef evaluation,induced segment,diverse set,traditional character n-grams,tel task,mean average precision,indexation | Conference | 5706 |
ISSN | ISBN | Citations |
0302-9743 | 3-642-04446-8 | 2 |
PageRank | References | Authors |
0.41 | 9 | 1 |
Name | Order | Citations | PageRank |
---|---|---|---|
Paul McNamee | 1 | 38 | 3.50 |