Abstract | ||
---|---|---|
Nowadays people are overwhelmed by more and more short information from lots of different applications, especially with the rapid development of mobile systems. One way to alleviate this issue is an automatic classification of the short texts before they are delivered to users. Several methods have been proposed to classify the short texts, and they are largely based on expanding the short texts to longer ones with external resources to solve the sparseness problem. Different from these studies, we tackle the sparseness problem by selecting a better feature space in which the feature vectors of the short texts are denser, and our method needs no external resources at all. The experimental results on an open dataset show that this method can significantly improve the short text classification accuracy comparing with the baseline, especially when the dimension of the feature space is low. |
Year | DOI | Venue |
---|---|---|
2013 | 10.1109/CIS.2013.32 | CIS |
Keywords | Field | DocType |
better feature space,short text,short information,different application,external resource,short text classification accuracy,better feature space selection,improving short text classification,feature vector,feature space,sparseness problem,automatic classification,vectors,text analysis,classification | k-nearest neighbors algorithm,Data mining,Feature vector,Computer science,Feature (computer vision),Artificial intelligence,Machine learning | Conference |
Citations | PageRank | References |
4 | 0.39 | 10 |
Authors | ||
3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Meng Wang | 1 | 4 | 1.41 |
Lanfen Lin | 2 | 78 | 24.70 |
Feng Wang | 3 | 20 | 2.34 |