Abstract | ||
---|---|---|
Feature selection has been studied and shown to improve classifier performance in standard batch data mining but is mostly unexplored in data stream mining. Feature selection becomes even more important when the relevant subset of features changes over time, as the underlying concept of a data stream drifts. This specific kind of drift is known as feature drift and requires specific techniques not only to determine which features are the most important but also to take advantage of them. This paper presents a novel method of feature subset selection specialized for dealing with the occurrence of feature drifts called Iterative Subset Selection (ISS), which splits the feature selection process into two stages by first ranking the features, and then iteratively selecting features from the ranking. Applying our feature selection method together with Naive Bayes or k-Nearest Neighbour as a classifier, results in compelling accuracy improvements, compared to prior work.
|
Year | DOI | Venue |
---|---|---|
2018 | 10.1145/3167132.3167188 | SAC 2018: Symposium on Applied Computing
Pau
France
April, 2018 |
Keywords | Field | DocType |
Data Stream Mining, Feature Selection, Concept Drift, Embedded Feature Selection, Iterative Subset Selection | Data stream mining,Naive Bayes classifier,Ranking,Pattern recognition,Feature selection,Data stream,Computer science,Concept drift,Artificial intelligence,Classifier (linguistics) | Conference |
ISBN | Citations | PageRank |
978-1-4503-5191-1 | 2 | 0.39 |
References | Authors | |
15 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Lanqin Yuan | 1 | 2 | 0.39 |
Bernhard Pfahringer | 2 | 10252 | 494.74 |
Jean Paul Barddal | 3 | 140 | 16.77 |