Abstract | ||
---|---|---|
Discretization of continuous attributes is an important task for certain types of machine learning algorithms. Bayesian approaches, for instance, require assumptions about data distributions. Decision Trees, on the other hand, require sorting operations to deal with continuous attributes, which largely increase learning times. This paper presents a new method of discretization, whose main characteristic is that it takes into account interdependencies between attributes. Detecting interdependencies can be seen as discovering redundant attributes. This means that our method performs attribute selection as a side effect of the discretization. Empirical evaluation on five benchmark datasets from UCI repository, using C4.5 and a naive Bayes, shows a consistent reduction of the features without loss of generalization accuracy. |
Year | DOI | Venue |
---|---|---|
1998 | 10.1007/3-540-49795-1_14 | IBERAMIA |
Keywords | Field | DocType |
attribute selection,detecting interdependency,bayesian approach,benchmark datasets,decision trees,continuous attribute,uci repository,continuous attributes,dynamic discretization,new method,certain type,account interdependency,feature selection | Decision tree,Discretization,Feature selection,Algorithm complexity,Naive Bayes classifier,Computer science,Sorting,Artificial intelligence,Machine learning,Discretization of continuous features,Bayesian probability | Conference |
Volume | ISSN | ISBN |
1484 | 0302-9743 | 3-540-64992-1 |
Citations | PageRank | References |
16 | 0.91 | 10 |
Authors | ||
3 |
Name | Order | Citations | PageRank |
---|---|---|---|
João Gama | 1 | 3785 | 271.37 |
Luís Torgo | 2 | 690 | 85.83 |
Carlos Soares | 3 | 95 | 18.18 |