Title
A new information measure based on example-dependent misclassification costs and its application in decision tree learning
Abstract
This article describes how the costs of misclassification given with the individual training objects for classification learning can be used in the construction of decision trees for minimal cost instead of minimal error class decisions. This is demonstrated by defining modified, cost-dependent probabilities, a new, cost-dependent information measure, and using a cost-sensitive extension of the CAL5 algorithm for learning decision trees. The cost-dependent information measure ensures the selection of the (local) next best, that is, cost-minimizing, discriminating attribute in the sequential construction of the classification trees. This is shown to be a cost-dependent generalization of the classical information measure introduced by Shannon, which only depends on classical probabilities. It is therefore of general importance and extends classic information theory, knowledge processing, and cognitive science, since subjective evaluations of decision alternatives can be included in entropy and the transferred information. Decision trees can then be viewed as cost-minimizing decoders for class symbols emitted by a source and coded by feature vectors. Experiments with two artificial datasets and one application example show that this approach is more accurate than a method which uses class dependent costs given by experts a priori.
Year
DOI
Venue
2009
10.1155/2009/134807
Adv. Artificial Intellegence
Keywords
Field
DocType
class symbol,class dependent cost,classic information theory,decision tree learning,classical information measure,decision alternative,cost-dependent generalization,decision tree,example-dependent misclassification cost,minimal error class decision,new information measure,cost-dependent probability,cost-dependent information measure
Information Fuzzy Networks,Decision tree,Data mining,Computer science,Artificial intelligence,ID3 algorithm,Decision stump,Information theory,Pattern recognition,Information gain ratio,Decision tree learning,Machine learning,Incremental decision tree
Journal
Volume
Citations 
PageRank 
2009,
2
0.41
References 
Authors
15
2
Name
Order
Citations
PageRank
Fritz Wysotzki145645.46
Peter Geibel228626.62