Abstract | ||
---|---|---|
Classifier learning with imbalanced data is one of the main challenges in the data mining community. An ensemble of classifiers is a popular solution to this problem, and it has acquired significant attention owing to its better performance as compared to individual classifiers. In this paper, we propose an imbalanced classification ensemble method, which is hereafter referred to as overlap and imbalanced sensitive random forest (OIS-RF). We consider the existence of overlap in imbalanced data and create a new coefficient called Hard To Learn (HTL) which aims to measure the degree of importance for each training instance. In this regard, OIS-RF focuses more on learning the instances with high importance in each sub-dataset. Furthermore, to encourage the diversity of the ensemble, a weighted bootstrap method is proposed to generate subdatasets containing diverse local information. The proposed method is evaluated on imbalanced datasets and is supported by statistical analyses. The results show that our method outperforms 9 state-of-the-art ensemble algorithms. |
Year | DOI | Venue |
---|---|---|
2021 | 10.1016/j.engappai.2021.104355 | ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE |
Keywords | DocType | Volume |
Class imbalance, Ensemble, Overlap, Diversity | Journal | 104 |
ISSN | Citations | PageRank |
0952-1976 | 0 | 0.34 |
References | Authors | |
0 | 6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Bo-Wen Yuan | 1 | 4 | 1.73 |
Zhongliang Zhang | 2 | 36 | 2.86 |
Xing-Gang Luo | 3 | 138 | 14.85 |
Yang Yu | 4 | 3 | 1.04 |
Xiao-Hua Zou | 5 | 0 | 0.34 |
Xiao-Dong Zou | 6 | 0 | 0.34 |