IMBoost: A New Weighting Factor for Boosting to Improve the Classification Performance of Imbalanced Data
Imbalanced datasets pose significant challenges in the field of machine learning, as they consist of samples where one class (majority) dominates over the other class (minority). Although AdaBoost is a popular ensemble method known for its good performance in addressing various problems, it fails wh...
Saved in:
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Wiley
2023-01-01
|
Series: | Complexity |
Online Access: | http://dx.doi.org/10.1155/2023/2176891 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | Imbalanced datasets pose significant challenges in the field of machine learning, as they consist of samples where one class (majority) dominates over the other class (minority). Although AdaBoost is a popular ensemble method known for its good performance in addressing various problems, it fails when dealing with imbalanced data sets due to its bias towards the majority class samples. In this study, we propose a novel weighting factor to enhance the performance of AdaBoost (called IMBoost). Our approach involves computing weights for both minority and majority class samples based on the performance of classifier on each class individually. Subsequently, we resample the data sets according to these new weights. To evaluate the effectiveness of our method, we compare it with six well-known ensemble methods on 30 imbalanced data sets and 4 synthetic data sets using ROC, precision-eecall AUC, and G-mean metrics. The results demonstrate the superiority of IMBoost. To further analyze the performance, we employ statistical tests, which confirm the excellence of our method. |
---|---|
ISSN: | 1099-0526 |