From: Boosting methods for multi-class imbalanced data classification: an experimental review

Approach | Brief description | Years |
---|---|---|

AdaBoost.M1 [43] |
A multiclass variation of AdaBoost which uses multiclass base classifier Weight of each base classifiers is a function of error rate | 1997 |

AdaBoost.M2 [43] |
A multiclass variation of AdaBoost Weight of each base classifiers is a function of pseudo-loss | 1997 |

GentleBoost [45] |
Extended version for AdaBoost which uses Newton steps Using weighted least-squares regression for fitting the base classifiers | 2000 |

CSB1 [58] |
A Cost-sensitive variation of AdaBoost proposed for handling imbalanced data Adding cost item into the weight update formula of AdaBoost Removing step size coefficient from the weight update formula of AdaBoost | 2000 |

CSB2 [58] |
A Cost-sensitive variation of AdaBoost proposed for handling imbalanced data Adding cost item into the weight update formula of AdaBoost The step size is considered in the weight update formula, like AdaBoost | 2000 |

MAdaBoost [59] |
Proposed with the goal of solving the AdaBoost's sensitivity to noise Modifying the weight update formula of AdaBoost | 2000 |

RareBoost [60, 61] |
An improvement for AdaBoost Using different weight update scheme for positive and negative predictions Considering False Positive, True Positive, True Negative and False Negative in step size calculation | 2001 |

Modest AdaBoost [62] |
An improvement of GentleBoost Using different weight update formula for misclassified and truly classified samples Using inverted distribution to assign larger weights to truly classified samples | 2005 |

JOUSBoost [63] |
Proposed with the goal of handling imbalanced data in AdaBoost algorithm Combining the jittering of the data and sampling techniques with AdaBoost | 2007 |

ABC-LogitBoost [47] |
An improvement of LogitBoost for multiclass classification Solving the difficulties of dense Hessian Matrix in Logistic loss | 2009 |

AdaBoost.HM [64] |
A multiclass variation of AdaBoost which uses hypothesis margin Using multiclass base classifiers instead of decomposing the multiclass classification problem into multiple binary problems | 2010 |

RAMOBoost [65] |
Proposed with the goal of imbalanced data handling Combining Ranked Minority Oversampling with AdaBoost.M2 Using the sampling probability distribution for ranking the minority class samples | 2010 |

AOSO-LogitBoost [48] |
One versus one version of LogitBoost for multiclass classification Solving the difficulties of dense Hessian Matrix in Logistic loss by utilizing vector tree and adaptive block coordinate descent techniques | 2011 |

CD-MCBoost [66] |
Performing coordinate descent on multiclass loss function Concentration of each base classifier on margin maximization of a single class | 2011 |

EUSBoost [67] |
An improvement of RUSBoost which uses evolutionary undersampling Using different subsets of majority class samples in the training phase of each base classifier to ensure diversity | 2013 |

RB-Boost [68] |
Combining Random Balance with AdaBoost.M2 Using SMOTE sampling to deal with imbalanced data problem The difference with SMOTEBoost is using random proportion of classes in each iteration of booting to ensure the diversity of base classifiers | 2015 |

LIUBoost [69] |
Proposed with the goal of imbalanced data handling Using undersampling in order to solve the imbalanced data problem Adding a cost term to the weight update formula of the samples | 2019 |

TLUSBoost [70] |
Proposed with the goal of imbalanced data handling Using Tomek-linked and redundancy-based undersampling for removing outlier samples | 2019 |