抄録
Decision tree boosting algorithms, such as XGBoost, have demonstrated superior predictive performance on tabular data for supervised learning compared to neural networks. However, recent studies on loss functions for imbalanced data have primarily focused on deep learning. The goal of this study is to improve the XGBoost algorithm for better performance on unbalanced data. To this end, Influence-balanced loss (IBL), originally introduced in deep learning, was applied to enhance the performance of the XGBoost algorithm. As a side effect, the proposed method was also found to perform well on datasets prone to over-specialization. Furthermore, we conducted a comparison between the proposed method and conventional techniques using 38 publicly available datasets. Our method outperforms other methods in terms of F1-score and Matthews correlation coefficient.
| 本文言語 | English |
|---|---|
| ページ(範囲) | 193473-193486 |
| ページ数 | 14 |
| ジャーナル | IEEE Access |
| 巻 | 12 |
| DOI | |
| 出版ステータス | Published - 2024 |
ASJC Scopus subject areas
- コンピュータサイエンス一般
- 材料科学一般
- 工学一般
フィンガープリント
「Influence-Balanced XGBoost: Improving XGBoost for Imbalanced Data Using Influence Functions」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。引用スタイル
- APA
- Standard
- Harvard
- Vancouver
- Author
- BIBTEX
- RIS