Influence-Balanced XGBoost: Improving XGBoost for Imbalanced Data Using Influence Functions

Akiyoshi Sutou*, Jinfang Wang

*この研究の対応する著者

研究成果: Article査読

抄録

Decision tree boosting algorithms, such as XGBoost, have demonstrated superior predictive performance on tabular data for supervised learning compared to neural networks. However, recent studies on loss functions for imbalanced data have primarily focused on deep learning. The goal of this study is to improve the XGBoost algorithm for better performance on unbalanced data. To this end, Influence-balanced loss (IBL), originally introduced in deep learning, was applied to enhance the performance of the XGBoost algorithm. As a side effect, the proposed method was also found to perform well on datasets prone to over-specialization. Furthermore, we conducted a comparison between the proposed method and conventional techniques using 38 publicly available datasets. Our method outperforms other methods in terms of F1-score and Matthews correlation coefficient.

本文言語English
ページ(範囲)193473-193486
ページ数14
ジャーナルIEEE Access
12
DOI
出版ステータスPublished - 2024

ASJC Scopus subject areas

  • コンピュータサイエンス一般
  • 材料科学一般
  • 工学一般

フィンガープリント

「Influence-Balanced XGBoost: Improving XGBoost for Imbalanced Data Using Influence Functions」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル