TY - GEN
T1 - Reliability-Based Imbalanced Data Classification with Dempster-Shafer Theory
AU - Tian, Hongpeng
AU - Zhang, Zuowei
AU - Martin, Arnaud
AU - Liu, Zhunga
N1 - Publisher Copyright:
© 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.
PY - 2022
Y1 - 2022
N2 - The classification analysis of imbalanced data remains a challenging task since the base classifier usually focuses on the majority class and ignores the minority class. This paper proposes a reliability-based imbalanced data classification approach (RIC) with Dempster-Shafer theory to address this issue. First, based on the minority class, multiple under-sampling for the majority one are implemented to obtain the corresponding balanced training sets, which results in multiple globally optimal trained classifiers. Then, the neighbors are employed to evaluate the local reliability of different classifiers in classifying each test sample, making each global optimal classifier focus on the sample locally. Finally, the revised classification results based on various local reliability are fused by the Dempster-Shafer (DS) fusion rule. Doing so, the test sample can be directly classified if more than one classifier has high local reliability. Otherwise, the neighbors belonging to different classes are employed again as the additional knowledge to revise the fusion result. The effectiveness has been verified on synthetic and several real imbalanced datasets by comparison with other related approaches.
AB - The classification analysis of imbalanced data remains a challenging task since the base classifier usually focuses on the majority class and ignores the minority class. This paper proposes a reliability-based imbalanced data classification approach (RIC) with Dempster-Shafer theory to address this issue. First, based on the minority class, multiple under-sampling for the majority one are implemented to obtain the corresponding balanced training sets, which results in multiple globally optimal trained classifiers. Then, the neighbors are employed to evaluate the local reliability of different classifiers in classifying each test sample, making each global optimal classifier focus on the sample locally. Finally, the revised classification results based on various local reliability are fused by the Dempster-Shafer (DS) fusion rule. Doing so, the test sample can be directly classified if more than one classifier has high local reliability. Otherwise, the neighbors belonging to different classes are employed again as the additional knowledge to revise the fusion result. The effectiveness has been verified on synthetic and several real imbalanced datasets by comparison with other related approaches.
KW - Dempster-Shafer theory
KW - Imbalanced data
KW - Reliability
UR - http://www.scopus.com/inward/record.url?scp=85140466119&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-17801-6_8
DO - 10.1007/978-3-031-17801-6_8
M3 - 会议稿件
AN - SCOPUS:85140466119
SN - 9783031178009
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 77
EP - 86
BT - Belief Functions
A2 - Le Hégarat-Mascle, Sylvie
A2 - Aldea, Emanuel
A2 - Bloch, Isabelle
PB - Springer Science and Business Media Deutschland GmbH
T2 - 7th International Conference on Belief Functions, BELIEF 2022
Y2 - 26 October 2022 through 28 October 2022
ER -