An Explainable Error Detection Approach for Machine Learning

  • Kaiyue WU
  • , Changwu HUANG*
  • , Xin YAO*
  • *Corresponding author for this work

Research output: Book Chapters | Papers in Conference ProceedingsConference paper (refereed)Researchpeer-review

Abstract

Detecting the prediction errors of machine learning model is critical for preventing undesirable consequence and enhancing the safety of machine learning systems. However, existing error detection methods often operate as black-box models, making their outputs difficult to interpret and consequently hindering user trust. Additionally, with the enhancement of model accuracy, the occurrences of errors decrease significantly, posing challenges in distinguishing error samples due to the imbalance between correct and erroneous predictions. This paper introduces a novel error detection method called ImExED, which is explainable and accounts for the imbalance between correct and erroneous samples. By integrating Explainable Boosting Machine (EBM) with techniques designed for handling class imbalances, ImExED improves both the interpretability and effectiveness of error detection. Through evaluations on diverse datasets and model families, along with an ablation study on class imbalance techniques, ImExED outperforms current methods in error detection while providing insights into the correctness or incorrectness of model predictions.
Original languageEnglish
Title of host publicationNeural Information Processing - 31st International Conference, ICONIP 2024, Proceedings
EditorsMufti MAHUMD, Maryam DOBORJEH, Kevin WONG, Andrew Chi Sing LEUNG, Zohreh DOBORJEH, M. TANVEER
PublisherSpringer Singapore
Chapter8
Pages104-118
Number of pages15
ISBN (Electronic)9789819665792
ISBN (Print)9789819665785
DOIs
Publication statusPublished - 24 Jun 2025

Publication series

NameLecture Notes in Computer Science
Volume15287 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Bibliographical note

Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.

Funding

This work was supported by the National Natural Science Foundation of China (Grant No. 62250710682), the Guangdong Provincial Key Laboratory (Grant No. 2020B121201001), the Program for Guangdong Introducing Innovative and Entrepreneurial Teams (Grant No.2017ZT07X386), and the Research Institute of Trustworthy Autonomous Systems.

Keywords

  • Confidence Estimation
  • Error Detection
  • Explainable Boosting Machine
  • Explainable Machine Learning

Fingerprint

Dive into the research topics of 'An Explainable Error Detection Approach for Machine Learning'. Together they form a unique fingerprint.

Cite this