Research on human-computer interaction emphasise the importance of reliability in hybrid decision-making systems. Trust hinges on the performance and trustworthiness of AI, achievable through accuracy metrics, confidence scores, eXplainable AI, and abstention mechanisms. This study presents an explainable abstaining classifier named Learning to Reject via Local Rule-based Explanations (L2loRe), a novel approach that leverages the distance between data points and counterfactuals to evaluate the confidence of predictions, thus facilitating the formulation of a rejection policy and generating clear explanations for the reasoning behind predictions or rejections.

L2loRe: a method for explaining the reject option

Clara Punzi;
2024-01-01

Abstract

Research on human-computer interaction emphasise the importance of reliability in hybrid decision-making systems. Trust hinges on the performance and trustworthiness of AI, achievable through accuracy metrics, confidence scores, eXplainable AI, and abstention mechanisms. This study presents an explainable abstaining classifier named Learning to Reject via Local Rule-based Explanations (L2loRe), a novel approach that leverages the distance between data points and counterfactuals to evaluate the confidence of predictions, thus facilitating the formulation of a rejection policy and generating clear explanations for the reasoning behind predictions or rejections.
File in questo prodotto:
File Dimensione Formato  
paper_191.pdf

accesso aperto

Tipologia: Versione finale editoriale
Licenza: Creative commons
Dimensione 584.76 kB
Formato Adobe PDF
584.76 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11568/1325250
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact