Unboxing AI | Live Webinar Every Friday | Sham Sokka, PhD, Deephealth (Dec 20) @ 11 am ET | Latha Poonamallee, Ph.D., InMed.ai (Dec 27) @11 am ET | Register Now
  • 2022-04-01

Clinical Explainability Failure (CEF) & Explainability Failure Ratio (EFR)–changing the way we validate classification algorithms

Adoption of Artificial Intelligence (AI) algorithms into the clinical realm will depend on their inherent trustworthiness, which is built not only by robust validation studies but is also deeply linked to the explainability and interpretability of the algorithms. Most validation studies for medical imaging AI report the performance of algorithms on study-level labels and lay little emphasis on measuring the accuracy of explanations generated by these algorithms in the form of heat maps or bounding boxes, especially in true positive cases. We propose a new metric – Explainability Failure Ratio (EFR) – derived from Clinical Explainability Failure (CEF) to address this gap in AI evaluation. We define an Explainability Failure as a case where the classification generated by an AI algorithm matches with study-level ground truth but the explanation output generated by the algorithm is inadequate to explain the algorithm's output. We measured EFR for two algorithms that automatically detect consolidation on chest X-rays to determine the applicability of the metric and observed a lower EFR for the model that had lower sensitivity for identifying consolidation on chest X-rays, implying that the trustworthiness of a model should be determined not only by routine statistical metrics but also by novel ‘clinically-oriented’ models.

Link to complete publication here: https://scholar.google.com/citations?view_op=view_citation&hl=en&user=dqpMNRUAAAAJ&citation_for_view=dqpMNRUAAAAJ:QIV2ME_5wuYC

Unlock the potential of CARPL platform for optimizing radiology workflows

Talk to a Clinical Solutions Architect