TY - GEN
T1 - Interpretable Explanations of Black Box Classifiers Applied on Medical Images by Meaningful Perturbations Using Variational Autoencoders
AU - Uzunova, Hristina
AU - Ehrhardt, Jan
AU - Kepp, Timo
AU - Handels, Heinz
PY - 2019/3/15
Y1 - 2019/3/15
N2 - The growing popularity of black box machine learning methods for medical image analysis makes their interpretability to a crucial task. To make a system, e.g. a trained neural network, trustworthy for a clinician, it needs to be able to explain its decisions and predictions. In this work, we tackle the problem of generating plausible explanations for the predictions of medical image classifiers, that are trained to differentiate between different types of pathologies and healthy tissues. An intuitive solution to determine which image regions influence the trained classifier is to find out whether the classification results change when those regions are deleted. This idea can be formulated as a minimization problem and thus efficiently implemented. However, the meaning of “deletion” of image regions, in our case pathologies in medical images, is not defined. We contribute by defining the deletion of pathologies, as the replacement by their healthy looking equivalent generated using variational autoencoders. The experiments with a classification neural network on OCT (Optical Coherence Tomography) images and brain lesion MRIs show that a meaningful replacement of “deleted” image regions has significant impact on the reliability of the generated explanations. The proposed deletion method is proven to be successful since our approach delivers the best results compared to four other established methods.
AB - The growing popularity of black box machine learning methods for medical image analysis makes their interpretability to a crucial task. To make a system, e.g. a trained neural network, trustworthy for a clinician, it needs to be able to explain its decisions and predictions. In this work, we tackle the problem of generating plausible explanations for the predictions of medical image classifiers, that are trained to differentiate between different types of pathologies and healthy tissues. An intuitive solution to determine which image regions influence the trained classifier is to find out whether the classification results change when those regions are deleted. This idea can be formulated as a minimization problem and thus efficiently implemented. However, the meaning of “deletion” of image regions, in our case pathologies in medical images, is not defined. We contribute by defining the deletion of pathologies, as the replacement by their healthy looking equivalent generated using variational autoencoders. The experiments with a classification neural network on OCT (Optical Coherence Tomography) images and brain lesion MRIs show that a meaningful replacement of “deleted” image regions has significant impact on the reliability of the generated explanations. The proposed deletion method is proven to be successful since our approach delivers the best results compared to four other established methods.
UR - https://www.researchgate.net/publication/331764452_Interpretable_explanations_of_black_box_classifiers_applied_on_medical_images_by_meaningful_perturbations_using_variational_autoencoders
U2 - 10.1007/978-3-658-25326-4_42
DO - 10.1007/978-3-658-25326-4_42
M3 - Conference contribution
VL - 10949
T3 - Proceedings of SPIE
SP - 1094911-1 – 1094911-8
BT - Medical Imaging 2019: Image Processing
PB - SPIE
T2 - SPIE MEDICAL IMAGING 2019<br/>
Y2 - 16 February 2019 through 21 February 2019
ER -