U-Noise: Learnable Noise Masks For Interpretable Image Segmentation
Teddy Koker, Fatemehsadat Mireshghallah, Tom Titcombe, Georgios Kaissis
-
SPS
IEEE Members: $11.00
Non-members: $15.00Length: 00:05:41
Deep Neural Networks (DNNs) are widely used for decision making in a myriad of critical applications, ranging from medical to societal and even judicial. Given the importance of these decisions, it is crucial for us to be able to interpret these models. We introduce a new method for interpreting image segmentation models by learning regions of images in which noise can be applied without hindering downstream model performance. We apply this method to segmentation of the pancreas in CT scans, and qualitatively compare the quality of the method to existing explainability techniques, such as Grad-CAM and occlusion sensitivity. Additionally we show that, unlike other methods, our interpretability model can be quantitatively evaluated based on the downstream performance over obscured images.