“…Pinpointing image sub-regions that were used by the model to make its global imageclass prediction not only provides weakly supervised segmentation, but also enables interpretable deep-network classifiers. It is worth noting that such interpretability aspects are also attracting wide interest in computer vision (Bach et al, 2015;Bau et al, 2017;Bhatt et al, 2020;Dabkowski and Gal, 2017;Escalante et al, 2018;Fong et al, 2019;Fong and Vedaldi, 2017;Goh et al, 2020;Osman et al, 2020;Murdoch et al, 2019;Petsiuk et al, 2020;2018;Ribeiro et al, 2016;Samek et al, 2020;Zhang et al, 2020;Belharbi et al, 2021) and medical imaging (de La Torre et al, 2020;Gondal et al, 2017;González-Gonzalo et al, 2020;Taly et al, 2019;Quellec et al, 2017;Keel et al, 2019;Wang et al, 2017). Deep learning classifiers are often considered as "black boxes" due to the lack of explanatory factors in their decisions.…”