Lately, there has been a substantial rise in the number of identified individuals with skin cancer, making it the most widespread form of cancer worldwide. Until now, several machine learning methods that utilize skin scans have been directly employed for skin cancer classification, showing encouraging outcomes in terms of enhancing diagnostic precision. In this paper, multimodal Explainable Artificial Intelligence (XAI) is presented that offers explanations that (1) address a gap regarding interpretation by identifying specific dermoscopic features, thereby enabling (2) dermatologists to comprehend them during melanoma diagnosis and allowing for an (3) evaluation of the interaction between clinicians and XAI. The specific goal of this article is to create an XAI system that closely aligns with the perspective of dermatologists when it comes to diagnosing melanoma. By building upon previous research on explainability in dermatology, this work introduces a novel soft attention mechanism, called Convolutional Spiking Attention Module (CSAM), to deep neural architectures, which focuses on enhancing critical elements and reducing noise-inducing features. Two instances of the proposed CSAM were placed inside the proposed Spiking Attention Block (SAB). The InceptionResNetV2, DenseNet201, and Xception architectures with and without the proposed SAB mechanism were compared for skin lesion classification. Pretrained networks with SAB outperform state-of-the-art methods on the HAM10000 dataset. The proposed method used the ISIC-2019 dataset for the crossdataset validation process. The proposed model provides attention regarding cancer pixels without using an external explainer, which proves the importance and significance of the SAB module.