Volume 4 , Issue 1 , PP: 24-33, 2023 | Cite this article as | XML | Html | PDF | Full Length Article
Wei Hong Lim 1 * , Marwa M. Eid 2
Doi: https://doi.org/10.54216/JAIM.040103
Deep Convolutional Networks (CNNs) have revolutionized various fields, including computer vision, but their decision-making process remains largely opaque. To address this interpretability challenge, numerous visual explanation methods have been proposed. However, a comprehensive evaluation and benchmarking of these methods are essential to understand their strengths, limitations, and comparative performance. In this paper, we present a systematic study that benchmarks and compares various visual explanation techniques for deep CNNs. We propose a standardized evaluation framework consisting of benchmark explain ability methods. Through extensive experiments, we analyze the effectiveness, and interpretability of popular visual explanation methods, including gradient-based methods, activation maximization, and attention mechanisms. Our results reveal nuanced differences between the methods, highlighting their trade-offs and potential applications. We conduce a comprehensive evaluation of visual explanation methods on different deep CNNs, the results demonstrate the ability to achieve informed selection and adoption of appropriate techniques for interpretability in real-world applications.
Convolutional Neural Networks (CNNs) , Benchmarking, Interpretability , Class Activation Maps (CAM) , Deep learning, Image classification , Explainable AI.
[1] Ramaswamy, Harish Guruprasad, Ablation-cam: Visual explanations for deep convolutional network via gradient-free localization. Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2020.
[2] Fei Z. et al., Deep convolution networkbased emotion analysis towards mental health care. Neurocomputing, 388, 212-227, 2020.
[3] Hägele M et al., Resolving challenges in deep learning-based analyses of histopathological images using explanation methods. Scientific reports, 10(1), 1-12, 2020.
[4] Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., & Batra, D. (2017). Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE international conference on computer vision, 618-626, 2017.
[5] Nayak R., Pati U. C., Das S. K., A comprehensive review on deep learning-based methods for video anomaly detection. Image and Vision Computing, 106, 104078, 2021.
[6] Mohamed Saber, Efficient phase recovery system, IJEECS, 5(1), 2017.
[7] Zhong B., Pan X., Love P. E., Ding L., Fang W. , Deep learning and network analysis: Classifying and visualizing accident narratives in construction. Automation in Construction, 113, 103089, 2020
[8] Lee J. H., Han S. S., Kim Y. H., Lee, C., Kim I, Application of a fully deep convolutional neural network to the automation of tooth segmentation on panoramic radiographs. Oral surgery, oral medicine, oral pathology and oral radiology, 129(6), 635-642, 2020.
[9] Mohamed Saber, A novel design and implementation of FBMC transceiver for low power applications. IJEEI, 8(1), 83-93, 2020.
[10] Kitaguchi D., Takeshita N., Matsuzaki H., Takano H., Owada Y., Enomoto T., Ito M., Real-time automatic surgical phase recognition in laparoscopic sigmoidectomy using the convolutional neural network-based deep learning approach. Surgical endoscopy, 34, 4924-4931, 2020.
[11] Van der Velden, B. H. Kuijf, H. J. Gilhuijs, K. G., Viergever, M. A., Explainable artificial intelligence (XAI) in deep learning-based medical image analysis. Medical Image Analysis, 102470, 2022.
[12] Roy S., Menapace W., Oei S., Luijten B., Fini E., Saltori C., Demi L., Deep learning for classification and localization of COVID-19 markers in point-of-care lung ultrasound. IEEE transactions on medical imaging, 39(8), 2676-2687, 2020.
[13] Wang H., Wang Z., Du M., Yang F., Zhang Z., Ding S.,Hu X., Score-CAM: Score-weighted visual explanations for convolutional neural networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition workshops, 24-25, 2020.
[14] Naeem H., Ullah F., Naeem M. R., Khalid S., Vasan D., Jabbar S., Saeed S., Malware detection in industrial internet of things based on hybrid image visualization and deep learning model. Ad Hoc Networks, 105, 102154, 2020.
[15] Arshad H., Khan M. A., Sharif M. I., Yasmin M., Tavares J. M. R., Zhang Y. D., Satapathy S. C., A multilevel paradigm for deep convolutional neural network features selection with an application to human gait recognition. Expert Systems, 39(7), e12541, 2022.
[16] Abouelatta, Mohamed A., Sayed A. Ward, Ahmad M. Sayed, Karar Mahmoud, Matti Lehtonen, and Mohamed MF Darwish, Measurement and assessment of corona current density for HVDC bundle conductors by FDM integrated with full multigrid technique. Electric Power Systems Research, 199, 107370, 2021.