Evaluating Trust in CNN Transfer Learning with Flower Image Classification via Heatmap-Based XAI

Main Article Content

Rawesak Tanawongsuwan
Sukanya Phongsuphap
Pattanasak Mongkolwat

Abstract

Convolutional neural networks (CNNs) have demonstrated impressive performance in image classification tasks but are often criticized for their black-box nature, which complicates understanding their decision-making and reliability. Transfer learning with pre-trained CNNs is a widely used approach for tasks with limited data. This study evaluates the performance and explainability of popular CNN models on over image classification using two custom datasets, Flower-8-One and Flower-8-Zoom. Employing Explainable AI (XAI) techniques, such as Grad-CAM, this research visualizes CNN decision-making to uncover its alignment with human perception. A human study assesses trustworthiness by analyzing participants' confidence scores based on model visualizations. Results indicate strong CNN performance but highlight disparities between model-extracted features and human expectations. Among the models evaluated, Xception and Inception-v3 consistently earn higher trust ratings. These findings emphasize the necessity of XAI-driven evaluations to enhance trust and reliability in CNN-integrated systems, particularly in applications requiring human-computer interaction.

Article Details

How to Cite
[1]
R. Tanawongsuwan, S. Phongsuphap, and P. Mongkolwat, “Evaluating Trust in CNN Transfer Learning with Flower Image Classification via Heatmap-Based XAI”, ECTI-CIT Transactions, vol. 19, no. 3, pp. 392–405, Jul. 2025.
Section
Research Article

References

A. Krizhevsky, I. Sutskever and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Communications of the ACM, vol. 60, no. 6, pp. 84–90, 2017.

O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg and L. FeiFei, “ImageNet Large Scale Visual Recognition Challenge,” International Journal of Computer Vision, vol. 115, no. 3, pp. 211-252, 2015.

N. K. Sarkar, M. M. Singh, A. Pradesh and U. Nandi, “Recent Researches on Image Classification Using Deep Learning Approach,” International Journal of Computing and Digital Systems, vol. 12, pp. 1357-1374, 2022.

C. R. Prasad, B. Arun, S. Amulya, P. Abboju, S. Kollem and S. Yalabaka, “Breast Cancer Classification using CNN with Transfer Learning Models,” International Conference for Advancement in Technology (ICONAT), pp. 1-5, 2023.

L. Abdelrahman, M. Al Ghamdi, F. Collado-Mesa and M. Abdel-Mottaleb, “Convolutional neural networks for breast cancer detection in mammography: A survey,” Computers in Biology and Medicine, vol. 131, no. 104248, 2021.

J. Gallego, A. Pedraza, S. Lopez, G. Steiner, L. Gonzalez, A. Laurinavicius and G. Bueno, “Glomerulus Classification and Detection Based on Convolutional Neural Networks,” Journal of Imaging, vol. 4, no. 1:20, 2018.

P. Chagas, L. Souza, I. Araújo, N. Aldeman, A. Duarte, M. Angelo, W. L. C. dos-Santos and L. Oliveira, “Classification of glomerular hypercellularity using convolutional features and support vector machine,” Artificial Intelligence in Medicine, vol. 103, no. 101808, 2020.

K. N. Rao, O. I. Khalaf, V. Krishnasree, A. S. Kumar, D. M. Alsekait, S. S. Priyanka, A. S. Alattas and D. S. AbdElminaam, “An efficient brain tumor detection and classification using pre-trained convolutional neural network models,” Heliyon, vol. 10, no. 17, pp. e36773, 2024.

D. Helen, and S. Gokila, “EYENET: An Eye Disease Detection System using Convolutional Neural Network,” International Conference on Edge Computing and Applications (ICECAA), pp. 839-842, 2023.

A. Islam, N. R. Shuvo, M. Shamsojjaman, S. Hasan, S. Hossain and T. Khatun, “An Automated Convolutional Neural Network Based Approach for Paddy Leaf Disease Detection,” International Journal of Advanced Computer Science and Applications, vol. 12, no. 1, 2021.

N. H. T. Nguyen, S. Perry, D. Bone, H. T. Le and T. T. Nguyen, “Two-stage convolutional neural network for road crack detection and segmentation,” Expert Systems with Applications, vol. 186, no. 115718, 2021.

A. Narin, C. Kaya and Z. Pamuk, “Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks,” Pattern Analysis and Applications, vol. 24, no. 3, pp. 1207-1220, 2021.

D. Gunning, “Explainable artificial intelligence (xAI),” Defense Advanced Research Projects Agency (DARPA), 2017.

A. Barredo Arrieta, N. Díaz-Rodríguez, J. Del Ser, A. Bennetot, S. Tabik, A. Barbado, S. Garcia, S. Gil-Lopez, D. Molina, R. Benjamins, R. Chatila and F. Herrera, “Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI,” Information Fusion, vol. 58, pp. 82-115, 2020.

K. Tung, “Flowers Dataset,” Harvard Dataverse, 2020.

“Flowers Recognition,” Source. [Online]. Available: https://www.kaggle.com/datasets/alxmamaev/flowers-recognition.

“Flower Datasets,” Source. [Online]. Available: https://www.robots.ox.ac.uk/~vgg/data/flowers/.

“Flowers dataset,” Source. [Online]. Available: https://www.kaggle.com/dataset/ imsparsh/flowers-dataset.

K. Simonyan and A. Zisserman, “Very Deep Convolutional Networks for Large-Scale Image Recognition,” 3rd International Conference on Learning Representations (ICLR), pp. 1-14, 2015.

C. Szegedy, L. Wei, J. Yangqing, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke and A. Rabinovich, “Going deeper with convolutions,” 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, pp. 1-9, 2015.

B. Zoph, V. Vasudevan, J. Shlens and Q. V. Le, “Learning Transferable Architectures for Scalable Image Recognition,” IEEE Conference on Computer Vision and Pattern Recognition, pp. 8697-8710, 2018.

J. Redmon. "Darknet: Open Source Neural Networks in C," https://pjreddie.com/darknet.

M. Tan and Q. V. Le, “EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks,” Proceedings of the 36th International Conference on Machine Learning, pp. 6105-6114, 2019.

X. Zhang, X. Zhou, M. Lin and J. Sun, “ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices,” IEEE Conference on Computer Vision and Pattern Recognition, pp. 6848-6856, 2018.

N. I. Forrest, W. M. Matthew, A. Khalid, H. Song, J. D. William and K. Kurt, “SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size,” CoRR, 2016.

G. Huang, Z. Liu, L. Van Der Maaten and K. Q. Weinberger, “Densely Connected Convolutional Networks,” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, pp. 2261-2269, 2017.

M. Sandler, A. Howard, M. Zhu, A. Zhmoginov and L. -C. Chen, “MobileNetV2: Inverted Residuals and Linear Bottlenecks,” 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, pp. 4510- 4520, 2018.

C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens and Z. Wojna, “Rethinking the Inception Architecture for Computer Vision,” 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, pp. 2818- 2826, 2016.

F. Chollet, “Xception: Deep Learning with Depthwise Separable Convolutions,” 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, pp. 1800-1807, 2017.

K. He, X. Zhang, S. Ren and J. Sun, “Deep Residual Learning for Image Recognition,” 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, pp. 770-778, 2016.

C. Szegedy, S. Ioffe, V. Vanhoucke and A. A. Alemi, “Inception-v4, inception-ResNet and the impact of residual connections on learning,” Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, pp. 4278–4284, 2017.

S. J. Pan and Q. Yang, “A Survey on Transfer Learning,” in IEEE Transactions on Knowledge and Data Engineering, vol. 22, no. 10, pp. 1345- 1359, Oct. 2010.

L. Torrey and J. Shavlik, “Transfer learning,” Handbook of research on machine learning applications and trends: algorithms, methods, and techniques, pp. 242-264, 2010.

J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng and T. Darrell, “DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition,” 31st International Conference on Machine Learning, pp. 647-655, 2014.

M. Oquab, L. Bottou, I. Laptev and J. Sivic, “Learning and Transferring Mid-level Image Representations Using Convolutional Neural Networks,” 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, pp. 1717-1724, 2014.

A. S. Razavian, H. Azizpour, J. Sullivan and S. Carlsson, “CNN Features Off-the-Shelf: An Astounding Baseline for Recognition,” 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops, Columbus, OH, USA, pp. 512-519, 2014.

P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus and Y. LeCun, “Overfeat: Integrated recognition, localization and detection using convolutional networks,” 2nd International Conference on Learning Representations (ICLR), 2014.

M. D. Zeiler and R. Fergus, “Visualizing and Understanding Convolutional Networks,” European Conference on Computer Vision (ECCV), pp. 818-833, 2014.

D. Erhan, Y. Bengio, A. Courville and P. Vincent, “Visualizing Higher-Layer Features of a Deep Network,” Technical Report, Univeristé de Montréal, 2009.

J. Springenberg, A. Dosovitskiy, T. Brox and M. Riedmiller, “Striving for Simplicity: The All Convolutional Net,” International Conference on Learning Representations (Workshop), 2015.

R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh and D. Batra, “Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization,” 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, pp. 618-626, 2017.

B. Zhou, A. Khosla, A. Lapedriza, A. Oliva and A. Torralba, “Learning Deep Features for Discriminative Localization,” 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, pp. 2921- 2929, 2016.

T. Szandala, “Enhancing Deep Neural Network Saliency Visualizations With Gradual Extrapolation,” IEEE Access, vol. 9, pp. 95155-95161, 2021.

Z. C. Lipton, “The Mythos of Model Interpretability,” ACM Queue, vol. 16, no. 3, pp. 31–57, 2018.

M. T. Ribeiro, S. Singh and C. Guestrin, “Why Should I Trust You?: Explaining the Predictions of Any Classifier,” International Conference on Knowledge Discovery and Data Mining, pp. 1135–1144, 2016.

E. Tjoa, H. J. Khok, T. Chouhan and C. Guan, “Enhancing the confidence of deep learning classifiers via interpretable saliency maps,” Neurocomputing, vol. 562, no. 126825, 2023.

T. Szandala, “Unlocking the black box of CNNs: Visualising the decision-making process with PRISM,” Information Sciences, vol. 642, no. 119162, 2023.