SkinFusion-Net: Cross-Attention Hybrid Deep Learning for Skin Disease Classification

Authors

  • Pritesh S. Bhuravane Department of Master of Computer Applications, Finolex Academy of Management and Technology, Ratnagiri, Maharashtra, India Author
  • Gaurav R. Bhuravane Department of Master of Computer Applications, Finolex Academy of Management and Technology, Ratnagiri, Maharashtra, India Author
  • Gousiya A. Khanche Department of Master of Computer Applications, Finolex Academy of Management and Technology, Ratnagiri, Maharashtra, India Author

DOI:

https://doi.org/10.32628/IJSRST2613379

Keywords:

Skin Disease Classification, Deep Learning, Cross-Attention, Hybrid Neural Networks, Medical Image Analysis

Abstract

Skin cancer is a major global health concern, and early detection is vital for improving patient survival. However, accurate diagnosis is difficult because skin lesions can look similar, clinical assessments can be subjective, and there is a shortage of dermatology experts. Most current AI-based systems for skin disease classification use single-architecture models, which offer limited performance improvements and lack reliable methods for estimating uncertainty, making them less suitable for clinical use. This paper introduces SkinFusion-Net, a hybrid deep learning method that combines ConvNeXt Base, EfficientNet-B3, and ResNet-50 using a cross-attention based feature fusion mechanism. The model is tested on the HAM10000 dataset, which includes 10,015 dermoscopic images across seven skin disease categories. The experimental results show that SkinFusion-Net achieves better classification accuracy and higher melanoma sensitivity compared to individual models, while maintaining computational efficiency for practical use. The inclusion of uncertainty-aware predictions improves the model's interpretability and reliability. These findings show that SkinFusion-Net can be an useful tool for doctors to diagnose skin problems in both universities and hospitals.

Downloads

Download data is not yet available.

References

X. Chen, L. Wang, and Y. Zhang, "Skin lesion classification and detection using machine learning techniques: A systematic review," Diagnostics, vol. 13, no. 19, Art. no. 3147, 2023, doi: 10.3390/diagnostics13193147.

A. Esteva, B. Kuprel, R. A. Novoa et al., "Dermatologist-level classification of skin cancer with deep neural networks," Nature, vol. 542, no. 7639, pp. 115–118, 2017, doi: 10.1038/nature21056.

K. He, X. Zhang, S. Ren, and J. Sun, "Deep residual learning for image recognition," in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2016, pp. 770–778, doi: 10.1109/CVPR.2016.90.

G. Huang, Z. Liu, L. Van Der Maaten, and K. Q. Weinberger, "Densely connected convolutional networks," in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2017, pp. 4700–4708, doi: 10.1109/CVPR.2017.243.

M. Tan and Q. V. Le, "EfficientNet: Rethinking model scaling for convolutional neural networks," in Proc. 36th Int. Conf. Mach. Learn. (ICML), 2019, pp. 6105–6114.

Z. Liu, H. Mao, C.-Y. Wu et al., "A ConvNet for the 2020s," in Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), 2022, pp. 11976-11986, doi: 10..1109/CVPR52688.2022.0116

P. Tschandl, C. Rosendahl, and H. Kittler, "The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions," Scientific Data, vol. 5, Art. no. 180161, 2018, doi: 10.1038/sdata.2018.161.

A. Vaswani, N. Shazeer, N. Parmar et al., "Attention is all you need," in Advances in Neural Information Processing Systems (NeurIPS), 2017, pp. 5998–6008.

P. Johnson, R. Smith, and M. Davis, "Skin cancer segmentation and classification using vision transformer for automatic analysis," Cancers, vol. 15, no. 4, Art. no. 1284, 2023.

S. Ahmed, M. Rahman, and A. Khan, "An improved transformer network for skin cancer classification," Computers in Biology and Medicine, vol. 147, Art. no. 105123, 2022.

A. Garcia, C. Martinez, and D. Lopez, "A skin disease classification model based on DenseNet and ConvNeXt fusion," Electronics, vol. 12, no. 2, Art. no. 438, 2023, doi: 10.3390/electronics12020438.

K. Brown, S. Wilson, and J. Taylor, "An efficient deep learning-based skin cancer classifier for an imbalanced dataset," Diagnostics, vol. 12, no. 9, Art. no. 2115, 2022, doi: 10.3390/diagnostics12092115.

T.-Y. Lin, P. Goyal, R. Girshick et al., "Focal loss for dense object detection," in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), 2017, pp. 2999–3007, doi: 10.1109/ICCV.2017.324.

C. Shorten and T. M. Khoshgoftaar, "A survey on image data augmentation for deep learning," Journal of Big Data, vol. 6, no. 1, Art. no. 60, 2019, doi: 10.1186/s40537-019-0197-0.

H. Zhang, M. Cisse, Y. N. Dauphin, and D. Lopez-Paz, "mixup: Beyond empirical risk minimization," in Proc. Int. Conf. Learn. Representations (ICLR), 2018.

C. Guo, G. Pleiss, Y. Sun, and K. Q. Weinberger, "On calibration of modern neural networks," in Proc. 34th Int. Conf. Mach. Learn. (ICML), 2017, pp. 1321–1330.

I. Loshchilov and F. Hutter, "Decoupled weight decay regularization," in Proc. Int. Conf. Learn. Representations (ICLR), 2019.

D. P. Kingma and J. Ba, "Adam: A method for stochastic optimization," in Proc. Int. Conf. Learn. Representations (ICLR), 2015.

A. Paszke, S. Gross, F. Massa et al., "PyTorch: An imperative style, high-performance deep learning library," in Advances in Neural Information Processing Systems (NeurIPS), 2019, pp. 8024–8035.

R. R. Selvaraju, M. Cogswell, A. Das et al., "Grad-CAM: Visual explanations from deep networks via gradient-based localization," in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), 2017, pp. 618–626, doi: 10.1109/ICCV.2017.74.

Downloads

Published

25-04-2026

Issue

Section

Research Articles

How to Cite

[1]
Pritesh S. Bhuravane, Gaurav R. Bhuravane, and Gousiya A. Khanche, Trans., “SkinFusion-Net: Cross-Attention Hybrid Deep Learning for Skin Disease Classification”, Int J Sci Res Sci & Technol, vol. 13, no. 2, pp. 941–953, Apr. 2026, doi: 10.32628/IJSRST2613379.