1. Introduction
The epidermis, the outermost layer of skin, is where malignant cells grow and multiply uncontrollably and abnormally to create skin cancer. The leading cause of skin cancer is prolonged direct exposure to ultraviolet sun rays, which causes melanin, a pigment, to be produced in the top layer of the skin [
1]. Moreover, a fair complexion, sunburns, a family history of the disease, and a weakened immune system are risk factors that might lead to the development of skin cancer [
2,
3]. Skin cancer can take on various forms, including squamous cell carcinoma, basal cell carcinoma, and melanoma [
4], with melanoma being the most severe type in comparison. Melanoma is a less frequent but more dangerous type of skin cancer and can invade surrounding tissue and cause disfigurement or even death if left untreated at an early stage.
The most prevalent form of cancer in the world is skin cancer. According to the World Health Organization (WHO), in 2020, a total of 1.5 million cases of skin malignancies were detected worldwide, with an accompanying report of over 120,000 fatalities attributed to skin cancer [
5]. It is becoming more common in many parts of the world and is now one of the top ten cancers worldwide. In South Africa, skin cancer is a significant public health concern and has one of the highest incidence rates [
6]. According to the South African Skin Cancer Foundation, skin cancer affects up to 80% of newly diagnosed cancer cases in South Africa. It affects one in three people throughout their lifetimes [
6]. In South Africa, the prevalence of skin cancer is high for several reasons, including the country's location in the southern hemisphere, where there is higher UV exposure and the large population of fair-skinned individuals of European descent [
7]. Other risk factors for skin cancer in South Africa include exposure to sunlight, outdoor occupations, and a lack of sun protection, such as using hats and sunscreen. To address the growing burden of skin cancer in South Africa, public health officials and health organizations have launched several initiatives to increase awareness of the disease and promote sun safety measures [
8].
Clinical diagnosis accuracy for skin lesions in typical clinical settings depends on clinician experience and training. Dermatologists and other healthcare professionals have accuracy rates of 60% to 90% for skin cancer identification, with more excellent rates for more experienced practitioners [
9,
10,
11]. Even skilled clinicians can misdiagnose or postpone diagnosis, resulting in poor patient outcomes. Lesions that look like benign skin lesions might make melanoma challenging to diagnose. Asymmetry, Border, Color, Diameter, and Evolution (ABCDE) criteria, biopsy, and histological investigation are used by most dermatologists [
12,
13,
14,
15,
16,
17]. Manual visualization and segmentation for pattern analysis make these methods time-consuming, expensive, and inaccurate [
18]. Photo or visual examination cannot distinguish malignant from benign lesions. Skin biopsy is limited by its invasiveness, pain, and requirement for additional samples in suspected lesions by various procedures. Non-invasive instruments aid clinical diagnosis [
19,
20]. Dermoscopy non-invasive procedures have collected crucial or irregular skin lesion features, removed reflection, and improved visual impression. Automatically detecting skin lesions may be complex due to artefacts, low contrast, skin tone, hairs, veins, and other visual characteristics like melanoma and non-melanoma [
21,
22]. Thus, computer-assisted methods that consider pigment networks, streaks, spots, globules, and different skin patterns are needed to help doctors diagnose accurately and quickly [
23].
1.1. Motivation & Objectives
In recent years, deep learning-based systems have achieved tremendous popularity in medical imaging and classification. Computer-assisted diagnostics improve skin cancer diagnosis by objectively and quantitatively studying skin anomalies [
24]. It can help physicians make better decisions, eliminate misdiagnosis and delay, enhance patient outcomes, increase efficiency, and lower costs. Deep learning algorithms have been proven to identify skin cancer with 90% accuracy, equivalent to or better than human doctors [
25]. For years, convolutional neural networks (CNNs) have dominated medical image classification and diagnoses. Their capacity to extract and analyze complex image patterns made them ideal for disease detection, anomaly identification, and tissue classification.
However, CNNs have some limitations, such as CNNs are not able to represent spatial relationships between the features, sensitive to noises [
26,
27] and limited at generalizing to new data due to down-sampling layers of CNN pooling layers leading to data loss [
28,
29,
30]. In addition, spatial information, and instantiation parameters (such as the position of low-level features to each other, deformation, and texture information) are not transferable in convolutional neural networks [
30]. Thus, the above restrictions result in five major problems:
Skin lesions are incredibly challenging to classify appropriately because of their similarities in size, color, and overall appearance. To address the first problem, the authors used data augmentation and normalization techniques to capture low-contrast skin lesions by eliminating air bubbles, noise, and artefacts [
30]. The active contour snake model has been proposed to effectively segment the region of interest and picture borders [
31]. The pre-trained model ResNet50 has also been employed to extract the most relevant features from the image and address the overfitting problem [
32]. These approaches have been proposed as potential solutions to address the second problem. Lastly, to address the third problem, we proposed a dynamic routing model known as Capsule neural networks (CapsNets) by fusing channel and spatial attention mechanisms [
33] to highlight the informative regions and improve the accuracy, generalization ability and interpretability of the model for skin lesion recognition and detection. The routing mechanism of the network suppresses the noise and focuses on the most relevant features of the image. Capsule networks can comprehensively record image features, positions, channels, and spatial relationships through neuron "packaging"[
34]. Therefore, capsules can identify specific patterns and mitigate the network's reliance on extensive datasets [
34], effectively improving the model's capacity to address a broader spectrum of pathological assessment demands [
35].
The main contribution to this research is as follows:
Develop and implement an Active contour segmentation technique for accurately localizing skin lesions within images and applying the ResNet50 pre-trained network to extract essential and relevant features of interest from images.
Propose a novel approach by integrating a Capsule Network architecture fused with the Convolutional Block Attention Module (CBAM), which includes dynamic routing and layer-based squashing for feature extraction and classification of segmented skin lesions. Stochastic Gradient Descent (SGD), a gradient-based optimization technique, is used to optimize the model parameters.
Evaluate the novel approach on a diverse dataset of skin lesion images and compare its performance against traditional methods and state-of-the-art techniques.
The rest of the article is organized as follows:
Section 2 discusses the related work;
Section 3 describes the proposed research technique, including the protocol, algorithm, mathematical representations, and pseudocode.; the proposed method is compared with existing approaches to offer simulated results in
Section 4; and finally, the conclusion and future work is presented in
Section 5.
2. Related Work
With the rising prevalence of skin malignancies, a growing population, and a lack of competent clinical experience and resources, there is a high demand for AI image diagnosis to assist physicians in medicine. Extensive research has been conducted on automated skin cancer diagnosis [
36]. Most skin lesion diagnostic studies followed the standard machine learning method, including preprocessing, segmentation, feature extraction and selection, and classification [
37]. Researchers have developed computer-aided diagnosis approaches based on deep learning techniques that differentiate between malignant and benign skin lesions using several image modalities, including histopathology, confocal, clinical follow-up, dermoscopy, and expert consensus [
36,
37,
38]. Deep learning algorithms have demonstrated notable achievements in the field of medical imaging, particularly in the realm of skin cancer detection [
39]. Traditional automated skin cancer diagnostic methodologies typically involve two primary components: developing handmade features and utilizing machine learning classifiers for classification [
40]. A computer-aided design (CAD) system encompasses several crucial stages, including preprocessing initial dermoscopy pictures, lesion detection by segmentation approaches, extraction of handcrafted features, selection of features, and classification using machine learning classifiers [
40].
Due to its excellent feature extraction, researchers use a Convolutional Neural Network (CNN) for skin cancer detection [
40]. However, convolutional neural networks (CNNs) require a lot of training data to recognize images with rotational invariance or other transformations accurately and record spatial relationships between features [
41]. Reinforcement learning and pre-trained models were used to solve CNN restrictions [
42,
43]. The approaches failed to improve, leading to capsule networks (Caps Nets) [
44]. This technique improved model accuracy to a greater extent than CNN [
45]. The authors of [
46] use "faster region-oriented convolutional neural networks (RCNN) and fuzzy k-means cluster (FKM)" to detect cutaneous melanoma. After refining dataset photos to improve visual information and remove noise and illumination, the faster RCNN constructs a feature vector of a predefined length. FKM breaks the image into different-sized and-boundary fragments. FKM cannot always accurately define skin lesion image borders. The Faster R-CNN model may overfit if the training dataset is too small or the model is not correctly regularized. The proposed skin cancer detection method separates benign, malignant, and typical carcinoma [
47]. After feature extraction, segmentation, and classification, fuzzy C-means clustering image segments is advised. LVP and LBP extract features from segmented images. The fuzzy classifier identifies images using LVP+LBP recovered features. The enhanced "Rider Optimization Algorithm (ROA)" "Distance Oriented ROA" is used best to find fuzzy classifier membership function boundaries in this work. FCM performance may decrease with complex or textured graphics. It may not converge or be locked in local minima. It can be challenging to predict the number of clusters in advance.
To classify dermoscopy images with benign or malignant lesions, [
48] offered two new hybrid CNN representations through an SVM categorizer at the output layer. The SVM classifier classifies the first and second CNN representations' concatenated characteristics. The framework's performance is compared to dermatology labelling. This model outperformed the latest CNN representations on the public ISBI 2016 dataset. A DNN with optimized training and dermoscopy image learning may identify skin cancer [
49]. Combining many dermoscopy datasets provides a solid foundation for DL representation training. The recommended framework trains faster on a small dataset utilizing transfer learning and fine-tuning. Data augmentation improves method performance. A total of 58,032 fine-tuned dermoscopy images were used in this education. The highlighted metrics suggest that the DNN network using customized EfficientNetV2-M outperforms recent deep learning-based multiclass classification representations. The deep neural network (DNN) architecture classifies lesions as benign or malignant. Labelled skin lesions in the data set are categorized using these binary classes. Due to several circumstances, including imaging equipment, illumination, and patient movements, medical images, especially dermoscopy images of skin lesions, can be influenced by noise. Identifying and assessing the lesion adequately might be challenging because noise can hide crucial features and produce artefacts.
In recent years, a deep neural network system used transfer learning to extract features from dermoscopy images and a classifier layer to predict class labels. This study [
50] recommends DL for exact lesion extraction. Image quality is improved by "Enhanced Super-Resolution Generative Adversarial Networks (ESRGAN)". ROI is identified from the complete image after segmentation. For image evaluation, CNNs and modified Resnet-50 models classify skin lesions. Seven skin cancer types from the HAM10000 dataset were randomly selected for this study. The recommended CNN-based technique outperformed the preceding analysis with 0.86 accuracy, 0.84 precision, 0.86 recall, and 0.86 F-score. Image processing and ML synthetically diagnose skin cancer [
51]. Graphic low-resolution images are employed to recreate high-resolution images or sequences. CNN representation precision improved with deep learning image super-resolution. CNN's decision-making and learned qualities are challenging to understand. Many retrieved features make the final prediction harder. CNN training requires a lot of labelled data, especially for high accuracy and generalization, which is time-consuming.
The ISR package and DLN, like ResNet, VGG16, and InceptionV3, can improve low-quality images for computer-assisted skin cancer detection. Skin cancer features like border, color, symmetry, diameter, texture, size, and form can be analyzed using neural networks. These features are used to classify healthy and cancerous skin using image-based data. The authors of [
52] propose "Teaching-Learning-Based Optimization (TLBO)" and the upgraded Extreme Learning Machine (ELM) algorithm for flexible melanoma diagnosis. ELM is a quick, accurate feed-forward neural network with one hidden layer, and TLBO optimizes system settings for optimal visual output. Combining these methods may improve melanoma detection by classifying skin lesions as benign or malignant [
52].
The study [
53] proposed an intelligible CNN-based stacked ensemble framework for initial melanoma skin cancer detection. The stacking collaborative structure employs the transfer learning concept to combine many CNN sub-methods that achieve the same categorization task. The last forecasts are generated by a novel kind called a meta-learner, which uses all of the sub-model predictions. The representation is assessed using benign and malignant melanoma images from an open-access dataset. Using an explainable technique, proficient adaptive clarifications generate heat maps that emphasize the areas within melanoma images exhibiting the highest degree of infection manifestation. Dermatologists can, therefore, understandably interpret the model's decision.
A new deep transfer learning standard for MobileNetV2 melanoma classification is proposed [
54]. MobileNetV2, a deep CNN, diagnoses skin lesions as benign or malignant. ISIC 2020 assesses the presentation of deep learning standards. Class imbalance arises when 2% or less of dataset samples are certain. Augmenting data with random elements reduces session inequality. Studies in [
55] show that deep learning outperforms cutting-edge DL algorithms' accuracy and computing power. The proposed system [
55] combines robotic "DL with a Class Attention Layer Oriented Skin Lesion Detection and Classification (DLCAL_SLDC)" to identify and classify skin lesions. The DLCAL-SLDC technique classifies skin tumors using dermoscopy. A dull razor removes hair; a typical average filter removes noise in image preparation procedures. Dermoscopy pictures are segmented using Tsallis entropy to locate suspicious lesions. Capsule Network, Computer Aided Diagnosis, and Adagrad optimizer use DLCAL-oriented feature extractors to extract features from segmented lesions. CAL is constructed to link CapsNets for processing and capture class-specific properties for dependency reporting. SSO-based CSAEs are classified last. DLCAL-SLDC is tested on a benchmark ISIC dataset. An unbalanced dataset in this work introduces a novel DL-based skin cancer detector [
55].
The author of [
56] employed RegNetY-320 deep learning models for skin cancer classification. Data augmentation was used to rectify the data imbalance to equalize the distribution of skin cancer classifications. Skin Cancer MNIST: HAM10000 has seven more skin lesions. RegNetY-320, InceptionV3, and AlexNet are deep learning-based skin cancer classifiers. Hyperparameters were varied in numerous combinations to adapt the suggested structure. RegNetY-320 outperformed InceptionV3 and AlexNet in accuracy, F1-score, and "receiver operating characteristic (ROC)" curve compared to the imbalanced and balanced datasets. The proposed structure outperformed more conservative techniques. They may help diagnose illnesses early, minimize unnecessary biopsies, save lives, and lower medical costs for patients, skin specialists, and doctors.
FixCaps enhanced dermoscopy image categorization in the study by [
57]. FixCaps uses a huge in-height presentation kernel, 31*31, at the lowest convolution layer instead of the more common 9*9. It may give FixCaps more viewers than CapsNets. Convolution and pooling lose three-dimensional data when the convolutional block attention segment is added. Collection convolution was used to avoid capsule layer underfitting. The system can reduce calculations and enhance detection accuracy compared to other methods. According to the investigational results, FixCaps had a higher accuracy rate than IRv2-SA, which had 96.49% on the HAM10000 dataset. The study aims to determine how the ability of DL models to generate large data networks affects pharmaceutical manufacturing [
58].
The researchers discovered that image determination does not diminish Sensitivity, Specificity, or accuracy when other features are present. The study by [
59] focuses on how DL-driven recordkeeping systems help doctors discover skin cancer early and how machinery helps doctors provide quality care. While different and effective augmentation approaches are used, training pictures improve CNN design accuracy, Sensitivity, and Specificity. This study proposes extracting and learning essential photo demos using MobileNetV3 design to improve skin cancer detection [
59]. Next, the features are used to modify the "Hunger Games Search (HGS) oriented on Particle Swarm Optimization (PSO) and Dynamic-Opposite Learning (DOLHGS)". This adaptation uses a novel feature selection method to determine the most crucial element to improve the classic's presentation. PH2 and ISIC-2016, two and three classification datasets, were used to evaluate the DOLHGS's effectiveness. The suggested technique achieves 88.19% accuracy on ISIC-2016 and 96.43% on PH2. According to the testing, the proposed method surpassed other popular algorithms in classification accuracy and ideal skin cancer analysis features.
The paper [
60] addresses possible drawbacks and issues with systems for detecting and classifying skin cancer and ML-based implementations. Additionally, they studied five dermatology-related fields using deep learning: skin disorder measurement using smartphones and personal monitoring systems, dermatopathology visual classification of malignancy, and clinical image categorization. By better understanding machine learning and its many applications, dermatologists will be better equipped to identify potential challenges. This study looked at profound learning studies on skin cancer diagnosis to evaluate alternative approaches. This study also laid the foundations for developing an application for diagnosing skin cancer, and it primarily addresses two problems: deep learning-based skin lesion tracking and image segmentation.
This research [
61] suggests a DL-oriented skin cancer categorization network (DSCC_Net) oriented on convolutional neural networks and using three widely accessible standard datasets (ISIC 2020, HAM10000, & DermIS). The suggested DSCC_Net is typically connected to six baseline deep networks for skin cancer classification: ResNet-152, Vgg-16, Vgg-19, Inception-V3, EfficientNet-B0, and MobileNet. To correct the minority classifications in this dataset, they employed SMOTE Tomek. Their DSCC_Net model beats baseline techniques, helping dermatologists and healthcare practitioners identify skin cancer. The purpose of the study [
62] was to (i) address a common class imbalance issue brought about by the fact that persons with skin cancer tend to be smaller than people in good physical shape and (ii) analyze typical production to identify better decision-making. (iii) create an Android application for a comprehensive intelligent healthcare plan to produce reliable deep-learning prediction models. The suggested DL approach was assessed for generalization ability and classification accuracy in association with six popular classifiers. Using an updated CNN and the HAM10000 dataset, this research detected seven cases of skin cancer. A skin lesion classification system utilizing Explainable Artificial Intelligence (XAI) was created, and the outcomes were explained using Grad-CAM and Grad-CAM++ techniques. This method aids in physicians' early skin cancer diagnosis with an 82% classification accuracy and a 0.47% loss accuracy. This work carefully categorized skin cancer using a two-tier approach [
63].
Data augmentation approaches were employed early in the framework to improve picture models for practical training. Based on the encouraging results of medical image processing obtained from the Medical Vision Transformer (MVT), they built an MVT-based classification typically utilized for SC in the second layer of the design. The input picture is divided by this MVT into many segments, which are then sent to the transformer in a sequentially similar term embedding. The input image is finally classified using the Multi-Layer Perceptron. Through tests on the HAM10000 datasets, they discovered that the MVT-based approach outperforms the most recent methods for classifying skin cancer. Deep learning algorithms can recognize melanoma from dermoscopy images [
64]. Fuzzy GrabCut-stacked convolutional neural networks (GC-SCNN) were employed for the imaging experiment. Several publicly available datasets were utilized to extract picture features and classify lesions. The recommended model was shown to detect and classify lesion segments more quickly and accurately compared to the performance of current approaches.
A novel and trustworthy feature fusion model for skin cancer identification was put out [
65]. First, the images are cleaned of noise using a Gaussian Filter (GF). While LBP was utilized for manual extraction, Inception V3 performed automatic feature extraction. The learning rate was controlled using an Adam optimizer. Malignant and benign skin cancers were categorized using an LSTM network based on fused features. Their system integrated techniques from DL and ML. For skin lesions on Kaggle, they used the DermIS dataset, which has 1000 images, 500 of which are benign and 500 of which are malignant. They tested their features-fusion approach against DL- and segmentation-based techniques. After cross-validating their model using a thousand Global Skin Image Collection images, they achieved a detection accuracy of 98.4%. Their method works better than other approaches and yields noteworthy outcomes.
Table 1 represents the research gap in the literature survey.
The research gaps are identified by analyzing the recent available literature and are summarized in
Table 1. To address these gaps, we used preprocessing methods to capture the low-contrast features and active contour segmentation to delineate the skin lesion's borders precisely. ResNet50 transfer learning network captures the textural feature maps and addresses the vanishing gradient problem. Also, the lightweight attention mechanism is integrated into convolutional blocks of the CapsNets network to identify the spatial relationship among various features. CapsNets network reduces the overfitting problem using regularizations and dynamic routing, enhancing model generalization performance.
Figure 1.
Proposed LA-CapsNet for Skin Lesion Classification.
Figure 1.
Proposed LA-CapsNet for Skin Lesion Classification.
Figure 2.
(a) RBB-1; (b). RBB-2.
Figure 2.
(a) RBB-1; (b). RBB-2.
Figure 3.
Design of ResNet50.
Figure 3.
Design of ResNet50.
Figure 4.
Convolutional Block Attention Mechanism [
81].
Figure 4.
Convolutional Block Attention Mechanism [
81].
Figure 5.
Proposed Capsule Network Model.
Figure 5.
Proposed Capsule Network Model.
Figure 6.
Augmented and Normalized Images.
Figure 6.
Augmented and Normalized Images.
Figure 7.
Absolute difference between original and resized image.
Figure 7.
Absolute difference between original and resized image.
Figure 8.
Absolute difference between original and Normalized Image.
Figure 8.
Absolute difference between original and Normalized Image.
Figure 9.
Comparison between Active Contour and Fuzzy K-Means.
Figure 9.
Comparison between Active Contour and Fuzzy K-Means.
Figure 10.
Hair Mask and Removal.
Figure 10.
Hair Mask and Removal.
Figure 11.
Convexity values for Labeled Region.
Figure 11.
Convexity values for Labeled Region.
Figure 12.
Circularity values for Labeled Region.
Figure 12.
Circularity values for Labeled Region.
Figure 13.
Irregularity Index values for Labeled Regions.
Figure 13.
Irregularity Index values for Labeled Regions.
Figure 14.
Texture Pattern Map.
Figure 14.
Texture Pattern Map.
Figure 15.
Color Feature Representation.
Figure 15.
Color Feature Representation.
Figure 16.
Color Feature Histogram Representation.
Figure 16.
Color Feature Histogram Representation.
Figure 17.
Performance Evaluation of the Proposed Model.
Figure 17.
Performance Evaluation of the Proposed Model.
Figure 18.
GUI Interface.
Figure 18.
GUI Interface.
Figure 19.
Number of images vs Accuracy.
Figure 19.
Number of images vs Accuracy.
Figure 20.
Number of Images vs Sensitivity.
Figure 20.
Number of Images vs Sensitivity.
Figure 21.
Number of images vs Specificity.
Figure 21.
Number of images vs Specificity.
Figure 22.
False positive rate vs True positive rate.
Figure 22.
False positive rate vs True positive rate.
Figure 23.
Number of Images vs F1 score .
Figure 23.
Number of Images vs F1 score .
Table 1.
Summary of Literature Review.
Table 1.
Summary of Literature Review.
Ref. |
Objective |
Methods/ Techniques |
Research Gap |
[46] |
Deep learning-based skin cancer detection using dermoscopy pictures. |
Deep neural network algorithms such as faster R-CNN and fuzzy k-means clustering (FKM) |
When employing FKM, the boundaries between distinct areas in the skin lesion images cannot always be clear and precise. |
[47] |
Techniques for detecting skin cancer that categorize the disease as benign, malignant, or normal. |
Fuzzy C-means Clustering (FCM), Rider Optimization Algorithm (ROA) |
FCM clustering faces challenges in complex or textured images, leading to weak convergence and local minima issues, impacting image segmentation quality. |
[48] |
To categorize dermoscopy images into benign or malignant lesions. |
CNN, Support vector machines (SVM) |
The proposed system does not emphasize preprocessing. Thus, it affects input image accuracy. |
[49] |
To improve dermoscopy image learning and skin cancer diagnosis training. |
DNN, DL models |
DNNs require a lot of labelled data for training, making it hard to find and annotate diverse and accurate skin lesion images, especially for rare or specialized malignancies. |
[50] |
Deep Learning-Based Melanoma Classification. |
CNN, Super-Resolution Generative Adversarial Networks (SRGAN) |
CNN may make decision-making and learning features challenging to interpret. The final prediction is complex, with more extracted features. |
[51] |
Skin cancer detection using ML and image processing |
Image Super-Resolution (ISR) algorithms |
ISR image artefacts can affect skin cancer detection. Abnormalities lead to generating diagnostic false positives and negatives. |
[52] |
Teaching-Learning-Based Optimization for Detecting Skin Cancer. |
TLBO algorithm, Extreme Learning Machine (ELM) |
The suggested technique requires a lot of computing power to handle large skin cancer imaging datasets, limiting its practical uses. |
[53] |
An explainable CNN-based method for early melanoma skin cancer detection. |
CNN-based stacked ensemble architecture |
Stacking ensemble frameworks with many models, such as CNNs, can create a complex architecture. Complexity needs more extended training and more resources. |
[54] |
Detecting Skin Cancer using Transfer Learning |
MobileNetV2 |
Due to its low capabilities, MobileNetV2 can have difficulty with complex skin diseases that demand fine-grained characteristics. |
[55] |
Deep learning-based skin cancer detection and categorization. |
Swallow Swarm Optimization (SSO), DLCAL-SLDC method, CAD model |
When CAD systems overlook carcinogenic lesions and misclassify benign lesions as malignant, false positives and negatives occur. Errors cause needless biopsies or missed diagnoses. |
[56] |
DL-Based Skin Cancer Classifier for Unbalanced Datasets. |
Modeling based on deep learning RegNetY-320, InceptionV3, and AlexNet |
Most of these parametric algorithms require uniform data but without controlling its nature. Thus, these approaches cannot accurately diagnose the condition. |
[57] |
Network of Capsules for Skin Cancer Diagnosis |
FixCaps, convolutional block attention module |
FixCaps's generalization performance has not been thoroughly investigated. |
[58] |
Detect Skin Cancer from Food Antioxidants via Deep Learning. |
CNN, DL model |
The suggested system for effective training considers features, classifications, and augmentations, which can overfit data. |
[59] |
A robust skin cancer detection system using transfer learning. |
Optimizing Particle Swarms (PSO) with Dynamic-Opposite Learning |
Proper transfer learning depends on the quantity and quality of the target skin cancer dataset. Transfer learning fails if the dataset is too small or has noisy or biased samples. |
[60] |
DL approaches for detecting and categorizing skin cancer |
CNN, Medical Vision Transformer's |
Privacy considerations and the rarity of some skin cancers make obtaining datasets for skin cancer detection and expert annotations difficult. |
[61] |
DL Models Based Classification for Skin Cancer. |
CNN, EfficientNet-B0, ResNet-152, Vgg-16, Vgg-19, Inception-V3, and MobileNet |
DSCC_Net model only works for light-skinned people. This study omitted dark-skinned people. |
[62] |
Convolutional Neural Network for Cancer Classification. |
CNN, Grad-CAM |
Due to computational costs, access to strong GPUs or cloud computing resources is necessary to train optimized CNN designs. |
[63] |
Skin cancer classification via medical vision. |
Medical Vision Transformer's (MVT), Multi-Layer Perceptron (MLP) |
MLPs don't capture image spatial connections. Skin cancer diagnosis often requires spatial patterns and specific features. |
[64] |
Melanoma identification from dermoscopy images using DL. |
GrabCut-stacked convolutional neural networks (GC-SCNN), SVM |
GrabCut can encounter issues with complex backdrops or parts with similar color distributions to the target object. The algorithm cannot distinguish foreground from background in some cases. |
[65] |
Skin cancer detection model based on feature fusion. |
Local binary patterns (LBP), LSTM, |
LSTM is commonly used for sequential data, including time series or natural language word sequences. This method can convert images into sequential representations, although it cannot be as efficient or precise as Convolutional Neural Networks. |
Table 2.
Skin Lesion Dataset-Class Assessment Metrics.
Table 2.
Skin Lesion Dataset-Class Assessment Metrics.
Class Assessment Metrics using randomly sampled datasets |
Dataset |
No of Images |
ImbR |
IntraC |
InterC |
DistR |
Silho |
HAM 10000 [95] |
7818 |
6.024 |
8705 |
9770 |
0.891 |
0.213 |
ISIC 2020 [96] |
25838 |
9.012 |
28786 |
32132 |
0.804 |
0.202 |
Table 3.
Performance of image after preprocessing phase.
Table 3.
Performance of image after preprocessing phase.
|
PSNR (dB) |
SSIM |
MSE |
Mean Absolute Difference |
Resized Image |
33.52 |
0.97 |
0.0023 |
109.01 |
Normalized Image |
44.90 |
0.97 |
0.0023 |
0.0052 |
Table 4.
Classification accuracy (%) on ISIC2020 test data.
Table 4.
Classification accuracy (%) on ISIC2020 test data.
Table 5.
Comparison of Proposed Model with SOTA Methods.
Table 5.
Comparison of Proposed Model with SOTA Methods.
Performance Metric |
DE-ANN |
ELM-TLBO |
R-CNN |
DCNN |
Fuzzy K-means |
LA-CapsNet |
Accuracy |
96,97 |
96,21 |
97,63 |
97,83 |
94,23 |
98,04 |
Sensitivity |
97,91 |
97,03 |
98,32 |
97,72 |
96,53 |
98,82 |
F1 Score |
97,01 |
85,00 |
98,42 |
90,00 |
96,07 |
98,87 |
AUC |
97,21 |
89,00 |
95,00 |
97,92 |
95,43 |
99,00 |
Specificity |
45,00 |
48,00 |
49,00 |
39,00 |
55,00 |
68,00 |