Altmetrics
Downloads
44
Views
19
Comments
0
This version is not peer-reviewed
Submitted:
29 October 2024
Posted:
31 October 2024
You are already at the latest version
Early detection of cervical cancer is the need of the hour to stop the fatality from this disease. There have been various CAD approaches in the past that promise to detect cervical cancer in an early stage, each of them having some constraints.This study proposes a novel ensemble deep learning model, VLR (variable learning rate), aimed at enhancing the accuracy of cervical cancer classification. The model architecture integrates VGG16, Logistic Regression, and ResNet50, combining their strengths in an offbeat ensemble design. VLR learns in two ways: firstly, by dynamic weights from three base models, each of them trained separately; secondly, by attention mechanisms used in the dense layer of the base models. Hyperparameter tuning is applied to further reduce loss, fine tune the model’s performance, and maximize classification accuracy. We performed K-fold cross-validation on VLR to evaluate any improvements in metric values resulting from hyperparameter fine tuning. We have also validated our model on images captured in three different solutions and on a secondary dataset. Our proposed VLR model outperformed existing methods in cervical cancer classification, achieving a remarkable training accuracy of 99.95% and a testing accuracy of 99.89%.
Algorithm 1VLR Ensemble Model incorporating Attention Mechanism, Static Weights, Fine-Tuning, and K-Fold Cross Validation for VLR |
Ep-och | VGG16 | ResNet50 | LR | VLR Loss Contrib | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Val-Acc (%) | Train-Loss | Train-Time (sec) | Dyn- | Val-Acc (%) | Train-Loss | Train-Time (sec) | Dyn- | Val-Acc (%) | Train-Loss | Train-Time (sec) | Dyn- | ||
1 | 90.05 | 0.00912 | 3240 | 0.099 | 91.22 | 0.00880 | 2640 | 0.098 | 90.87 | 0.00850 | 220 | 0.088 | 0.004423 |
20 | 94.12 | 0.00750 | 3600 | 0.109 | 95.67 | 0.00710 | 3200 | 0.109 | 95.80 | 0.00690 | 180 | 0.098 | 0.003753 |
40 | 95.13 | 0.00730 | 3960 | 0.111 | 96.12 | 0.00690 | 3520 | 0.110 | 96.50 | 0.00660 | 140 | 0.100 | 0.003666 |
60 | 95.65 | 0.00720 | 4320 | 0.112 | 96.75 | 0.00680 | 3840 | 0.111 | 96.90 | 0.00640 | 100 | 0.101 | 0.003620 |
80 | 96.25 | 0.00711 | 4560 | 0.113 | 97.20 | 0.00660 | 3960 | 0.113 | 97.30 | 0.00620 | 60 | 0.103 | 0.003554 |
100 | 96.55 | 0.00700 | 4700 | 0.114 | 97.65 | 0.00650 | 4100 | 0.114 | 97.80 | 0.00600 | 40 | 0.104 | 0.003505 |
120 | 96.77 | 0.00711 | 4800 | 0.114 | 98.10 | 0.00631 | 3920 | 0.115 | 98.02 | 0.00349 | 20 | 0.135 | 0.003315 |
140 | 96.77 | 0.00711 | 4840 | 0.114 | 98.23 | 0.00631 | 3960 | 0.115 | 98.92 | 0.00349 | 17 | 0.136 | 0.003317 |
150 | 96.77 | 0.00711 | 4860 | 0.114 | 98.23 | 0.00631 | 3960 | 0.115 | 98.92 | 0.00349 | 15 | 0.136 | 0.003317 |
Model | Raw-Score () | Exponential () | Attention-Weight-Formula () | Attention-Weight-Value () |
---|---|---|---|---|
ResNet50 | 3.0 | 20.09 | 0.574 | |
VGG16 | 2.5 | 12.18 | 0.348 | |
Logistic-Regression (LR) | 1.0 | 2.72 | 0.078 |
Ep-och | Learn Rate | Batch Size | Val Loss | Train Loss | Val Acc (%) | Pre-cision (%) | Recall (%) | TP / FP / FN |
---|---|---|---|---|---|---|---|---|
90 | 0.001 | 100 | 0.0065 | 0.003317 | 98.56 | 98.10 | 97.80 | 2188 / 12 / 8 |
100 | 0.0005 | 100 | 0.0052 | 0.003304 | 99.10 | 98.25 | 97.85 | 2189 / 11 / 7 |
110 | 0.0001 | 80 | 0.0041 | 0.00325 | 99.45 | 98.40 | 97.88 | 2190 / 9 / 7 |
120 | 0.0001 | 60 | 0.0035 | 0.00275 | 99.60 | 98.50 | 97.90 | 2191 / 8 / 6 |
130 | 0.00005 | 60 | 0.0030 | 0.00250 | 99.68 | 98.55 | 97.92 | 2192 / 7 / 6 |
140 | 0.00005 | 50 | 0.0026 | 0.00230 | 99.72 | 98.60 | 97.95 | 2193 / 6 / 5 |
150 | 0.00001 | 50 | 0.0023 | 0.00214 | 99.77 | 98.65 | 97.97 | 2194 / 5 / 5 |
160 | 0.000005 | 40 | 0.0021 | 0.00205 | 99.80 | 98.68 | 97.99 | 2194 / 5 / 4 |
170 | 0.000005 | 40 | 0.0019 | 0.00200 | 99.82 | 98.70 | 98.00 | 2194 / 4 / 4 |
180 | 0.000001 | 35 | 0.0018 | 0.00192 | 99.85 | 98.72 | 98.03 | 2194 / 4 / 3 |
190 | 0.000001 | 35 | 0.0017 | 0.00185 | 99.87 | 98.74 | 98.05 | 2195 / 3 / 2 |
200 | 0.0000005 | 30 | 0.0015 | 0.00175 | 99.89 | 98.75 | 98.08 | 2195 / 3 / 2 |
Data-set | VGG16 | ResNet50 | LR | VLR-Loss | VLR-vali-dation-acc | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Val-Acc (%) | Train-Loss | Static | Dyn-mic | Final | Val-Acc (%) | Train-Loss | Static | Dyna-mic | Final | Val-Acc (%) | Train-Loss | Static | Dyna-mic | Final | Train-Loss-Contrib | Final-Acc | |
Lugol’s-iodine | 91.67 | 0.00821 | 0.348 | 0.3352 | 0.3389 | 92.43 | 0.00731 | 0.574 | 0.3713 | 0.4670 | 94.42 | 0.00649 | 0.078 | 0.2935 | 0.1903 | 0.00708 | 99.21 |
Acetic-acid | 90.17 | 0.00864 | 0.348 | 0.3360 | 0.3396 | 91.23 | 0.00791 | 0.574 | 0.3745 | 0.4705 | 92.12 | 0.00744 | 0.078 | 0.2895 | 0.1883 | 0.00746 | 99.10 |
Normal-saline | 87.87 | 0.00978 | 0.348 | 0.3366 | 0.3400 | 88.73 | 0.00912 | 0.574 | 0.3764 | 0.4723 | 90.04 | 0.00842 | 0.078 | 0.2870 | 0.1865 | 0.00802 | 99.02 |
Malhari | 93.45 | 0.00814 | 0.348 | 0.3332 | 0.3366 | 94.21 | 0.00701 | 0.574 | 0.3695 | 0.4642 | 95.66 | 0.00631 | 0.078 | 0.2973 | 0.1931 | 0.00695 | 99.41 |
Model | Val-Acc (%) | Prec-ision (%) | Recall (%) | F1-Score (%) | ROC-AUC (%) | Train Time (sec-(TPU-V28) ) | Train Loss |
---|---|---|---|---|---|---|---|
VGG16 | 96.77 | 95.47 | 96.12 | 95.84 | 93.255 | 4860 | 0.00711 |
ResNet50 | 98.23 | 96.65 | 97.24 | 97.02 | 95.854 | 3960 | 0.00631 |
LR | 98.92 | 97.25 | 97.99 | 97.50 | 97.891 | 15 | 0.00349 |
VLR | 99.89 | 98.75 | 98.08 | 98.41 | 99.991 | 5040 | 0.00175 |
Metric | Formula | VGG16 (%) | ResNet50 (%) | LR (%) | VLR (%) |
---|---|---|---|---|---|
Validation Accuracy | 96.77 | 98.23 | 98.92 | 99.89 | |
Precision (P) | 95.47 | 96.25 | 97.25 | 98.75 | |
Recall (R) | 96.12 | 97.24 | 97.99 | 98.08 | |
True Positives (TP) | 2115 | 2160 | 2189 | 2195 | |
False Positives (FP) | 30 | 20 | 7 | 3 | |
False Negatives (FN) | 55 | 20 | 4 | 2 |
Metric | Fold 1 | Fold 2 | Fold 3 | Fold 4 | Fold 5 | Mean (%) | Standard Deviation (%) |
---|---|---|---|---|---|---|---|
Validation Accuracy (%) | 99.90 | 99.88 | 99.91 | 99.87 | 99.89 | 99.89 | 0.02 |
Training Accuracy (%) | 99.96 | 99.95 | 99.95 | 99.94 | 99.95 | 99.95 | 0.01 |
Precision (%) | 98.76 | 98.77 | 98.74 | 98.75 | 98.76 | 98.75 | 0.01 |
Recall (%) | 98.09 | 98.08 | 98.11 | 98.06 | 98.07 | 98.08 | 0.02 |
F1-Score (%) | 98.47 | 98.45 | 98.48 | 98.44 | 98.46 | 98.46 | 0.02 |
ROC-AUC (%) | 99.99 | 99.98 | 99.99 | 99.98 | 99.99 | 99.99 | 0.01 |
Training Loss | 0.00176 | 0.00174 | 0.00175 | 0.00173 | 0.00175 | 0.00175 | 0.00002 |
Solution | Train Accuracy (%) | Validation Accuracy (%) | Pre-cision% | Recall-% | F1-Score% |
---|---|---|---|---|---|
Lugol’s iodine | 99.41 | 99.21 | 98.69 | 97.81 | 98.22 |
Acetic acid | 99.23 | 99.10 | 98.14 | 97.23 | 97.64 |
Normal saline | 99.14 | 99.02 | 97.12 | 96.04 | 96.56 |
Malhari | 99.74 | 99.41 | 98.62 | 98.11 | 98.36 |
Model | Accuracy (%) | Precision | Recall | F1-Score |
---|---|---|---|---|
MFEM-CIN[7] | 89.2 | 92.3 | 88.16 | 90.18 |
DL[4] | 92 | 79.4 | 100 | 97.58 |
CNN[44] | 87 | 75 | 88 | 80.55 |
CNN[45] | 90 | NA | NA | NA |
CNN[46] | 84 | 96 | 99 | 97.47 |
DenseNet-CNN[47] | 73.08 | 44 | 87 | 58.44 |
CLD Net[48] | 92.53 | 85.56 | NA | NA |
Kappa[49] | 67 | 89 | 33 | 76.44 |
Caps Net[50] | 99 | NA | NA | NA |
E-GCN[3] | 81.85 | 81.97 | 81.78 | 81.87 |
VLR | 99.95 | 98.75 | 98.08 | 98.41 |
Reference | Dataset | Method | Accuracy | Remarks |
---|---|---|---|---|
[7]2024 | Colposcopy | CNN, Transformer | 89.2% | Employs a hybrid model of MFEM-CIN and Transformer. |
[12]2024 | Pap Smear | CNN, ML, DL | 99.95% | Combines computational tools like CNN-LSTM hybrids, KNN, and SVM. |
[16]2024 | Pap Smear | Deep Learning | 92.19% | Employs CerviSegNet-DistillPlus for classification. |
[15]2024 | Biopsy | Machine Learning | 98.19% | Employs ensemble of machine learning algorithms for classification. |
[17]2024 | Colposcopy | Deep Learning | 94.55% | Has used a hybrid deep neural network for segmentation. |
[2]2023 | Pap Smear | Deep Learning | 97.18% | Uses deep learning techniques integrated with advanced augmentation techniques such as CutOut, MixUp, and CutMix. |
[4]2023 | Colposcopy | Deep Learning | 92% | Uses predictive deep learning model. |
[30]2022 | Pap Smear | CNN, PSO, DHDE | 99.7% | Applies CNN for a multi-objective problem, and PSO and DHDE are used for optimization. |
[14]2022 | Cytology | ANN | 98.87% | Uses artificial Jellyfish Search optimizer combined with an ANN. |
[9]2021 | Colposcopy | Deep Learning | 92% | Uses Deep neural techniques for cervical cancer classification. |
[11]2021 | Colposcopy | Deep Learning | 90% | Using deep neural network generated attention maps for segmentation. |
[10]2021 | Colposcopy | Residual Learning | 90%, 99% | Employed residual network using Leaky ReLU and PReLU for classification. |
[29]2021 | MR-CT Images | GAN | - | Uses a conditional generative adversarial network (GAN). |
[21]2021 | Pap Smear | Biosensors | - | Uses biosensors for higher accuracy. |
[5]2021 | Cervical Pathology Images | SVM, k-Nearest Neighbors, CNN, RF | 90%-89.1% | Uses ResNet50 model as a feature extractor and selects k-NN, Random Forest, and SVM for classification. |
[1]2021 | Colposcopy | CNN | 99% | Uses Faster Small-Object Detection Neural Networks. |
[13]2021 | Pap Smear | Deep Convolutional Neural Network | 95.628% | Constructs a CNN called DeepCELL with multiple kernels of varying sizes. |
[28]2020 | MRI Data of Cervix | Statistical Model | - | A statistical model called LM is used for outlier detection in lognormal distributions. |
[3]2020 | Colposcopy | CNN | 81.95% | Employs a graph convolutional network with edge features (E-GCN). |
[8]2020 | Colposcopy | Deep Learning | 83.5% | Has used K-means for classification, CNN and XGBoost to combine the CNN decesion. |
[6]2019 | Colposcopy | CNN | 96.13% | Uses a recurrent convolutional neural network for classification of cervigrams . |
Our Method | Colposcopy | Deep Neural Network | 99.95% | An ensemble model called Variable Learning Rate specially designed to increase the accuracy. |
Sl.No. | Gaps | Corrective Measures |
---|---|---|
1.[3,9] | less classification accuracy | VLR projects a training accuracy of 99.95% |
2.[12,29] | variations in dataset used for classification | correct dataset used; colposcopy |
3.[4,44] | recall values are higher than precision | VLR projects precision as 98.75% and recall as 98.08% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
Xia LI
et al.
,
2021
Tengku Muhammad Hanis
et al.
,
2023
Carlos Julio Macancela
et al.
,
2023
© 2024 MDPI (Basel, Switzerland) unless otherwise stated