SciELO - Scientific Electronic Library Online

 
vol.33 issue67Development of an Application for Calculating the Power Flow of Bipolar DC Networks Using the MATLAB Environment author indexsubject indexarticles search
Home Pagealphabetic serial listing  

Services on Demand

Journal

Article

Indicators

Related links

  • On index processCited by Google
  • Have no similar articlesSimilars in SciELO
  • On index processSimilars in Google

Share


Revista Facultad de Ingeniería

Print version ISSN 0121-1129On-line version ISSN 2357-5328

Rev. Fac. ing. vol.33 no.67 Tunja Jan./Mar. 2024  Epub Mar 30, 2024

https://doi.org/10.19053/01211129.v33.n67.2024.16943 

Articles

HASCC: A Hybrid Algorithm for Skin Cancer Classification

HASCC: Algoritmo Híbrido para Clasificación de Cáncer de Piel

HASCC: Algoritmo Híbrido para Classificação de Câncer de Pele

Carlos-Vicente Niño-Rondón1 
http://orcid.org/0000-0002-3781-4564

Diego-Andrés Castellano-Carvajal2 
http://orcid.org/0000-0002-4530-1136

Sergio-Alexander Castro-Casadiego3 
http://orcid.org/0000-0003-0962-9916

Byron Medina-Delgado4 
http://orcid.org/0000-0003-0754-8629

Karla-Cecilia Puerto-López5 
http://orcid.org/0000-0003-3749-676X

1 M. Sc. Pontificia Universidad Javeriana Cali (Cali-Valle del Cauca, Colombia). carlosvicenteninor@javerianacali.edu.co. ORCID: 0000-0002-3781-4564

2 Universidad Francisco de Paula Santander (Cucutá-Norte de Santander, Colombia). diegoandrescc@ufps.edu.co. ORCID: 0000-0002-4530-1136

3 M. Sc. Universidad Francisco de Paula Santander (Cucutá-Norte de Santander, Colombia). sergio.castroc@ufps.edu.co. ORCID: 0000-0003-0962-9916

4 Ph. D. Universidad Francisco de Paula Santander (Cucutá-Norte de Santander, Colombia). byronmedina@ufps.edu.co. ORCID: 0000-0003-0754-8629

5 M. Sc. Universidad Francisco de Paula Santander (Cucutá-Norte de Santander, Colombia). karlaceciliapl@ufps.edu.co. ORCID: 0000-0003-3749-676X


Abstract

Skin cancer is a dangerous and potentially lethal disease that is steadily increasing worldwide. Signs of skin cancer may include changes in the appearance of moles or the emergence of new spots on the skin. Early detection is crucial, as many types of skin cancer respond well to treatment when addressed in the early stages. Computer-aided diagnostic tools are employed to aid in the diagnosis of this disease. This article introduces HASCC, a hybrid algorithm implemented through a graphical user interface for skin cancer classification. The algorithm integrates image processing, feature extraction using the VGG16 algorithm with component reduction through PCA, and classification using XGBoost trained on images from the HAM10000 dataset. The hybrid algorithm was executed and tested on a Raspberry Pi 4 embedded system. HASCC was compared at both hardware and software levels with other computational intelligence methods and architectures, revealing notable improvements in terms of accuracy, ranging from 88.2% to 93.2%, with an average execution time of 250 milliseconds at low machine resource demand during the diagnostic process. Additionally, HASCC's performance was compared against previous research focused on skin cancer detection and classification. The hardware performance demonstrates that HASCC can be implemented on single-board microprocessor devices, and its software performance suggests viability for supporting the diagnosis and classification of skin cancer.

Keywords: computer-aided diagnosis; embedded system; graphical user interface; hybrid algorithm; open-source; skin cancer

Resumen

El cáncer de piel es una enfermedad peligrosa y potencialmente letal que aumenta constantemente en los reportes de casos de cáncer a nivel mundial. Los signos de cáncer de piel pueden incluir cambios en la apariencia de los lunares o la aparición de nuevas manchas en la piel. La detección temprana es fundamental, ya que muchos tipos de cáncer de piel responden bien al tratamiento si se abordan en las etapas iniciales. Para el apoyo en el diagnóstico de esta enfermedad se emplean herramientas de diagnóstico asistido. Este artículo presenta HASCC, un algoritmo híbrido implementado mediante una interfaz gráfica de usuario para la clasificación del cáncer de piel. El algoritmo integra procesamiento de imágenes, extracción de características mediante el algoritmo VGG16 con reducción de componentes mediante PCA y clasificación mediante XGBoost entrenado con imágenes del Conjunto de Datos HAM10000. El algoritmo híbrido se ejecutó y se probó sobre un sistema embebido Raspberry Pi 4. HASCC se comparó a nivel hardware y a nivel software con otros métodos y arquitecturas de inteligencia computacional, y se obtuvo que el sistema propuesto mostró mejores notables en términos de precisión, que osciló entre el 88.2 % y 93.2 %, con un tiempo promedio de ejecución de 250 milisegundos a baja demanda de recursos de máquina durante el proceso de diagnóstico. Adicionalmente, el rendimiento de HASCC se comparó contra investigaciones previas enfocadas a la detección y clasificación de cáncer de piel. El rendimiento a nivel hardware demuestra que HASCC es viable para implementación en dispositivos microprocesadores de placa única, y con su desempeño a nivel de software se infiere que es viable para el apoyo en el diagnóstico y clasificación del cáncer de piel.

Palabras clave: algoritmo híbrido; cáncer de piel; código abierto; diagnóstico asistido por computador; interfaz gráfica de usuario; sistema embebido

Resumo

O câncer de pele é uma doença perigosa e potencialmente letal que vem aumentando constantemente nos relatos de casos de câncer em todo o mundo. Os sinais de câncer de pele podem incluir alterações na aparência de manchas ou aparecimento de novas manchas na pele. A detecção precoce é essencial, pois muitos tipos de câncer de pele respondem bem ao tratamento se tratados nos estágios iniciais. Ferramentas de diagnóstico assistido são utilizadas para apoiar o diagnóstico desta doença. Este artigo apresenta o HASCC, um algoritmo híbrido implementado usando uma interface gráfica de usuário para classificação do câncer de pele. O algoritmo integra processamento de imagens, extração de características usando o algoritmo VGG16 com redução de componentes usando PCA e classificação usando XGBoost treinado com imagens do Dataset HAM10000. O algoritmo híbrido foi executado e testado em um sistema embarcado Raspberry Pi 4 e comparado a nível de hardware e a nível de software com outros métodos e arquiteturas de inteligência computacional, e obteve-se que o sistema proposto apresentou melhores resultados em termos de precisão. , que variou de 88,2% a 93,2%, com tempo médio de execução de 250 milissegundos com baixa demanda de recursos da máquina durante o processo de diagnóstico. Além disso, o desempenho do HASCC foi comparado com pesquisas anteriores focadas na detecção e classificação do câncer de pele. O desempenho ao nível do hardware demonstra que o HASCC é viável para implementação em dispositivos microprocessados ​​de placa única, e com o seu desempenho ao nível do software infere-se que é viável para apoiar o diagnóstico e classificação do cancro da pele.

Palavras-chave: algoritmo híbrido; câncer de pele; código aberto; diagnóstico auxiliado por computador; interface gráfica do usuário; sistema embarcado

I. Introduction

Image processing and artificial intelligence are used to enhance and extract features from images, facilitating machine learning of models for skin cancer disease detection and classification. Currently, classical machine learning and deep learning models with variations and integrations of learning models are used. These models present a high efficiency in benign and malignant diagnostic classifications, although, in some cases, a special hardware tool is needed to execute the computational models.

In the detection of skin cancer, classification between benign and malignant lesions is prioritized [1], with a benign classification success rate of 82.77% and 88.61% using the ResNet101 and InceptiónV3 models, and a malignant classification success rate of 85.66 % and 86% respectively for these models [2]. Similarly, support vector machines have been used for classification, where sensitivity reaches 90% for benign lesions, and the classification efficiency is just over 93% [3]. However, works have been developed with an emphasis on skin cancer type identification, where the melanoma type classification reached an efficiency of 91% with 111 dermatoscopic images, and a success rate of more than 90% in carcinoma classification with 135 images with test data. On the other hand, with validation data, the accuracy of the employed framework drops to 72.1% when tested with two classes. When adding more classes, its overall accuracy performance drops to 55.4% [4]. Additionally, combined neural networks have been employed for the diagnosis of non-pigmented skin cancer using the InceptionV3 model and a ResNet50, where cancer identification presented an accuracy of 72.5% in 95% of clinical practices, and specific diagnosis among malignant types of skin cancer showed an increase to 74.2% in 95% of practices [5].

VGG is a neural network architecture used in skin lesion detection and identification processes. In its VGGNet extension, it has been fused and compared with other CNN architectures such as GoogleNet, AlexNet, and ResNet, where with VGGNet, the average hit rate was 81.3%, expressing the need to use high-level and cost hardware tools [6]. Furthermore, a VGG architecture with 13 convolutional layers, 5 clustering layers, and 3 fully connected layers with input sizes of 227*227*3 and 200 iterations has been applied, where the hit rate in lesion classification ranged from 60% to 90% [7]. Likewise, XGBoost has been implemented as a classification model, where with the default hyperparameters defined and with test image sizes of 64x64, a recall of 44% in the identification of melanoma, and specifics of 69% in the identification of non-melanoma lesions was presented [8]. Additionally, the performance of XGBoost classifications has been compared with RNNs, where XGBoost decreases the efficiency with respect to RNNs. However, XGBoost does not suffer from information loss effects due to data transformation for RNN model fits [9].

This paper presents HASCC, a hybrid algorithm developed for skin cancer classification between carcinoma (C1), benign keratosis (C2), and melanoma (C3) with images from the HAM10000 dataset. HASCC contains image processing, feature extraction, and reduction stages based on VGG16 and PCA, and a classifier based on XGBoost. This algorithm is integrated into a CAD-like graphical interface, runs on a Raspberry Pi 4 board, and its performance is measured in hardware and software.

II. Methodology

A methodology based on three stages is proposed. In the first stage, the hybrid algorithm developed for detecting and classifying skin cancer, such as carcinoma, benign keratosis, and melanoma, is detailed. The second stage presents the graphical user interface designed for the CAD tool. The third stage validates the tool developed on the Raspberry Pi 4 through a hardware and software comparison against classification algorithms.

A. Hybrid Algorithm for Computer-Aided Diagnosis

The hybrid algorithm developed for skin cancer classification on dermoscopic images integrates algorithms based on image processing techniques, such as feature extraction and reduction, using the VGG16 algorithm for extraction and PCA for dimensionality reduction of image components, as well as XGBoost for classification.

1) Image Processing Techniques for Body Hair Suppression. The input image is initially subjected to a preprocessing step by converting to grayscale with median filter smoothing and kernel size with k=3. Subsequently, the Sobel operator is applied to identify body hair in the lesion. In this case, the first derivative for the horizontal and vertical components is applied to the grayscale image. In addition, erosion morphology is applied to attenuate the effects generated by the noise produced by the body hair on the lesion, using a cross-shaped structural element with a kernel size of 5x5. Moreover, the thresholding method is applied to separate the lesion region from the image background. The binary thresholding method (with a variable value between 0 and 255 pixels from the graphical interface) is used, complemented with contour search stages, to detect the region of interest in the image. In addition, the Otsu adaptive thresholding method is employed for a first approximation of the background lesion segmentation. In this case, the reinforcement threshold is 127.

Image processing culminates with the application of morphological aperture to smooth the contours of the thresholded image with an elliptical structural element with k=5. Furthermore, a superposition is performed between the image with median smoothing and the thresholded region of interest with morphological aperture, where the image without body hair is represented in the foreground.

2) VGG16 and PCA Algorithms. The images used are part of the ISIC Challenge skin cancer dataset consisting of 10,015 pigmented dermoscopic image samples validated with different modalities, in which the dimension of each image is 600x450 pixels. For feature extraction, a VGG16 convolutional neural network architecture is made without layers fully connected by its extraction application, composed of 5 convolutional layers of 64, 128, 256, and 512 channels and 5 grouping layers with 2x2 stride, optimized and grouped with 3x3 size kernels for the convolutional layers and 2x2 size kernels for the grouping layers that allow improving the performance of the network for the extraction. According to the above, an input image size of 128x128 pixels in RGB scale is proposed for feature extraction. The data distribution is 60% for training and 40% for validation and testing.

An image component reduction method is applied to reduce the computational requirements. In this case, the Principal Component Analysis (PCA) method is used [10]. With the image in RGB color space, the channels are separated, and a normalized data raster is obtained and transformed so that the number of components to be reduced is input. In the case of the ISIC Challenge Dataset images, the initial and original 1,890 image components extracted with the VGG16 network were reduced to 375 components for the hybrid algorithm with a total explained variance ratio of 96.6%, representing most of the variance of the information.

Following the reduction of the components by separate channels, the components are integrated in a form that generates a multi-channel image for classification.

3) XGBoost Classifier. The XGBoost hyperparameters referred to the learning rate, the number of reinforced trees per gradient (equivalent to iterations per reinforcement), the maximum tree depth set that makes the classifier less susceptible to overfitting, the Lagrangian operator (gamma), the ratio of random training subsamples to train each tree, the subsampling ratio of each column when constructing each tree and the random seed were set to 0.1, 400, 5, 0.04, 0.8, 0.85 and 27, respectively.

B. Graphical User Interface

For the design of the graphical user interface, two factors are taken into consideration: the identification of user requirements and the identification of the task [11]. Regarding the identification of user requirements, it is necessary to protect the personal information provided to the system, so a credential validation stage is required to access the system. Likewise, a characterization stage is required so that potential patients can be individualized. In addition, this information must be sent to a spreadsheet file in real time. Moreover, the assisted diagnosis tool for the identification and classification of skin cancer through images requires the specialist to adapt and segment the image using sliders, and also to determine the kernel size of the blurring filter to be applied to the image.

Regarding task identification, the specialist must log in to the system by validating credentials to access the characterization stage, in which they enter the necessary information to individualize the patient and save this information. Then, they access HASCC and enter the input image into the system. The specialist has the autonomy to use horizontal sliders to segment the image as far as they wish. They can also apply smoothing filtering using horizontal sliders if the image has body hair. In addition, the method used for diagnosis must be selected and visualized.

C. Validation of HASCC on the Embedded System

The HASCC was validated on a Raspberry Pi 4 embedded system. Its performance was weighted against three methods based on computational intelligence algorithms: AutoKeras (a machine learning distribution of Keras), LightGBM +GLCM (Light Gradient Boosting Machine + Grey-Level Co-occurrence Matrix), and PCA+VGG16 (as a classifier). The comparison is performed both for hardware and software [12]. Regarding hardware, the response time from the time of execution, RAM used, CPU used, and the power consumed by the Raspberry Pi 4B at the time of HASCC execution were analyzed. In addition, the software comparison was performed considering parameters such as accuracy, precision, recall, and F1-Score [13]. Additionally, the HASCC was compared to previous work developed for the identification and classification of skin cancer.

III. Results and Discussion

Figure 1 shows the image processing applied as the initial process of the proposed hybrid algorithm. Section A shows 3 original images corresponding to the classification classes: benign keratosis, carcinoma, and melanoma. Similarly, in B, the grayscale images are shown, while in C and D, those resulting from applying the first derivative for edge detection using Sobel for the horizontal and vertical components are presented, respectively. D illustrates the Sobel resultant, while E and F show the thresholded images using the binary thresholding method and Otsu's method, respectively. The graphical interface allows the determination of the best segmentation using any of the thresholding techniques since the efficiency of this process depends on factors such as the level of incident light at the time of image capture.

Fig. 1 Image processing techniques. 

In Figure 2, the distribution of PCA components, the confusion matrix, and the ROC curve obtained for the hybrid algorithm are depicted.

Fig. 2 HASCC performance. 

Figure 2a shows the dispersion for the first and last pair of components when applying PCA, where the overlapping effect between classes is mitigated. In addition, the confusion matrix is shown in Figure 2b, where the true positives are represented with values of 382, 343, and 344 for C1, C2, and C3, respectively. Moreover, the HASCC ROC curves are presented in Figure 2c, from which the sensitivity rate of 88.2% for C1, 80.9% for C2, and 85.4% for C3 is inferred, indicating the stability of the model in determining the positive proportions of C1, C2, and C3.

Figure 3 shows the system architecture for HASCC. Access to the system is initially represented by the integrated RPi 4 board, validating the login credentials. Subsequently, the diagnostic GUI is accessed, where the image is loaded into the system and the skin cancer type is classified using the HASCC.

Fig. 3 Architecture of the computer-aided diagnosis tool based HASCC. 

Table 1 shows the results of the comparison between the proposed hybrid skin cancer classification algorithm (HASCC) and the other hardware and software classification algorithms, in which the shaded values represent the best results. HASCC takes approximately 40 times less time to train than AutoKeras, 1.39 times less time than LightGBM + GLCM, and 1.76 times less time than PCA + VGG16. Likewise, HASCC had the shortest response time upon execution, being 7.6 times faster than AutoKeras, 6 times faster than LightGBM + GLCM, and 1.6 times faster than VGG16. In terms of RAM memory, HASCC was the least demanding, with a decrease of 12.6%, 10.2%, and 2 % when compared to the other algorithms. It also required 3.1 times less CPU level during processing than AutoKeras, 1.4 times less than LightGBM + GLCM, and approximately the same percentage as VGG16. In terms of power, it was only outperformed by the LightGBM + GLCM algorithm, which consumed 2.95 W, while HASCC consumed 3.05 W.

Table 1 Performance of the HASCC considering hardware and software efficiency. 

Accuracy
Method Training time Response time RAM % CPU % Power C1 C2 C3
HASCC 83.4 0.25 42.7 21.4 3.05 93.2 88.5 88.2
AutoKeras 3325.1 1.9 55.3 65.8 3.2 87.9 80.3 84.8
LightGBM +GLCM 116.2 1.5 52.9 30 2.95 91.8 82.1 84.4
PCA+VGG16 146.55 0.4 44.7 21.6 4.35 92 85.8 86

Method Precision Recall F1-Score
C1 C2 C3 C1 C2 C3 C1 C2 C3
HASCC 88.2 80.9 85.6 91.6 84.3 79.1 89.9 82.6 82.2
AutoKeras 89.6 65.3 74.2 78.2 73.3 77.5 83.5 69.1 75.8
LightGBM +GLCM 90.1 71.2 75.7 86.7 74.4 75.5 88.3 72.8 75.6
PCA+VGG16 90.8 76.9 77.7 86.6 80.1 78.4 88.6 78.5 78.1

At the software level, HASCC showed better performance concerning the accuracy of other algorithms since, for each of the three classes, it reached 93.2% for C1, 88.5% for C2, and 88.2% for C3. In terms of precision, it was only surpassed by VGG16 in the C1 class, which improved performance by 2.6%, while HASCC improved accuracy by up to 15.6% for the C2 class, and by up to 11.4% for the C3 class. Likewise, HASCC performed better for each of the three skin cancer type classes, improving by up to 13.4%, 11%, and 3.7% at the recall level, indicating an improvement of the proposed hybrid algorithm to identify the cancer type.

The values were confirmed with the F1-Score metric, which compares the combined performance of the metrics, in which HASCC presented the highest values with 89.9%, 82.6%, and 82.2% for the three classes. The improvement of HASCC with respect to the other algorithms and methods at the software level can be explained by the reduction of components prior to the classification stage due to its representation in terms of least squares, decreasing from 1,890 to 375 image components. With this, both the execution time and the percentages of RAM and CPU used are improved.

In terms of power, the method based on LightGBM + GLCM, although it was 1.03 times faster, it gives up performance at the software level due to the consideration of the image purely in grayscale. HASCC also improved software-level performance in terms of accuracy, decreasing the gap between the numbers of false and true positives and false negatives and true and negatives in the classification of each of the three classes. At the accuracy level, HASCC presented the lowest performance with 88.2% in the carcinoma class, which was an expected result since only true and false positives are taken into consideration, not including the number of negatives represented in the classifier confusion. In contrast, it showed better performance for the classes C2 and C3, which are the most complex to classify due to the similarity in appearance of the lesion. In terms of recall, HASCC was the best performer for each of the three classes, indicating a remarkable improvement of this method at the level of correct classifications of the classes (true positives). This justifies the improvement in the F1-Score, indicating that HASCC presents a better balance between precision and recall for each of the classes through the harmonic mean.

Table 2 presents the comparison between HASCC and related work. [14] did not differentiate between specific types of skin lesions but between benign and malignant. Still, HASCC improved on average the classification performance of benign malignant lesions considered in the study by 5.75% compared to Resnet101, and by 2.66% compared to InceptionV3. In [15], all classes available in the ISIC-Archive dataset were considered, and the Resnet 50 architecture was used. The accuracy, on average, outperformed HASCC by 2.13%. However, the method used in this study contemplated the data augmentation of the HAM10000 dataset in classes containing a small number of images, which, by increasing proportionally, increases the chances of overfitting and overlearning in training. Similarly, in [16], the same classes tested by HASCC were considered, using the MobileNet and Modified MobileNet architecture, where HASCC improves performance by 9.83% over MobileNet, and by 6.04% over Modified MobileNet.

Table 2 HASCC vs related works. 

Work Dataset Class Architecture Result
HASCC ISIC-Archive (HAM 10000) Akiec, bkl, mel VGG16 - PCA - XGBoost ACC: 93.2 %, 88.5 %, 88.2 %,
[14] ISIC-Archive Malign and no malign Resnet 101 - Inception V3 ACC: 85.66 %; 82.77 % - 86 %; 88.61 %
[15] ISIC-Archive (HAM 10000) All classes of HAM10000 Resnet 50 ACC: 92.1 %
[16] ISIC-Archive (HAM 10000) Mel, akiec (malign), bkl (benign) MobileNet, modified MobileNet ACC: 80.14 %, 83.93 % SPE: 80 %, 84 %
[17] ISIC-Archive Mel, benign U-MobileNetV1, U-DenseNet121 ACC: 87.6 %

HASCC improves in specificity by 5% over MobileNet and by 1% over MobileNet Modified. Furthermore, in [17], only melanoma class and benign lesions were considered for classification, using a fusion between U-MobileNetV1 and U-DenseNet121 architectures, where classification hits were 2.37% lower on average than those developed by HASCC. Although HASCC improves the classification processes of skin lesions, there is a need to improve the melanoma class samples to improve differentiation, given their similarity to other lesion types.

Figure 4 shows the integrated graphical user interface for HASCC. It shows the stages of credential validation, patient characterization, visualization, and lesion diagnosis.

Fig. 4 Graphical user interface developed for HASCC. 

IV. Conclusion

In this work, a hybrid algorithm for skin cancer classification is proposed. The use of image processing techniques enabled body hair removal, feature extraction using a robust method such as VGG16, component reduction using PCA, and selection of information of interest to deliver to the XGBoost classifier, which directly influenced the software performance and model hardware requirements for implementation. The results showed an improvement of HASCC over conventional and merged architectures, as well as related work in skin cancer classification. By running HASCC on a Raspberry Pi 4B embedded system, the possibility of replicating and improving the proposed architecture on high and low-cost hardware tools is inferred. Also, it is concluded that there is space for model optimization considering feature extraction and classification parameters, such as the susceptibility to improve the data provided by the datasets to avoid over-fitting, over-learning, and increasing the hit rate in classifications.

References

[1] O. O. Abayomi-Alli, R. Damaševičius, S. Misra, R. Maskeliūnas, A. Abayomi-Alli, “Malignant skin melanoma detection using image augmentation by oversampling in nonlinear lower-dimensional embedding manifold,” Turkish Journal of Electrical Engineering and Computer Science, vol. 29, no. 8, pp. 2600-2614, 2021. https://doi.org/10.3906/elk-2101-133Links ]

[2] A. Demir, F. Yilmaz, O. Kose, “Early detection of skin cancer using deep learning architectures: Resnet-101 and inception-v3,” in Medical Technologies Congress, pp. 1-4, 2019. https://doi.org/10.1109/TIPTEKNO47231.2019.8972045Links ]

[3] J. Jaworek-Korjakowska, “Computer-aided diagnosis of micro-malignant melanoma lesions applying support vector machines,” BioMed Research International, vol. 2016, pp. 1-8, 2016. https://doi.org/10.1155/2016/4381972Links ]

[4] A. Esteva et al., “Dermatologist-level classification of skin cancer with deep neural networks,” Nature, vol. 542, no. 7639, pp. 115-118, 2017. https://doi.org/10.1038/nature21056Links ]

[5] P. Tschandl et al., “Expert-Level Diagnosis of Nonpigmented Skin Cancer by Combined Convolutional Neural Networks,” JAMA Dermatology, vol. 155, no. 1, pp. 58-65, 2019. https://doi.org/10.1001/JAMADERMATOL.2018.4378Links ]

[6] B. Harangi, “Skin lesion classification with ensembles of deep convolutional neural networks,” Journal of Biomedical Informatics, vol. 86, pp. 25-32, 2018. https://doi.org/10.1016/j.jbi.2018.08.006Links ]

[7] M. Chen, P. Zhou, D. Wu, L. Hu, M. M. Hassan, A. Alamri, “AI-Skin: Skin disease recognition based on self-learning and wide data collection through a closed-loop framework,” Information Fusion, vol. 54, pp. 1-9, 2020. https://doi.org/10.1016/j.inffus.2019.06.005Links ]

[8] H. N. Pham et al., “Lesion Segmentation and Automated Melanoma Detection using Deep Convolutional Neural Networks and XGBoost,” in Proceedings of 2019 International Conference on System Science and Engineering, 2019, pp. 142-147. https://doi.org/10.1109/ICSSE.2019.8823129Links ]

[9] A. Kumar, A. Vatsa, “Untangling Classification Methods for Melanoma Skin Cancer,” Frontiers in Big Data, vol. 5, pp. 1-11, 2022. https://doi.org/10.3389/fdata.2022.848614Links ]

[10] G. A. Kukharev, N. L. Shchegoleva, “Algorithms of Two-Dimensional Projection of Digital Images in Eigensubspace: History of Development, Implementation and Application,” Pattern Recognition and Image Analysis, vol. 28, no. 2, pp. 185-206, 2018. https://doi.org/10.1134/S1054661818020116Links ]

[11] Y. Batko, G. Melnyk, O. Pitsun, “Graphical interface of hybrid intelligent systems for biomedical imaging analysis,” in Proceedings of the 2016 IEEE 1st International Conference on Data Stream Mining and Processing, 2016, pp. 121-124. https://doi.org/10.1109/DSMP.2016.7583521Links ]

[12] A. A. Jordan, A. Pegatoquet, A. Castagnetti, J. Raybaut, P. Le Coz, “Deep Learning for Eye Blink Detection Implemented at the Edge,” IEEE Embedded Systems Letters, vol. 0663, pp. 57-60, 2020. https://doi.org/10.1109/LES.2020.3029313Links ]

[13] N. Tariq, R. A. Hamzah, T. F. Ng, S. L. Wang, H. Ibrahim, “Quality Assessment Methods to Evaluate the Performance of Edge Detection Algorithms for Digital Image: A Systematic Literature Review,” IEEE Access, vol. 9, pp. 87763-87776, 2021. https://doi.org/10.1109/ACCESS.2021.3089210Links ]

[14] S. A. A. Ahmed, B. Yanikoglu, O. Goksu, E. Aptoula, “Skin Lesion Classification with Deep CNN Ensembles,” in The 28 th IEEE Conference on Signal Processing and Communications Applications, 2020. https://doi.org/10.1109/SIU49456.2020.9302125Links ]

[15] M. Arshad et al., “A Computer-Aided Diagnosis System Using Deep Learning for Multiclass Skin Lesion Classification,” Computational Intelligence and Neuroscience, vol. 2021, e9619079. https://doi.org/10.1155/2021/9619079Links ]

[16] W. Sae-Lim, W. Wettayaprasit, P. Aiyarak, “Convolutional Neural Networks Using MobileNet for Skin Lesion Classification,” in 16 th International Joint Conference on Computer Science and Software Engineering, 2019, pp. 242-247. https://doi.org/10.1109/JCSSE.2019.8864155Links ]

[17] L. Wei, K. Ding, H. Hu, “Automatic Skin Cancer Detection in Dermoscopy Images Based on Ensemble Lightweight Deep Learning Network,” IEEE Access, vol. 8, pp. 99633-99647, 2020. https://doi.org/10.1109/ACCESS.2020.2997710Links ]

Citation: C.-V. Niño-Rondón, D.-A. Castellano-Carvajal, S.-A. Castro-Casadiego, B. Medina-Delgado, K.-C. Puerto-López, “HASCC: A Hybrid Algorithm for Skin Cancer Classification,” Revista Facultad de Ingeniería, vol. 33, no. 67, e16943, 2024. https://doi.org/10.19053/01211129.v33.n67.2024.16943

Received: December 07, 2023; Accepted: March 25, 2024; Accepted: March 30, 2024

Authors’ Contribution

Carlos-Vicente Niño-Rondón: conceptualization; investigation; methodology; writing-review and editing. Diego-Andrés Castellano-Carvajal: conceptualization; investigation; methodology; writing-review and editing. Sergio-Alexander Castro-Casadiego: conceptualization; methodology. Byron Medina-Delgado: conceptualization; investigation; methodology. Karla-Cecilia Puerto-López conceptualization; writing-original draft.

Creative Commons License This is an open-access article distributed under the terms of the Creative Commons Attribution License