Ijraset Journal For Research in Applied Science and Engineering Technology
Authors: Lohith Vattikuti, Sai Vignesh Chintala, Sudheer Kumar Dola, Vuyyala Poorna Sai, Dondapati Amarnadh, SrinivasaRao Tummalapalli
DOI Link: https://doi.org/10.22214/ijraset.2023.55896
Certificate: View Certificate
Melanoma is caused by abnormal growth of skin cells which is most often devloped on skin exposed to UV rays. It is less common but one of the most dangerous diseases in the world. Classifying skin lesions correctly at an early stage could avoid some types of skin cancers like melanoma and increases the chances of cure before cancer spreads. One of the efficient methods to accurately identify Melanoma is using deep learning. Deep learning methods such as AlexNet, Inception V3 were proposed in the past. In this project, other methods like MobileNet, CNN, ensemble method, ReNet, VGG16 will be used to detect the type of cancer using dermoscopic image data. Layers of these models are applied for fine-tuning which allows us to differentiate between different classes of skin lesions. There are two kinds of Melanoma, malignant and benign and we will also perform grading to early diagnose the disease which will ultimately increase the survival rate.
I. INTRODUCTION
Unchecked development of skin pigmented cells is one of the characteristics of skin cancer melanoma. It is one of the most lethal types of skin cancer and, if not found and treated quickly, can cause severe health issues. Using deep learning models is one potential method for melanoma detection and classification. Artificial neural networks are used in deep learning, a type of machine learning, to evaluate and learn from large datasets. Deep learning models can be taught to correctly identify and classify various types of melanomas by training them on large datasets of medical images of melanoma. Deep learning models can also be used to assess the severity of melanoma, which can assist medical professionals in making better treatment choices. The classification of melanoma skin diseases has been an active topic of research, with numerous studies suggesting various strategies to increase the precision and understandability of the classification models.
Deep learning model development for melanoma detection and classification is still an active research field, though. The need for big and diverse datasets. the interpretability of the models, and validation in actual clinical settings are just a few of the unresolved issues. Deep learning models have the potential to increase the precision and effectiveness of melanoma identification and classification in spite of these difficulties. If these models are successful, they could have a significant influence on melanoma early detection and treatment, improving patient outcomes. Convolutional neural networks (CNNs), for example, are a deep learning technology that may be used to automatically extract features from input photos. The performance of CNNs in melanoma classification tasks has been demonstrated to be very high, with many research obtaining accuracy rates above 90%.
II. LITERATURE REVIEW
The categorization of melanoma skin diseases has become a prominent research area in recent years, with numerous publications offering various techniques to increase the precision of classification models.
Scalability, computational power, and storage capacity are just a few benefits that cloud computing can provide for melanoma detection. Researchers have effectively processed massive volumes of data and trained intricate deep learning models by using cloud technology. Real-time analysis, remote access, and collaboration among healthcare experts are also made possible by cloud-based technologies[1]. The Dataset for ISIC 2019 which consisting of 25331 dermatoscopy pictures of skin lesions, was used by the authors to train and assess their models. There are two stages to the suggested hybrid strategy. In the first step, features from the input photos are obtained by employing a deep CNN model. In the second stage, the collected features are used to categorize skin lesions into benign or malignant using a traditional machine 4 earning method like support vector machines (SVMs).
The suggested method outperforms the current approaches with area under the receiver operating characteristic curve, sensitivity, specificity, and accuracy of 93.78%, 89.27%, and 0.975, respectively [2].
Three most common types of skin cancer: melanoma, squamous cell carcinoma, and basal cell carcinoma the scientists trained a deep convolutional neural network (CNN) model. The International Skin Imaging Collaboration (ISIC) and the University of Heidelberg, among other clinics and sources, provided the dataset. The trained model's effectiveness in identifying skin cancer was compared to a panel of 21 board-certified dermatologists using a separate dataset of 514 dermatoscopic images of skin lesions. In contrast to dermatologists, the area under the receiver operating characteristic curve for the CNN model was 0.91 [3].
The ISIC 2017 dataset was used by the authors to train and test a CNN model, and they used approaches for data augmentation to increase the training dataset. Using metrics for accuracy, sensitivity, specificity, and AUC-ROC, they assessed the model's performance and contrasted it with other deep learning models. When compared to other machine learning techniques, the suggested CNN model performed better at melanoma detection and attained high accuracy and sensitivity rates. The model achieved 88%, 84%, and 89%, respectively, for accuracy, sensitivity, specificity, and AUC- ROC values. The authors also contrasted their CNN model with cutting-edge deep learning models like ResNet50, VGG16, and Inception V3 [4]. The dataset was subsequently processed using both conventional image processing methods, such as color-based segmentation and feature extraction, and modern methods, such as deep learning algorithms. The findings demonstrated that when compared to conventional methods, deep learning algorithms had a greater level of melanoma detection accuracy. The maximum accuracy, 94%, was attained by the deep learning method utilizing the Inception V3 model [10].
Deep learning methods including explainable deep learning. transfer learning, and CNNs have showed considerable potential in this field. A paper published by using a hybrid approach-2020 used technique to detect hair lines from the dermoscopic images is done based on the 2-D derivatives of Gaussian (DOG) of the blue component of the images. Otsu method is used for separating hair lines. For the lesion segmentation, we used morphological snakes. After extracting the texture features SVM, KNN and CNN are used. Compound coefficient method is used to enhance the ability to capture richer and much more complex features for melanoma recognition and transfer learning h3 ped the model to find better model weights for inference. EfficientNet on ImageNet to a new skin lesion image has been used for classification domain.
III. SYSTEM DESIGN AND IMPLEMENTATION
A. Data Set
The SIIM-ISIC Melanoma Classification dataset is a sizable, openly accessible dataset of skin lesion images that is frequently used in studies for the classification of melanoma skin diseases. It includes both benign and malignant skin lesions in its 33,126 photos. The dataset, which is accessible 12 for download on the ISIC website, was assembled by the International Skin Imaging Collaboration (ISIC) and 4he Society for Imaging Informatics in Medicine (SIIM).The images in the dataset were obtained using a variety of tools and imaging modalities, and they range in size and quality. The collection also contains meta-data that can be utilised to further analyse the photos, such as patient age, sex, and lesion location. With 22,934 photos in the 15 training set and 10,192 images in the validation set, the dataset is divided into training and validation sets. Out of these we have taken 1440 for testing and 8164 for testing.
B. Methods
To create a reliable classification system for melanoma skin diseases, various procedures must be taken. First and foremost, gathering reliable data is crucial to the system's performance. To prepare the data for the model training process, it must be resized, normalised, and enhanced with additional data. The best model for classifying skin lesions has proven to be CNNs, and once chosen, the model must be trained using the preprocessed data. To make sure the model is accurate, its performance on a different set of data must be assessed. The model can be used to categorise skin lesions in a web-based application, once it has been reviewed and found to be satisfactory. By retraining the model on new data, performance may be continuously improved.
After building few models such as VGG16, Inception, Mobilenett, ResNetV2 and comparing it with Ensemble methods such as VGG16 and Mobilenett, VGG16, inception and Mobilenett. It was observed that mobilenett have achieved the greatest accuracy i.e., 92% among all the models built. Due to its minimal processing demands, MobileNet is a well- known CNN architecture for deep learning models. The initial stage in using MobileNet for the classification of melanoma skin diseases is to gather a high-quality dataset and preprocess it by scaling the photos to a particular size, normalising the pixel values, and carrying out data augmentation to enhance the dataset size. Next, the preprocessed data must be used to train the MobileNet model using methods like transfer learning or fine-tuning. Transfer learning entails utilising the MobileNet model's pre-trained weights on a sizable dataset like ImageNet before retraining the model's final few layers using the melanoma skin disease dataset. The entire model must be fine-tuned, and it must be retrained.
Our proposed method consists of the steps starting with data pre-processing followed by data augmentation then the model has been trained the data with different deep learning and ensemble models by compared the accuracies and by providing the testing images to the models in which the model will be predicting whether the predicted image as malignant or benign. Then the model which gave the highest accuracy i.e., Mobile net has been chosen as a base model to perform grading. By using the base model, the stage of cancer is predicted i.e., Stage 1 to 3.
The architectural implementation of our model is given below:
A convolutional neural network with a basic MobileNet architecture, also called MobileNetV1, is made up of a number of convolutional layers with depth wise separable convolutions. The fundamental design of MobileNet is as follows:
In Input layer 224x224 RGB images are commonly used as the input image for this layer and in the convolutional layer the input image is processed by 13 convolutional layers, each of which has a different number of filters and kernel sizes. The input tensor is halved in size in the first layer, which has a stride of 2. A stride of 1 is used in the other convolutional layers. D 15h wise separable convolutions: MobileNetV1 employs depth wise separable convolutions, which are made up of a pointwise convolution and a depth wise convolution. The pointwise convolution combines the result of the depth wise convolution using 1x1 convolutions, whereas the depth wise convolution applies a single filter to each input channel independently. Batch Normalizations: MobileNet employs batch normalization after each convolutional layer, which helps to normalize the layer's output and enhances the network's accuracy. Activation function: A ReLU activation function, which introduces non-linearity into the network and aids in the capture of more complex features in the input data, is applied to the out 14 of each batch normalization layer and in poolinng layer a global average pooling layer, which calculates the average value of each feature map in the output tensor, is applied after the last convolutional layer's output. Fully connected layer: The network's ultimate output is created by flattening and passing the output of the pooling layer via a fully connected lay.
The MobileNet ModelNet parameters used are same as VGG16 and that of activation function I.e., Relu, the output activation functions were taken as softmax and sigmoid. All the top layers were trained with imagenet dataset, and the last layers are trained with our dataset. While training the model it is trained using adgard optimizer at a learning rate of 0.01. We took 10 epochs to train the model and defined all the parameters which are to be included for training the images. As a result, the validation loss was improved from epoch 1 to 7, for the 8th and 9th epoch the validation loss did not improve so on the 9th epoch learning rate was reduced. After that the validation loss did not improve for the 9th and 10th epoch and finally, we got an accuracy of 0.916 and a loss of 0.1987. Using the MobileNet model we have predicted whether the skin lesions images are malignant or benign. Out of 25 images only one image was incorrectly predicted.
Given that the validation loss is similarly falling and is not considerably larger than the training loss, this suggests that the model is not overfitting the training set. This may indicate that the model generalizes effectively to fresh, untested data.
Last but not least, while training goes on, the loss curve gradually flattens out at a low value, demonstrating that the model has converged and improved in performance. This is a good result since it indicates that the model is successfully identifying the underlying patterns in the data. We see a consistent rise in accuracy over the training epochs in the accuracy curve that is supplied. This shows that as the model gains knowledge from the training data, it performs better and generates predictions that are more accurate.The accuracy curve initially begins at a low value and steadily increases. This implies that the model could initially have trouble making accurate predictions but eventually improves as training goes on. The increasing trend shows how the model can identify underlying patterns in the data over time and produce more accurate predictions.
The accuracy curve achieves a high value plateau as the training epochs go on. This suggests that the model has stabilized its performance and is operating at its highest level of accuracy. The plateau means that the model has made good progress in generalization and is reliably producing precise predictions on both the training and validation data.
Melanoma is a form of skin cancer that, if not found and treated quickly, can be incurable. Dermatologists employ a variety of methods, such as clinical examination, ceroscopy, and histology, to recognize and classify skin lesions. The models generated in this project have produced encouraging results, with good accuracy in identifying benign and malignant skin lesions as well as in identifying the stage of the disease by using MobileNett. To create deep learning models for the classification of melanoma, numerous studies have been undertaken. These models used a variety of clinical, dermoscopic, and histological features that are derived from skin lesions. By identifying and categorizing skin lesions as benign or malignant using image analysis techniques, deep learning algorithms can compete in overcoming these constraints. To understand the visual patterns and traits that distinguish between benign and malignant lesions, these algorithms are trained on massive datasets of skin lesion photos with dermatologist annotations The future scope for deep learning in melanoma classification is promising, and there are several potential directions for further research and development like, incorporating more representative and diverse datasets can help deep learning models perform better and be more generalizable.
[1] L. D. Biasi, A. A. Citarella, M. Risi and G. Tortora, \"A Cloud Approach for Melanoma Detection Based on Deep Learning Networks,\" in IEEE Journal of Biomedical and Health Informatics, vol. 26, no. 3, pp. 962-972, March 2022, doi: 10.1109/JBHI.2021.3113609. [2] J. Daghrir, L. Tlig, M. Bouchouicha and M. Sayadi, \"Melanoma skin cancer detection using deep learning and classical machine learning techniques: A hybrid approach,\" 2020 5th Intemational Conference on Advanced Technologies for Signal and Image Processing (ATSIP), Sousse,Tunisia,2020, pp.1-5, doi:10.1109/ATSIP49331.2020.9231544. [3] Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, Thrun S. \"Dermatologist-level classification of skin cancer with deep neural networks\". Nature. 2017 Feb 2;542(7639):115-118. doi: 10.1038/nature21056. Epub 2017 Jan 25. Erratum in: Nature. 2017 Jun 28;546(7660):686. PMID: 28117445; PMCID: PMC8382232. [4] R. Zhang, \"Melanoma Detection Using Convolutional Neural Network,\" 2021 IEEE International Conference on Consumer Electronics and Computer Engineering (ICCECE), Guangzhou, China, 2021, pp. 75-78, doi: 10.1109/ICCECE51280.2021.9342142. [5] Xie, Peizhen and Zuo, Ke and Zhang, Yu and Li, Fangfang and Yin, Mingzhu and Lu, Kai, \"Interpretable Classification from Skin Cancer Histology Slides Using Deep Learning: A Retrospective Multicenter Study\", 2019 [6] Peizhen Xie, Ke Zuo, Jie Liu, Mingliang Chen, Shuang Zhao, Wenjie Kang, Fangfang Li, \"Interoretable Diagnosis for Whole-Slide Melanoma Histology Images Using Convolutional Neural Network\", Journal of Healthcare Engineering, vol. 2021, Article ID 8396438, 2021. [7] L. B. Maia, A. Lima, R. M. Pinheiro Pereira, G. B. Junior, J. Dallyson Sousa de Almeida and A. C. de Paiva, \"Evaluation of Melanoma Diagnosis using Deep Features,\" 2018 25th International Conference on Systems, Signals and Image Processing (IWSSIP), Maribor, Slovenia, 2018, pp. 1- 4, doi: 10.1109/IWSSIP.2018.8439373. [8] Chendage, Bapu & Mente, Rajivkumar & Pawar, Sunil. (2021). Detection and Classification of Melanoma Skin Cancer Analysis. International Journal of Scientific Research in Computer Science Engineering and Information Technology. 7. 150-154. 10.32628/CSEIT217130. [9] Gerges, Firas & Shih, Frank. (2021). A Convolutional Deep Neural Network Approach for Skin Cancer Detection Using Skin Lesion Images. 15. 475-478. [10] B. Sreedhar, M. Swamy B.E and M. S. Kumar, \"A Comparative Study of Melanoma Skin Cancer Detection in Traditional and Current Image Processing Techniques,\" 2020 Fourth International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud) (I-SMAC), Palladam, India, 2020, pp. 654-658, doi: 10.1109/I-SMAC49090.2020.9243501. [11] Phillips M, Greenhalgh J, Marsden H, Palamaras I. of Malignant Melanoma Using Artificial Detection of Intelligence: An Observational Study of Diagnostic Accuracy. Dermatol Pract Concept. 2019 Dec 31;10(1):e2020011. doi: 10.5826/dpc.1001a11. PMID: 31921498; PMCID: PMC6936633.
Copyright © 2023 Lohith Vattikuti, Sai Vignesh Chintala, Sudheer Kumar Dola, Vuyyala Poorna Sai, Dondapati Amarnadh, SrinivasaRao Tummalapalli. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Paper Id : IJRASET55896
Publish Date : 2023-09-26
ISSN : 2321-9653
Publisher Name : IJRASET
DOI Link : Click Here