Ijraset Journal For Research in Applied Science and Engineering Technology
Authors: Urmila S, Dr Rudraswamy S B
DOI Link: https://doi.org/10.22214/ijraset.2022.47232
Certificate: View Certificate
It became clear that humanity must learn to live with and adapt to any pandemic such has Covid - 19 , especially in light of the fact that the vaccines currently in development do not prevent the infection but only lessen the intensity of the symptoms. It is crucial to diagnose both pneumonia and COVID- 19 since they both have an impact on the lungs. In this study, the automatic detection of the Coronavirus disease was carried out using a data-set of X-ray pictures from patients with common bacterial pneumonia, confirmed Covid-19 disease, and normal occurrences.The current approach for detection and diagnosis of COVID-19 is the RT-PCR and rapid test as known rapid test is not so effective and RT-PCR is time-consuming and, in lots of instances, no longer less expensive as a consequence the development of new low-price rapid tests of diagnostic gear to aid medical evaluation is needed. The study’s objective is to assess the effectiveness of cutting- edge convolutional neural network designs for medical picture categorization that have been recently suggested. In particular, the Transfer Learning method was utilised. Transfer learning makes it possible to detect many problems in small collections of medical image data, frequently with outstanding outcomes. The data sets used in this investigation are a collection of 5144 X-ray images, including 460 images with verified Covid-19 illness, 3418 photos with confirmed common bacterial pneumonia, and 1266 images of healthy conditions. The information was gathered from X-ray pictures that were accessible in public medical repositories. According to the data, Deep Learning combined with X-ray imaging may be able to identify important biomarkers for the Covid-19 disease. CNN achieved the highest accuracy 99.49% and specificity, followed by VGG-16 at 67.19% and dense net at 91.94
I. INTRODUCTION
Although COVID-19 is an acutely resolving illness, it has the potential to be fatal. Due to extensive alveolar damage and developing respiratory failure, severe illness may cause death when it first manifests . Countries may benefit from the early and automatic diagnosis of Covid-19 for the timely referral of the patient to quarantine and the quick incubation of dangerous cases.In specialised hospitals, and the disease’s spread is tracked. Although the diagnosis procedure has be- come reasonably quick, the financial problems brought on by the expense concerns both states and patients, particularly in nations, regarding diagnostic examinations. The X-rays from healthy instances are now more frequently available to the public than they were in 2020, although additionally from individuals who have Covid-19. As a result, we can examine the medical photographs and recognise any patterns that could result in the condition being automatically diagnosed.
Pneumonia is a condition when the lungs’ air sacs fill with fluid, causing inflammation in the lungs as a result. This leads to other problems like chills, weariness, and fever as well as breathing difficulties, coughing, and chest pain. The common cold and the flu are both causes of pneumonia. However, bacterial infections and viral infections can also be problematic at times. This is the cause; one of the factors cited as contributing to pneumonia is COVID 19.
The distinction between COVID 19-infected pneumonia and regular pneumonia has recently been the subject of numerous investigations and experiments conducted in laboratories by medical professionals. In order to illustrate how both have some significant differences, IDSA has developed a thorough knowledge using CT scan images and other sources. Therefore, it is essential to diagnose Pneumonia and COVID.
Deep learning applications seem to have emerged through- out the last five years at the perfect time. Automatically identifying features in photographs and classifying them is a major focus of the ”Deep Learning” class of machine learning techniques. Its main applications are in tasks involving the classification of medical images and item identification. When it comes to using artificial intelligence to mine, analyse, and uncover patterns in data, deep learning and machine learning are well-established fields of research. The contributions made by those disciplines to clinical decision making and computer- aided systems are getting harder to recapture as new data become accessible.
Deep neural networks, also known as deep learning, are artificial neural networks (ANN) with several layers. Due to its capacity to manage massive volumes of data, it has emerged during the last few decades as one of the most effective tools and has seen significant literary success. In a number of applications, including pattern recognition, deeper hidden layers are already beginning to perform better than conventional methods. One of the most popular deep neu- ral networks is the convolutional neural network (CNN). Convolutional Neural Networks have made groundbreaking discoveries over the past 10 years in a variety of pattern recognition-related fields, including speech recognition and picture processing. The reduction of ANN’s parameter count is CNNs’ most advantageous feature. The most crucial premise regarding issues that CNN solves is that they shouldn’t include features that are spatially dependent. This is because larger models may be used to handle sophisticated tasks that were not conceivable with classic ANNs.In other words, we don’t need to focus on where the faces are in the photographs while using a face detection tool, for instance. The only thing that matters is finding them, regardless of where they are in the provided photographs. Obtaining abstract characteristics when input propagates toward the deeper layers is another crucial component of CNN.
II. RELATED WORK
The author investigated a number of well-known pretrained deep CNN models in a transfer learning setup. For the purpose of identifying COVID-19 from chest X-ray images.Different setups were tested using the two separate publically accessible datasets either alone or jointly. Different setups were tested using the two different publically accessible datasets either alone or together.
Limitation:The COVID-19 vs. non-COVID-19 chest X-ray classification by VGG models was unsuccessful. The COVID-19 vs. non-COVID-19 chest X-ray classification by VGG models was unsuccessful.(from reference [6])
To assess and contrast the created models, three distinct experiments are conducted in accordance with three prepro- cessing approaches. The objective is to assess the effects of data preprocessing on the outcomes and how well they can be explained. Similar to this, a comprehensive investigation of various variability problems that could potentially ruin the system and its impacts is carried out. The methodology used yields a classification accuracy of 91.5 percent, with an average recall of 87.4 percent for the weakest but most explicable experiment.
Limitation:In this approach, manual methods based on visual observation of the images were used in earlier CT studies of COVID-19 to test the infection extent evaluation.(from reference [1]).
A proposed approach for analysing chest X-ray pics has been created to come across COVID-19 for binary instructions with an accuracy of 78 percentage and validation accuracy of 80 percent, wherein the loss is more or less zero.15 percent. This method applies convolution 2D techniques to COVID-19 open supply datasets which can be available at GitHub and Kaggle
Limitation:massive dataset working with GPU taking into ac- count many more attributes to evaluate for high computational speed, performance, and effective deep learning approaches implementation.(from reference [7])
A streaming diagnosis based on a deep learning-based retrospective analysis of laboratory data in the form of chest X-rays is required in this COVID-19 pendamic situation. In this paper, a method to detect COVID-19 using deep learning to assemble medical images was proposed.
Limitation:For the datasets used to determine if a person is contagious or not, classification techniques must be ap- plied.(from reference [2])
The suggested approaches were validated using a dataset that was specifically retrieved for this study. Despite the poor quality of the chest X-ray images that are a drawback of portable equipment, the proposed approaches produced global accuracy values of 79.62 percent, 80.27 percent, and 79.86 percent, respectively. This allowed a reliable analysis of portable radiographs to support clinical decision-making. Limitation: Although the decision-making accuracy is lower, the results show that the proposed technique and the tested ap- proaches permit a robust and reliable analysis.(from reference [3])
In this article, CNNs are briefly introduced, along with recently released papers and innovative methods for creating these amazingly outstanding image recognition models. The introduction makes the assumption that you are already famil- iar with the basics of ANNs and machine learning.
Limitation:In contrast to other artificial neural network types, convolutional neural networks concentrate on a specific type of input rather than the entire problem domain.(from reference [5])
Textual clinical reports were divided into four categories in this study using conventional and ensemble machine learning techniques. These features were applied to conventional and group machine learning classifiers.
Limitation:By adding more data, models’ effectiveness can be increased. In order to determine whether men or women are more likely to contract the disease, it can be categorised based on gender. Deep learning approaches can be applied in the future, but more feature engineering is required for improved outcomes. [4])
In this study, the author present a unique Support Vector Regression approach to analyse five distinct tasks associated with novel coronaviruses. To improve classification accuracy in this work, supported vectors are also used in place of a simple regression line.
Limitation:The encouraging outcomes show its inferior superi- ority in terms of both efficiency and accuracy..(from reference [8])
III. PROPOSED METHOD
A. Data Collection
The process of gathering and coordinating information from an unlimited number of various sources is known as data collection. In order to process the data after the statistics have been collected, the information is recorded as an image (chest x-ray). A dataset in machine learning is, to put it simply, a collection of data samples that can be analysed and forecasted by a computer as a single entity. This implies that the data gathered should be uniform and understandable because machines don’t perceive data in the same manner that people do. A good dataset should also adhere to strict quality and quantity standards. The dataset should also be relevant and evenly distributed for a rapid and simple training procedure. For our work, we created three classes of chest X-ray image datasets (normal people, pneumonia patient, and COVID - 19 Patients).
It has 5144 total images, of which 460 are covid- 19, 1266 are images of people who are healthy or normal, and 3418 are images of people who have pneumonia. Covid-19 is a recent infection, thus there aren’t many pictures of it. The data set is obtained from the Kaggle website and divided into two portions: 20% for validation and 80% for training.
B. Data Preprocessing
Preprocessing is a data mining technique used to transform raw data into a format that is appropriate and useful. The actions needed for data preprocessing are:
a. Normalization: To get the values of the data in the specified range, normalisation is done.
b. Attribute Selection:The set of already provided at- tributes is used to construct entirely new attributes.
c. Discretization: Restoring the raw values of numeri- cal properties at interval levels is done..
d. Concept Hierarchy Generation: Based on hierarchy, the qualities are changed from a low level to a high level.
3. Data Reduction: Since managing large amounts of data requires a certain approach, data mining. It was chal- lenging to assess in these situations while working with a significant amount of data. So we used a data reduc- tion strategy to get away from this. Reduced expenses and increased storage efficiency were the goals of the reduction.
C. Split Data
In order to evaluate how well the machine learning algo- rithm is working, data is split into training and testing sets.
The major goal is to assess the model’s performance using the most recent data, i.e., data that were not utilised in the model’s training. When there are several data sets gathered, the train-test methodology is appropriate.
The size of the training and testing data sets is a crucial configuration constant for the course of action. For either the training or testing data sets, it is often shown as a percentage between 0 and 1. For instance, if the size of the training set is 0.60 (60 percent), the testing set will receive the remaining 0.40 (40 percent). For dividing data sets, there is no ideal split %. Taking into account the needs of our project, we choose a percentage to divide the data sets by.
3. Cost of computation for training a model
4. Computed costs for model evaluation
5. Representation of the training set’s behaviour
6. Behaviour of the testing set’s representation
We took into account the aforementioned elements when choosing the percentage to divide the data into training and testing sets for this project. We used the Python machine learning toolkit scikit-learn, which provides a training and testing split assessment application. The ”test size” option, which accepts an input of the number of rows or a percentage of the size of the data set between 0 and 1, can be used to describe the split size.
D. Algorithms
The deep learning algorithm is a method for the system of AI capabilities to carry out the activities, typically by anticipating the values as output from previously provided data as input.
The important actions of algorithms of deep learning is image classification.
E. Transfer learning with CNNs
Transfer learning is a method that trains a CNN from start to perform a different but related job using new data, often from a smaller population. Transfer learning makes advantage of the data-mining knowledge produced by a CNN.
In this deep learning process, a CNN is first trained using big datasets for a particular objective (like classification). The availability of data for the first training is the most crucial component for training to be successful since CNN may dis- cover critical properties (features) of the image. The suitability of this model for transfer learning depends on the CNN’s capacity to recognise and extract the most exceptional visual characteristics. The processing of a brand-new collection of images of a different kind and the use of the CNN to extract features based on its experience with feature extraction since its first training constitute the following phase.
F. CNN
Convolutional neural network (CNN) techniques are typi- cally applied to designs with a large number of training layers. Weights and bias are the two factors for every layer. Later layers search for mid-level and high-level features such as structures, objects, and shapes, while earlier layers concentrate on low-level features like corners, edges, and lines. Using the prediction result, it is then possible to calculate the loss cost to the ground truth labels. In order to compute each parameter’s gradient based on the loss cost, the backward stage secondly uses chain rules. For the future gradient-based forward computation, every parameter has been updated and prepared. Following several cycles in both the forward and backward stages, network learning might come to an end.
It would be advantageous to define the term since it is frequently used to refer to mathematical notions and ideas connected to the feature transformation strategy and procedure. One of the starting functions in mathematics, most notably algebraic topology, is turned into a third function by the mathematical procedure known as convolution. This third function is frequently viewed as a transformed version of the two initial functions in mathematics, u and v. The convolution is a distinct function that is frequently represented by the letters u and v and discussed by (Hirschman and Widder, 2017) in the context of two real or complex functions, u and v.
Rectified Linear Unit (ReLU) transform functions only turn on a node if the input value exceeds a specific threshold. When the information rises above a threshold, the output changes from zero while the data is below zero. It and the dependent variable are related linearly.
H. Fully connection layer
An input vector of numbers is provided to this layer. The phrase ”fully connected” describes a system in which every input is coupled with every output. In the CNN procedure, it often comes after the final pooling layer. 90% of the CNN’s parameters are located in fully linked layers, which function similarly to a traditional neural network. This layer basically receives the output of the previous pooling layer and creates an N-dimensional vector, where N is the total number of classes the computer can choose from. We can use it to feed a vector with a specific length forward into the neural network. Additionally, the result is a vector of numbers. This
simulates higher-level reasoning in which all potential routes from the input to the output are taken into account. After passing through two layers of convolution, relu, and pooling, and being converted into a single file or a vector, take the reduced image and place it in the single list.
a. Forward (input, target): based on input and target value, estimates loss value.
b. Backward (input, target): identifies the criterion, calcu- lates the gradient of the loss function associated with it, and returns the outcome.
c. In a CNN, the back propagation principle is employed to determine the gradient of the loss function, which determines the cost related to a particular state.
2. VGG - 16:
I. VGG - 16 Architecture
The term ”ConvNets” is frequently used to refer to convo- lutional neural networks, a subset of artificial neural networks. Input, output, and multiple hidden layers are the components of a convolutional neural network. One of the most effective computer vision models available right now is the CNN(Convolutional Neural Network) variant known as VGG16. In order to analyse the networks and deepen the model, the developers developed an architecture with incredibly small (3 x 3) convolution filters, which was a substantial advance over the state-of-the-art setups. Around 138 trainable parameters were created with the depth increased to 16–19 weight layers. An input image with dimensions is sent to the network (224, 224, 3). In the first and second layers, there is the same padding and 64 channels with a 3*3 filter size. Following a stride (2, 2) max pool layer, two layers have convolution layers with a 128 filter size and a filter size (3, 3). Max-pooling stride (2, 2) layer with the exact same properties as the layer before follows. Then, 256 filters are spread out over 2 convolution layers with 3 and 3 filter widths.
A max pool layer comes next, then two sets of three convolution layers. With the same spacing and 512 filters per filter (3, 3). Then, this image is subjected to the stack of two convolution layers. We employ 3*3 filters in these convolution and max-pooling layers as opposed to 7*7, ZF-11*11, and AlexNet filters. Additionally, several of the layers change the number of input channels by using 1*1 pixels. A 1-pixel padding is offered after each convolution layer to prevent the spatial characteristic of the image.
As Convolution and max-pooling layers were added to the stack, and the result was a (7, 7, 512) feature map. A feature vector with the value 1, 25088 is produced by flattening this result. There are then three fully connected layers: the first layer uses the most recent feature vector as input and outputs a vector with a size of (1, 4096); the second layer also does so; the third layer, however, also produces a vector with a size of (1, 1000), which is used to implement the softmax function to divide the data into 1000 categories. Every hidden layer has a ReLU activation function. ReLU is more efficient since it encourages faster learning and reduces the possibility of error.
J. Densenet
One of the most recent advancements in neural networks for visual object detection is called DenseNet. Despite their apparent closeness, ResNet and DenseNet have some impor- tant differences. While DenseNet uses an additive approach to combine the previous and subsequent levels, ResNet concate- nates the layers.
Composition Layer Pre-Activation For each composition layer that produces feature maps with k channels, Batch Norm (BN), ReLU, and 33 Conv are executed. For instance, X0, X1, X2, and X3 may all be altered to X4. The concept for this was generated by Pre-Activation ResNet.
L. DenseNet-BC (Further Compression)
The transition layer produces m output feature maps, with 01 being the compression factor, if there are m feature-maps in a dense block. On each transition level when =1, a fixed number of feature-maps are present. DenseNet-C, also known as DenseNet, had a value of 1 and was equal to 0.5 throughout the trial. When the bottleneck layer and the transition layer with 1 are both used, the model is referred as as DenseNet- BC. Currently, DenseNets with/without B/C, various L layers, and various k growth rates are also being trained.
M. Prediction
The result of the algorithm, which shows if a specific illness as covid-19 or pnemonia can be predicted based on the data sets, has been trained on a genuine data-set and reinforced to recent data.
N. Accuracy
The study used to show which machine learning algorithm’s model is best at identifying connections and patterns between different variables in data sets is known as the model accuracy.
O. Confusion Matrix
It is a 2x2 binary classification matrix with real values on one axis and predicted values on the other. we introduce two concepts: false positives and false negatives.
In this study, CNN, dense net, and vgg16 were enhanced to better detect COVID, viral pneumonia, and to differentiate COVID-19 cases from non-COVID-19 cases on chest X-ray pictures. The proposed model eliminates the need for human feature extraction due to its automated nature and end-to-end structure. Less cases were used to train the deep models in the vast majority of earlier research. A variety of example edged images are produced by the proposed model’s multi- image augmentation technique, which is supported by first and second order derivative edge operators. The classification accuracy of xray scans and photographs is evaluated usingCNN that has been trained using these improved images. CNN has a classification accuracy of 100% for X-Ray pictures and photos when it is trained with these improved images. The experimental outcomes were deemed to be quite convincing, and they were helpful for COVID-19 screening on chest X-ray images of individuals who potentially have a corona. Future research may focus on the early diagnosis of several illnesses, such as pneumonia, bronchitis, and tuberculosis, as well as the COVID-19 of those who are suspected of having a respiratory disorder. Our suggested multi-model ensemble detection strategy has improved compared to the original detection effect for CNN, even though the detection effect for VGG-16 is still insufficient because of over fitting. The lengthy training schedule is another problem. To improve the model’s accuracy, each of these elements should be considered. Future study will focus on developing more precise classification systems for the diagnosis of two types of pneumonia caused by bacteria and viruses. The CNN-based version is a promising way to use X-rays to identify the disease, according to the description that was previously provided.
[1] Julia´n D Arias-London˜o, Jorge A Gomez-Garcia, Laureano Moro- Vela´zquez, and Juan I Godino-Llorente. Artificial intelligence applied to chest x-ray images for the automatic detection of covid-19. a thoughtful evaluation approach. Ieee Access, 8:226811–226827, 2020. [2] Asma Channa, Nirvana Popescu, et al. Robust technique to detect covid- 19 using chest x-ray images. In 2020 International Conference on e- Health and Bioengineering (EHB), pages 1–6. IEEE, 2020. [3] Joaquim De Moura, Luc´?a Ramos Garc´?a, Pla´cido Francisco Lizancos Vidal, Milena Cruz, Laura Abelairas Lo´pez, Eva Castro Lopez, Jorge Novo, and Marcos Ortega. Deep convolutional approaches for the analysis of covid-19 using chest x-ray images from portable devices. Ieee Access, 8:195594–195607, 2020. [4] Akib Mohi Ud Din Khanday, Syed Tanzeel Rabani, Qamar Rayees Khan, Nusrat Rouf, and Masarat Mohi Ud Din. Machine learning based approaches for detecting covid-19 using clinical text data. International Journal of Information Technology, 12(3):731–739, 2020. [5] Hagyeong Lee and Jongwoo Song. Introduction to convolutional neural network using keras; an understanding from a statistician. Communica- tions for Statistical Applications and Methods, 26(6):591–610, 2019. [6] Iosif Mporas and Prasitthichai Naronglerdrit. Covid-19 identification from chest x-rays. In 2020 International conference on biomedical innovations and applications (BIA), pages 69–72. IEEE, 2020. [7] T Padma and Ch Usha Kumari. Deep learning based chest x-ray image as a diagnostic tool for covid-19. In 2020 international conference on smart electronics and communication (ICOSEC), pages 589–592. IEEE, 2020. [8] Milind Yadav, Murukessan Perumal, and M Srinivas. Analysis on novel coronavirus (covid-19) using machine learning methods. Chaos, Solitons & Fractals, 139:110050, 2020.
Copyright © 2022 Urmila S, Dr Rudraswamy S B. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Paper Id : IJRASET47232
Publish Date : 2022-10-30
ISSN : 2321-9653
Publisher Name : IJRASET
DOI Link : Click Here