Ijraset Journal For Research in Applied Science and Engineering Technology
Authors: Aashish Nayak, Dr. Dharamveer Singh, Vikas Gupta
DOI Link: https://doi.org/10.22214/ijraset.2022.45227
Certificate: View Certificate
Glaucoma is a disease of the retina caused by high intraocular pressure. The intraocular pressure in people with glaucoma can reach 60-70 mm Hg. This disease is characterized by an increasing cup to disc ratio size. Glaucoma has three levels, namely mild with a cup to disc ratio value of 0.3-0.5, moderate with a cup to disc ratio value of 0.5-0.7 and severe with a cup to disc ratio value above 0.7. For retinal analysis and calculating the cup to disc ratio value taken from a fundus camera, it must be done by an expert ophthalmologist, but it takes a long time. Therefore, feature detection and automatic cup to disc ratio value calculation are expected to assist doctors in analyzing glaucoma. The data used were 132 retinal fundus images consisting of 66 mild glaucoma images, 26 moderate glaucoma images and 40 severe glaucoma images taken from the RIM-ONE dataset (http://medimrg.webs.ull.es). Pre-processing techniques like cropping, resizing, brightness, Median Filter are used for noise removal. Subsequently, feature extraction with the help of GLCM. Consequently, the method used to classify the degree of glaucoma is the Deep Belief Network. The test simulation results obtained accuracy value of 99% with 99% of precision and 100% of recall.
I. INTRODUCTION
A. Background
The development of vision systems has significantly simplified tasks in different areas of science in which activities such as machine learning, evolutionary computing, image processing, among others, are developed. In recent years, numerous vision systems have been developed that help solve problems in multiple real-life applications. Computer vision and image processing have a lot in common, the purpose of each being different. Image processing seeks to improve the quality of an image, such as eliminating defects, correcting blur, improving the quality of colors, among others. In which there is an input image and the improved image as an output. In the vision system the objective is to obtain the description and interpretation by computers, having an image as input, which will be processed to extract attributes and as output a description of the image. In the present work, a comparison and analysis of the performance of extraction and classification of characteristics in digital images that are obtained from patients with glaucoma and diabetic retinopathy are made, from which the most significant qualities of these are obtained, which are textural and chromatic characteristics. Glaucoma and diabetic retinopathy are eye diseases that affect vision causing total loss. For the first condition, the aqueous humor contained in the eye does not drain properly due to a failure of the ocular drainage system causing a higher intraocular pressure, this causes the nerve fibers that meet in the optic disc to be damaged and may die, preventing signals from reaching the optic nerve. Glaucoma is the second largest cause of blindness in the world (Bulletin of the WHO) and an estimated 80 million people will develop glaucoma by 2021 [1]. For retinopathy produced in diabetic patients, high sugar levels cause the blood vessels that feed the retina to die, the eye tries to create new of these structures to continue with the correct blood supply but these new vessels are weak which causes let them fracture and bleed. In the early stages of this condition, weak blood vessels dilate, generating lumps that protrude from the walls. The more blood vessels are damaged, the more serious it will be, because it causes the nerve fibers and the macula to become impawned. At this stage of the condition there is still treatment. The problem is aggravated, as already mentioned, with the death of the blood vessels, causing the detachment of the retina. Retinopathy can lead to glaucoma, as scar tissue affects the drainage of watery fluid. Automatic vision systems help efficient and effective detection without the intervention of a professional, speeding up the results. For them the importance of the best selection of textural and chromatic characteristics [2]. Glaucoma has been known for a long time, but not many people know about the dangers of this disease. If it is too late or not treated properly, glaucoma can cause permanent blindness in sufferers. Lack of awareness of the dangers of glaucoma is due to the symptoms of this disease that the glaucoma sufferer cannot feel directly [3]. Research to detect glaucoma has been carried out by several previous researchers, including research conducted by [4] who developed a glaucoma disease identification system through fundus images using the Backpropagation method which has a system accuracy rate of 98.45% [5] developed a system to detect glaucoma by combining Region of Interest (ROI) segmentation and automated techniques system by using hemorrhage detection in certain areas of the fundus image with an regression rate of 86.17.57%. Another research was conducted by [6] who developed an automatic glaucoma detection system that was identified by calculating the cup to disc ratio (CDR). They have proven that if the CDR value is between 0.0-0.3 then the input image is normal. Meanwhile, if the CDR value obtained is greater than 0.3, the image is identified as glaucoma. This system has an F-score of 96%. Subsequent research conducted by [7] who developed a system for detecting glaucoma through optical disc and cup segmentation using K-mean clustering. The accuracy of the system they developed reached 92%. Based on the background and several studies using fundus images to identify glaucoma, the authors submitted a research proposal entitled "Identification of Glaucoma through Fundus Images Using a Deep Belief Network". The American Academy of Ophthalmology (2018) divides glaucoma into 3 types, namely open-angle glaucoma, closed-angle glaucoma, and glaucoma in children (childhood glaucoma). Open-angle glaucoma is further divided into primary open-angle glaucoma, normal-tension glaucoma, juvenile open-angle glaucoma, suspected glaucoma (glaucoma suspect), and secondary open-angle glaucoma. Angle closure glaucoma is also subdivided into primary angle - closure glaucoma with relative pupillary block, acute angle closure glaucoma, subacute angle closure glaucoma, chronic angle closure glaucoma, secondary closed angle glaucoma with and without pupillary block, and iris plateau syndrome. Primary open-angle glaucoma, the most common form of black and white races, causes an asymptomatic, progressive, bilateral visual field narrowing that arises slowly and often goes undetected until a broad narrowing of the field of view occurs. It is estimated that the prevalence of primary open-angle glaucoma in the United States in individuals over 40 years of age is 1.86% based on a population meta-analysis study (American Academy of Ophthalmology, 2018) [1-3]. Image is a representation (picture), similarity, or imitation of an object. Image is divided into 2, there is an image that is analog and there is an image that is digital. Analog image is an image that is continuous, such as images on television monitors, X-ray photos, and CT scan results. Meanwhile, digital images are images that can be processed by computers [8]. Image processing is a form of processing an image or image with a numerical process of the image, in this case the individual pixels or points of the image are processed. Digital image processing techniques help manipulate digital images using computers. The purpose of image processing is to improve image quality. Where the resulting image can display information clearly and extract feature information from the image. Some of the techniques used in image processing are as follows[9-11]. Brightness [12-14] is a process for adjusting the brightness of an image. If the pixel intensity is reduced by a certain value, the image will be darker and vice versa, if the pixel intensity is increased by a certain value, the image will be lighter. Brightness can be done with equation 2.1.
Median Filter [15,16] is used to remove salt & pepper type noise that is often found in retinal images. Noise has a fixed value of 0 (pepper noise) and 255 (salt noise). This noise filtering technique is divided into two, namely linear and non-linear. The most widely used non-linear filter is the Median Filter which uses the median value to replace the damaged pixels and this filter is able to remove noise without removing the edges (Shrestha, 2014). This median filter technique focuses on the median or middle value of the total number of total pixel values ??contained in the image. The equation used to get the median value of an image can be seen in the equation:-
The grayscale technique is used to convert an RGB color image into a gray image. A gray image (grayscale) is an image where each pixel contains information on the intensity value of white and black. The lowest or no intensity value in a grayscale image is represented in black and the highest value represents white. All values ??between these highs and lows represent shades of gray. The number of these shadows depends on the possible values ??a pixel can have. if the pixel is expressed in bits, it can store two values, 0 and 1 and thus a black and white image, no grayscale. If pixels are expressed in bytes, there will be 256 gray levels starting with 0 as black and ending with 255 as white [16-23]. The results obtained from the grayscale process will make it easier for the image to be processed further in the thresholding process. The equation used to convert an RGB color image to a grayscale image can be seen in equation 2.6:
The thresholding procedure is used to adjust the degree of gray in the image which will produce a binary image [16-23]. The image has two gray levels, namely black and white. The process of determining the image color level at the thresholding is done by obtaining a threshold value. The thresholding process will change the pixel color to black if the image intensity value is less than the threshold value and if the image intensity value is more than the threshold value, the pixel color will turn white. The equation used to calculate the threshold value can be seen in equation 2.7:
rtificial1intelligence has actually1started since1the season1heat in11956. at1that time1a group1of computer1experts, experts1and researchers1from other1disciplines from1various academies,1industry as1well as1various circles1converged on1Dartmouth College1to discuss1the potential1of from1the laboratory1to real1work implementation[24-27].Atfirstartificialintelligence1only existed1inuniversitiesand research1laboratories, and1very few1- if1any practical1product that1has beendeveloped. Towards1the end1of the1year 1970s1and early11980s, began1to be1developed gradually1full computers1in order1imitate or1simulate human1intelligence. Some1the scientists1involved were1Allen Newel,1Herbert Simon,1Marvin Miskey,1Oliver Selfridge,1and John1McCarthy. Since1at that1time, experts1began to1work hard1to make,discuss, change1and develop1until reaching1full point1of progress.1Starting and1the results1are gradually1starting to1be marketed.1Currently, many1research results1are being1and have1been 11converted into1tangible products1that bring1profits for1the wearer.1 Artificial1intelligence (AI),1the definition1according to some1experts:
Artificial intelligence is a technology that simulates human intelligence. Computers can quickly solve a problem by imitating how humans solve the problem. One branch of AI is focused on developing algorithms for analyzing the content of an image. Computer vision tries to imitate the workings of the human visual system. Computer vision aims to allow computers to recognize an image and make decisions. The rapid development of the internet has led to the development of an extraordinary volume of textural information. Basically, the world community is very dependent on the existence of information from various fields, be it information in the form of text, images, videos or information from a web page [28-31].
Artificial Neural Network is a logical model that works based on the human brain. The workings of the brain using a number of simple neurons and interconnected with a weight value that transmits the signal from one neuron to another neuron can be modeled by an artificial neural network. Input will be received by each neuron through its connection. An output that corresponds to the weighted value of the relationship will be generated by the neuron, then the output will be forwarded back to the other neurons. Each neuron in an artificial neural network consists of several layers or layers. An artificial neural network generally consists of three layers, namely: the input layer, namely the nodes that receive the input signal, the middle layer which is also known as the hidden layer, which is the node that connects the nodes in the input layer with the nodes in the output layer and the output layer, namely nodes- the node that generates the output signal. Artificial neural networks learn by adjusting the weight values ??used to transmit values ??from one neuron to another [32-36]. Neural networks are one of the most advanced programming methods ever invented. In the conventional approach to programming, the computer is told what to do in solving a big problem by dividing it into many small precisely defined tasks so that problem solving can be done easily by the computer. On the other hand, in an artificial neural network, the computer is not told how to know which problem to solve. Instead, it learns from observational data then looks for solutions to solve problems. However, until 2006 researchers did not know how to train neural networks to be able to go beyond the computer problem solving approach in a more traditional way, except for a few specific problems [37-41]. Deep1Belief Network (DBN) is1one of1the most1popular deep1learning models1consisting of1a number1of Restricted1Boltzmann Machine (RBM) layers1and a1Backpropagation (BP) layer. RBM is a1generative model1that can1collect structural1information in1data and1can effectively1train non-linear1data through1unsupervised training.1Each RBM1layer extracts1input data1in a1bottom-up manner1and the1output information1from the1last RBM1layer is1used as1input data1from the1BP neural1network.TheRBMtrainingprocess makesit suitable1as a1DBN module.1The objective1of DBN1is to1extract and1separate bottom-up1input data1for each1RBM layer1and store1all important1information. The1multi-layer RBM1layer uses1the unsupervised1learning method,while the1Backpropagation Neural1Network (BNN)1uses the1supervised learning1method. Each layer of the RBM extracts input1data bottom-up1and the1output information1from the1last layer1of the1RBM network1is used1as input1data at1the BNN.1Because the1training carried1out by1each RBM1layer can1only make1the parameters1at that1layer achieve1optimization, we1use BNN1for top-down1tuning of1the entire1model. Meanwhile, the information1obtained from1the optimization1of the1RBM network1is used1as the1BNN data1input which1solves the1problem of1the BNN,1namely that1it is1easy to1fall into1local minimums1and has1slow1convergence [42-47].
The ReLU (Rectified Linear Unit - see figure 2.19 function is the most used (at the time of writing this work) and is used in almost all types of networks convolutional as well as in other types of deep learning networks. It is derivable and it is monotonic (its derivative is also monotonic). A disadvantage of the ReLU function is that the weights could be updated in a certain way in which the neuron never fires, that is, the output of that neuron is always zero and when that condition is reached during training. However, it has important advantages compared to the Sigmoid or Tangential Hyperbolic function, such as the acceleration in the convergence of the network (by means of the stochastic descending gradient) because it does not involve costly calculations in training [42-47]. Furthermore, it does not suffer from the phenomenon of gradient disappearance. This function is a variation of the ReLU function, where the values ofX less than 0 are mapped to a function and y = aX (with to equal to a small constant and greater than 0), instead of y = 0 as we saw previously (see equation 2.27) [42-47].
The data used were 50 data. From the combination of feature extraction used by the author, the system accuracy rate for glaucoma classification results reaches 86% [51]. The study by Golob, Anna & Takahashi, Traci & Johnson, Kay [53] in the year 2020 which made the system detect the location of a tumor or breast cancer obtained an accuracy rate of 91%. The data used for this detection process uses a mammogram image with a combined data processing method of MLP algorithms and multilevel threshold. Medical images were obtained from the segmentation results, namely the gray image which was used as input in the extraction process using CNN and RGB images used in the process of determining the location of a tumor or cancer.
Another D. Varshni, K. Thakral, L. Agarwal, R. Nijhawan and A. Mittal [54] in the year 2019 methodology used by the authors works under an architecture of densely connected neural network or known as DenseNet-169. The authors determined that the ideal feature extraction architecture to support loss of gradient is DenseNet-169, as your pre-trained network consists of 169 layers of equal sizes directly to each other. In the last phase, the classification, different techniques were used such as Support Vector Machine (SVM), Random Forest, Naive Bayes, K- nearest, among others. Research determined that the combination of the DenseNet-169 architecture as a feature extractor and SVM as the classifier. Research by Torres, C. and López, C., [55] in the year 2019 proposed a model for the classification of cases of schizophrenia through electroencephalography (EEG) signal analysis with the use of deep learning methods. Using other studies carried out for the same type of disease using trait classifiers such as Random Forest, they realized that the information obtained from the characteristics of EGG signals are of high dimensionality, variability and multichannel, for which they proposed to apply the technique of Pearson's correlation coefficient (PCC), thereby reducing this type of characteristics and bring them to a heat matrix that could serve as input for a neural network convolutional. Finally, the results that his study shows are based on metrics of precision, specificity and sensitivity. The ultimate goal is to establish a prediction model with techniques such as PCC for precision in the diagnosis and classification of different mental illnesses that use EGG signals.
II. PROPOSED METHODOLOGY
This chapter will discuss the analysis and design of glaucoma identification applications. The analysis stage discusses the steps taken to identify glaucoma starting from the data analysis stage, the prepocessing stage to the identification stage using the Deep Belief Network. At the next design stage, the system interface display design will be carried out.
A. General Architecture
This section will discuss the steps involved in developing a glaucoma identification application. The steps taken are as follows: the stage of collecting fundus image data consisting of normal images and glaucoma images which will be used as training data and test data; the prepocessing stage consisting of brightness, median filter, grayscale, and thresholding; the image feature extraction stage uses the Gray Level Co-occurrence Matrix (GLCM); and the image classification stage using the Deep Belief Network. After these steps are carried out, the application can produce output in the form of glaucoma identification results. The stages can be seen in the form of a general architecture in Figure 2.1. Figure 3.1 is an overview of the general architecture of the research classification using deep learning with an adaptive learning rate. The initial process is to collect data related to the data to be examined, then normalized and transformed the dataso as to produce data that is ready to be tested. Furthermore, the data is divided into two namely data for training and data for testing. Classification process using a standard deep neural network on the one hand, and on the other using a deep neural network with an adaptive learning rate. The training process the deep neural network has 2 training phases, the first phase is unsupervised learning deep belief network, and the second phase of supervised.learning fine-tuning deep neural network. The final result of the classification will be analyzed for each of its performance.
B. Preprocessing Data
The data preprocessing stage is the stage of data selection that aims to get data that is suitable for use. The value is often found in raw data missing value, not stored value (misrecording), data sampling which is not good enough and whatnot. Therefore it takes steps like discard some records that will not be analyzed or that do not becomedeterminants of data classification and transformation.
Data transformation was carried out to obtain the overall value for each attributewith a scale of 0 to 1 so that the analysis process is easier to do. The value transformation in the data uses the following equation:
The data used in this study are fundus images consisting of normal images and glaucoma images. This data was obtained from the RIM-ONE database (http://medimrg.webs.ull.es). RIM-ONE is a retinal fundus image database developed by 3 hospitals, namely Hospital Universitario de Canarias, Hospital Clinico San Carlos, and Hospital Universitario Miguel Servet. The types of images used in this study are normal fundus images and glaucoma fundus images. In Figure 3.2 part (a) is a normal fundus image and in part (b) is a glaucoma fundus image.
In the pre-processing stage, several processes are carried out to produce an image that is it is better to process it at a later stage. The pre-pocessing stage is brightness, median filter, grayscale, and thresholding.
Image brightness is the step to adjust the brightness of an image. This stage done to increase the brightness of the image so that the optical disc looks even more it is clear where this stage is carried out so that the image can be more easily processed at the stage next. It can be seen in Figure 3.3 part (a) is an original image that has not been processed and part (b) is the result of the brightness enhancement process carried out on original image.
2. Median Filter
This stage is done to remove salt & pepper type noise that is often found in the image. Median filter works by processing every pixel in the image then replacing each pixel value with the nearest pixel median value. The pattern of the median value of the surrounding pixels is called a window. Figure 3.4 is the image resulting from the median filter.
C. Feature Extraction
After going through the preprocessing stage, the next step is the extraction stage feature (feature extraction). The feature extraction method used in this study is the Gray Level Co-occurrence Matrix (GLCM). Here are the steps feature extraction using GLCM.
The co-occurrence matrix is ??calculated based on 1 distance and 4 directions, then the value of 6 features calculated for each co-occurrence matrix that has been normalized, so that there will be 24 features produced. An example of a feature value from a retinal image in Figure 2.6. shown in Table 1
NO. |
QVALUE |
DIRECTION |
FEATURE |
1 |
QCONTRASTQ |
10O |
13.176131 |
2 |
QCONTRASTQ |
145 O |
16.654541 |
3 |
QCONTRASTQ |
190 O |
14.047411 |
4 |
QCONTRASTQ |
1135 O |
16.409021 |
5 |
QHOMOGENEITYQ |
10 O |
10.565751 |
6 |
QHOMOGENEITYQ |
145 O |
10.452881 |
7 |
QHOMOGENEITYQ |
190 O |
10.538271 |
8 |
QHOMOGENEITYQ |
1135 O |
10.456371 |
9 |
QENERGYQ |
10 O |
10.006421 |
10 |
QENERGYQ |
145 O |
10.004521 |
11 |
QENERGYQ |
190 O |
10.005841 |
12 |
QENERGYQ |
1135 O |
10.004561 |
13 |
QENTROPYQ |
10 O |
15.628521 |
14 |
QENTROPYQ |
145 O |
15.98941 |
15 |
QENTROPYQ |
190 O |
1574,6351 |
16 |
QENTROPY |
1135 O |
15.970091 |
17 |
QVARIANCEQ |
10 O |
1145.39951 |
18 |
QVARIANCEQ |
145 O |
1144.08761 |
19 |
QVARIANCEQ |
190 O |
1145.55771 |
20 |
QVARIANCEQ |
1135 O |
1144.7491 |
21 |
QCORRELATIONQ |
10 O |
10.989081 |
22 |
QCORRELATIONQ |
145 O |
10.976911 |
23 |
QCORRELATIONQ |
190 O |
10.98611 |
24 |
QCORRELATIONQ |
1135 O |
10.977861 |
Table 1 GLCM Features
D. Proposed Algorithm
The next step after getting the GLCM feature value is image identification via Deep Belief Network (DBN) method. The preliminary stage carried out in the image identification procedure is the training dataset training process. In this case study, researcherscast-off 80 input data to be trained. The Restricted Boltzmann Machine (RBM) algorithm is also used during the training process.
The parameters used in the DBN method are:
Before data processing, it is necessary to determine the network architecture, such as determining the number of layers used, the number of neurons in each layer used, the activation function used, and other parameter values.
a. Input layer Based on the attributes to be used, there are contrast, homogeneity, energy, variance and correlations like attributes that will be used as neuron input.
b. Hidden layer / dense layer In a neural network, it is generally not more than 2 layers, but in deep learning it is possible to have more than two hidden layers. However two layers can also be used for simple datasets. There are three rules in choosing the number of neurons in each hidden layer
1. The number of neurons between the input and output sizes of neurons
2. The number of neurons is 2/3 of the input neurons plus the output of the neurons
3. The number of neurons is less than 2 times the input of neurons From the above criteria, this study will use 2 hidden layers with 9 neurons on each layer.
c. Output layer Generally, the number of neurons in the output layer depends on the classification problem. This study includes a multi-class classification, which will classify three types of mangrove sprouts, so that the number of neurons in the output layer is 3.
d. Activation function The activation function commonly used in classification on neural networks with a value between 0-1 is the sigmoid activation function, but the sigmoid activation function is generally used for binary classification. This study uses a multi-class classification which has 3 neurons in the output layer. The softmax activation function can be used in multi-class classification which has an output of more than 2 neurons, this activation function is generally used at the output layer in a neural network. So this research uses the sigmoid activation function in the hidden layer, and the softmax activation function in the output layer.
e. Parameters Parameters such as weights and bias used are random, and the learning rate used is 0.1. Other parameters in this study such as momentum are not used, this is based on the use of standard values ??of momentum on several problems that cannot have a major effect, such as doing a combination with softmax. Architectural specifications to be built is depicted below.
2. Steps for Training using DBN
The deep belief network (DBN) that has been trained in the first phase of training will be transformed into a deep neural network (DNN) by adding a discriminative layer ???? at the top layer of the RBM, then changing the connection within the DBN (from two directions to one direction) becomes feed-forward. The parameters that have been previously estimated by DBN can be used directly as DNN parameters, or can also be improved, for example by applying the backpropagation algorithm. The flow of the backpropagation algorithm is as follows:
1. Using the same number of neurons and structures as the DBN that was pre-trained using the Contrastive Divergence algorithm in the first training phase.
2. Change the direction of the weight which was originally two-way or not directional to one direction like the Multi-layer Perceptron
3. Set weight for each weight and bias is the final weight in the DBN Contrastive Divergence process above.
Feed Forward
This section contains implementation and testing of applications based on analysis and system design that has been discussed in the previous chapter. This stage aims to displays the results of the system design that has been built and the testing process a system for identifying glaucoma.
E. Application Requirements
In designing the glaucoma disease identification application through fundus images using the Deep Belief Network requires both hardware and software supporters include:
The hardware specifications used in this application are:
a. 6 Core vCPU VM Processor (2 units)
b. 2 Core vCPU VM Processor
c. 56 GB DDR4 RAM (2 units)
d. 8GB DDR4 RAM
e. 340 GB Temporary Storage (2 units)
f. 110 GB SSD Storage (3 units)
g. Nvidia Tesla K80 (2 units)
2. Software
The software specifications required in this study include:
a. CentOS6.064 bit Operating System [Server].
b. Windows 10 64 bit Operating System along with Browser (Chrome) [Client].
c. Anaconda Python 3.6.
d. Library: Pandas,Numpy, Matplotlib, PIL, Keras, Tensorflow, OpenCV.
e. Jupyter Notebook
3. Data Implementation
This study uses a dataset in the form of digital images derived from retinal photographs. The dataset obtained is sourced from an online data science and artificial intelligence platform called RIM-ONE dataset (http://medimrg.webs.ull.es) however the same data is also available on kaggle.com. From these data, the classification can be divided into 2 categories of eye blindness, namely normal and glaucoma.
4. Preprocessing
To start the research, pre-processing was carried out which consisted of several stages. Pre-processing is done to convert digital raw into digital images that are ideal for the training process. The first thing to do is to make sure the dataset used has a label for the type of blindness that each eye has and can be used for the training process. The data that has been prepared must be visualized to show the eye and the type of label as shown below.
From the results of the visualization above, it can be ascertained that the data used already has a label and is a dataset that is in accordance with the needs of this study. Then the thing that must be done is pre-processing so that the dataset becomes an ideal dataset so that there are no obstacles in the training process.
5. Brightness
This stage increases the brightness of the image so that the optical disc looks even more it is clear where this stage is carried out so that the image can be more easily processed at the stage next.
7. Grayscale
Grayscale means a process to turn an eye digital image into a monochromatic digital image. In this study, the purpose of carrying out the grayscale process is to ensure that the eye anomaly that has been described as a parameter can be seen when the training process is in progress. At this stage the grayscale process is carried out using the gaussian blur method. The gaussian method is combined with the image blending process. The image blending formula used is as follows.
The image blending process with the Gaussian blur method allows for several variations. Variation is carried out at α using a value between 0 and 1. The process of changes can produce different transformations depending on the type of requirement.
F. GLCM
The extraction1of1second-order statistical1features is1carried out1with a1co-occurrence matrix,1which is1an intermediate1matrix that1represents the1adjacency relationship1between pixels1in an1image in1various orientations1and spatial1distances .1The co-occurrence1matrix is1a matrix1of size1L x1L (L1represents the1number of1gray levels)1with the1element P1(x1, x2)1which is1a joint1probability distribution1of a1pair of1points with1a gray1level x11located at1coordinates (j,1k )1where x21is located1at coordinates1(m, n).1The coordinates1of the1pair of1points are1r with1the angle1θ. Thesecondlevelhistogram P(x1, x2)1is calculated1with the1following approximation results.
G. System Testing using DBN
System testing is carried out to determine the ability of the system that is built to identify fundus images. The system's ability to identify fundus images depends on the training process using the Deep Belief Network because the training procedure produces the final weights that will be used in the testing phase. Previously, 66 glaucoma images and 66 normal images were tested with different epoch parameters. The graph of the system accuracy value after testing 40 images with different epoch parameters havebe see in Figure 4.6. From tests carried out by different epoch parameter standards as under with accuracy of 98.00%.
Based on the system evaluation score using confusion matrix, the system is considered quite effective in identifying glaucoma through fundus images using the method that has been applied to the system. The system achieves a precision in advance glaucoma the precision score of 99%, recall of 100%, and the f1 score of 99% whereas in normal control or glaucoma the precision score of 100%, recall of 98%, and the f1 score of 99% and the overall system accuracy score of 99.0%.
Based on the results of the experiment from this study, it can be concluded that the use of the Deep Belief Network architectural model results in a maximum and linear change in process. The results of research using Deep Belief Network produce a more rational accuracy when compared to other Neural Architectural models as student in literature with several iterations of almost 99% of accuracy. Therefore, this scheme can act as defacto standard for Glaucoma Detection.
The conclusions and suggestions obtained in building an application for glaucoma identification through fundus images using the Deep Belief Network. A. Conclusion The conclusions that can be drawn from the results of testing the identification of glaucoma through fundus images using the Deep Belief Network are as follows: 1) The Deep Belief Network method is able to identify glaucoma through fundus images well. The results of the identification process of glaucoma through fundus images with an advance glaucoma the precision score of 99%, recall of 100%, and the f1 score of 99% whereas in normal control or glaucoma the precision score of 100%, recall of 98%, and the f1 score of 99% and the overall system accuracy score of 99.0%. 2) The selection of DBN parameter values has an influence on the accuracy results. The parameters used are supervised epoch 100. 3) The difference in the values used for the parameters also affects the duration of the training. The higher the epoch value, the longer the training data will be. B. Suggestions The suggestions that can be given by the author regarding this research for further development are as follows: 1) Using other neural network methods to compare with the classification results obtained from the Deep Belief Network method. 2) Inculcate the layers which can reduce the training time with same level of accuracy.
[1] https://www.medicinenet.com/glaucoma/article.htm [2] https://www.mayoclinic.org/diseases-conditions/glaucoma/symptoms-causes/syc-20372839 [3] https://www.webmd.com/eye-health/glaucoma-eyes [4] Karegowda, Asha &Nasiha, Asfiya& M.A, Jayaram & Manjunath, A.. (2011). Exudates Detection in Retinal Images using Back Propagation Neural Network. International Journal of Computer Applications. 25. 10.5120/3011-4062. [5] Li, Yunghui& Yeh, Nai-Ning & Chen, Shih-Jen & Chung, Yu-Chien. (2019). Computer-Assisted Diagnosis for Diabetic Retinopathy Based on Fundus Images Using Deep Convolutional Neural Network. Mobile Information Systems. 2019. 1-14. 10.1155/2019/6142839. [6] T. Khalil, M. U. Akram, H. Raja, A. Jameel and I. Basit, \"Detection of Glaucoma Using Cup to Disc Ratio From Spectral Domain Optical Coherence Tomography Images,\" in IEEE Access, vol. 6, pp. 4560-4576, 2018, doi: 10.1109/ACCESS.2018.2791427. [7] Ahmad, Jamil & Muhammad, Jan & Aziz, Lubna&Ayub, Sara &Akram, M. & Basit, Imran. (2016). Glaucoma detection through optic disc and cup segmentation using K-mean clustering. 143-147. 10.1109/ICECUBE.2016.7495212. [8] Galperin, Yevgeniy. (2021). Introduction to Basics of Digital Images. 10.1201/9780429400612-1. [9] Gan, Woon. (2020). Digital Image Processing. 10.1007/978-981-10-5550-8_10. [10] Seeram, Euclid. (2021). Digital Image Processing Concepts. 10.1007/978-981-15-6522-9_2. [11] Reis, George. (2017). Digital Image Processing. 10.4324/9781315693125-12. [12] Rizani, Fitri. (2021). IMAGE QUALITY IMPROVEMENT USING IMAGE PROCESSING METHOD IMAGE BRIGHTNESS CONTRAST AND IMAGE SHARPENING. MULTICA SCIENCE AND TECHNOLOGY (MST). 1. 6-12. 10.47002/mst.v1i1.200. [13] Kuznetsov, Yuri. (2021). Image Processing. 10.1007/978-3-030-60955-9_2. [14] Zielinski, Tomasz. (2021). Image Processing. 10.1007/978-3-030-49256-4_16. [15] Micek, J. &Kapitulík, Ján. (2003). Median filter. Journal of Information, Control and Management Systems. 1. [16] Pitas, I. &Venetsanopoulos, A.. (1990). Median Filters. 10.1007/978-1-4757-6017-0_4. [17] Kumar, Abhishek &bhargava, Prof.Neeeaj. (2016). A Practical approach for image processing & computer vision in MATLAB. [18] Bhargav, Dr.Neeraj& BHARGAV, Dr.Ritu& Kumar, Abhishek. (2016). A Practical Approach of Image Processing &Computer Vision in MATLAB. [19] Rahman, Mansib. (2017). Computer Vision & Image Processing. 10.1007/978-1-4842-2316-1_6. [20] Shapiro, Linda & Rosenfeld, Azriel. (1992). Computer vision and image processing.. [21] Marques, Oge. (2020). Image Processing and Computer Vision in iOS. 10.1007/978-3-030-54032-6. [22] Fisher, Robert & Dawson-Howe, K & Fitzgibbon, Andrew & Robertson, C & Trucco, Emanuele. (2005). Dictionary of Computer Vision and Image Processing. 10.1002/0470016302. [23] Parker, James. (1997). Algorithms for Image Processing and Computer Vision. [24] Eisner, Howard. (2021). Artificial Intelligence. 10.1201/9781003160618-8. [25] Lang, Volker. (2021). Artificial Intelligence. 10.1007/978-1-4842-6774-5_4. [26] C.R, Rathish. (2021). Artificial Intelligence. [27] Monte-Serrat or Monte Serrat, Dionéia&Cattani, Carlo. (2021). Artificial intelligence. 10.1016/B978-0-12-824118-9.00009-6. [28] Szeliski, Richard. (2010). Computer Vision: Algorithms and Applications (Texts in Computer Science). [29] Szeliski, Richard. (2011). Computer Vision: Algorithms and Applications. 10.1007/978-1-84882-935-0. [30] Zhang, Xin & Xu, Shuo. (2020). Research on Image Processing Technology of Computer Vision Algorithm. 122-124. 10.1109/CVIDL51233.2020.00030. [31] Prince, Dr. (2012). Computer Vision: Models, Learning, and Inference. [32] Perros, Harry. (2021). Artificial Neural Networks. 10.1201/9781003139041-10. [33] Cartwright, Hugh. (2021). Artificial Neural Networks. 10.1007/978-1-0716-0826-5. [34] Okwu, Modestus &Tartibu, Lagouge. (2020). Artificial Neural Network. 10.1007/978-3-030-61111-8_14. [35] Czischek, Stefanie. (2020). Artificial Neural Networks. 10.1007/978-3-030-52715-0_3. [36] Bell, Jason. (2020). Artificial Neural Networks. 10.1002/9781119642183.ch9. [37] Baldi, Pierre. (2021). Deep Learning in Science. 10.1017/9781108955652. [38] Watson, Samuel. (2020). Deep Learning. 10.1201/9780429398292-9. [39] Karimi, Zohreh. (2021). What is Deep Learning?. [40] Paper, David. (2021). Introduction to Deep Learning. 10.1007/978-1-4842-6649-6_1. [41] Ketkar, Nikhil &Moolayil, Jojo. (2021). Deep Learning with Python: Learn Best Practices of Deep Learning Models with PyTorch. 10.1007/978-1-4842-5364-9. [42] Khan, Asifullah& Islam, muhammad. (2016). Deep Belief Networks. 10.13140/RG.2.2.17217.15200. [43] Lopes, Noel & Ribeiro, Bernardete. (2015). Deep Belief Networks (DBNs). 10.1007/978-3-319-06938-8_8. [44] Hua, Yuming& Guo, Junhai& Zhao, Hua. (2015). Deep Belief Networks and deep learning. Proceedings of 2015 International Conference on Intelligent Computing and Internet of Things, ICIT 2015. 1-4. 10.1109/ICAIOT.2015.7111524. [45] Koo, Jaehoon&Klabjan, Diego. (2020). Improved Classification Based on Deep Belief Networks. 10.1007/978-3-030-61609-0_43. [46] Mclaughlin, Tom & Le, Mai &Bayanbat, Naran. (2021). Emotion Recognition with Deep-Belief Networks. [47] Fan, Rong & Hu, Wenxin. (2017). Face recognition with improved deep belief networks. 1822-1826. 10.1109/FSKD.2017.8393043. [48] Nayak, Jagadish & Acharya, U Rajendra & Bhat, P & Shetty, Nakul & Lim, Teik-Cheng. (2009). Automated Diagnosis of Glaucoma Using Digital Fundus Images. Journal of medical systems. 33. 337-46. 10.1007/s10916-008-9195-z. [49] Sengar, N., Dutta, M.K., Burget, R. &Ranjoha, M. 2017. Automatic Detection of Suspected Glaucoma in Digital Fundus Images. 2017 40th International Conference on Telecommunications and Signal Processing (TSP). [50] Atheesan, S. &Yashothara, S. 2016. Automatic Glaucoma Detection by Using Funduscopic Images. IEEE Wispnet 2016 Conference. [51] S Kumar and D. Singh, Energy and exergy analysis of active solar stills using compound parabolic concentrator, International Research Journal of Engineering and Technology (IRJET), 6 (2019) 12. [52] R. Shanker, D. Singh, D. B. Singh “Performance analysis of C.I. engine using biodiesel fuel by modifying injection timing and injection pressure” International Research Journal of Engineering and Technology(IRJET) 6 (2019) 12. [53] A. K. Anup and D. Singh, FEA analysis of refrigerator compartment for optimizing thermal efficiency, International Journal of Mechanical and Production Engineering Research and Development, 10 (2020) 3, 3951-3972. [54] S Kumar and D. Singh, Optimizing thermal behavior of compact heat exchanger, International Journal of Mechanical and Production Engineering Research and Development, 10 (2020) 3, 8113-8130. [55] Dharamveer and Samsher, Comparative analyses energy matrices and enviro-economics for active and passive solar still, materialstoday: proceedings, https://doi.org/10.1016/j.matpr.2020.10.001 (2020). [56] Dharamveer, Samsher, Anil Kumar, Analytical study of Nth identical photovoltaic thermal (PVT) compound parabolic concentrator (CPC) active double slope solar distiller with helical coiled heat exchanger using CuO Nanoparticles, Desalination and water treatment, 233 (2021) 30-51, https://doi.org/10.5004/dwt.2021.27526 [57] Dharamveer, Samsher, Anil Kumar, Performance analysis of N-identical PVT-CPC collectors an active single slope solar distiller with a helically coiled heat exchanger using CuO nanoparticles, Water supply, October 2021, SCI-E Index, IWA Publication. I.F 1.275,https://doi.org/10.2166/ws.2021.348 [58] M. Kumar and Dharamveer Singh, Comparative analysis of single phase microchannel for heat flow Experimental and using CFD, International Journal of Research in Engineering and Science (IJRES), 10 (2022) 03, 44-58. [59] Kavya, N. &Katmaja, Dr. K.V. 2017. Glaucoma Detection Using Texture Features Extraction. 2017 51st Asilomar Conference on Signals, Systems, and Computers. [60] Hyde, Kayleigh & Novack, Marlena & LaHaye, Nick &Parlett-Pelleriti, Chelsea &Anden, Raymond & Dixon, Dennis &Linstead, Erik. (2019). Applications of Supervised Machine Learning in Autism Spectrum Disorder Research: A Review. Review Journal of Autism and Developmental Disorders. 6. 10.1007/s40489-019-00158-x. [61] Golob, Anna & Takahashi, Traci & Johnson, Kay.(2020).Breast Cancer Screening.10.1007/978-3-030-50695-7_18. [62] Varshni, K. Thakral, L. Agarwal, R. Nijhawan and A. Mittal, \"Pneumonia Detection Using CNN based Feature Extraction,\"2019IEEEInternational Conferenceon Electrical,Computer and Communication Technologies (ICECCT),2019,pp.1-7,doi:10.1109/ICECCT.2019.8869364. [63] Carlos Alberto Torres Naira and Cristian Jos´eL´opez Del Alamo, “Classification of People who Suffer Schizophrenia and Healthy People by EEG Signals using Deep Learning” International Journal of Advanced Computer Science and Applications(IJACSA), 10(10), 2019. http://dx.doi.org/10.14569/IJACSA.2019.0101067
Copyright © 2022 Aashish Nayak, Dr. Dharamveer Singh, Vikas Gupta. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Paper Id : IJRASET45227
Publish Date : 2022-07-01
ISSN : 2321-9653
Publisher Name : IJRASET
DOI Link : Click Here