Ijraset Journal For Research in Applied Science and Engineering Technology
Authors: Rahila Ayoub, Dr. Ashish Oberoi
DOI Link: https://doi.org/10.22214/ijraset.2021.39245
Certificate: View Certificate
Within the biometric industry, computerized person identification using ear pictures is a hot topic. The ear, like other biometrics like the face, iris, and fingerprints, contains a huge number of particular and unique traits that may be used to identify a person. Due to the mask-wearing scenario, most face detection methods fail in this present international COVID-19 pandemic. The eardrum is a great data source for inactive person authentication since it doesn\'t necessitate the person we\'re attempting to pinpoint to cooperate, and the structure of the ear doesn\'t change significantly over time.. The acquisition of a human ear is also simple because the ear is apparent even while wearing a mask. An ear biometric system can enhance other biometric technology in an automated person identification system by giving authentication cues when other information is unreliable or even missing. We provide a six-layer deep convolutional architecture for ear identification in this paper. On the IITD ear dataset, the deep network\'s potential efficiency is assessed. The IITD has a detection performance of 97.36 percent for the deep network model and 96.99 percent for the IITD. When paired with a competent surveillance system, this approach can be beneficial in identifying people in a large crowd.
I. INTRODUCTION
We need to show that utilising the ear as the base for a new category of biometrics is feasible. We can't prove that every person has a fully unique pair of ears in the same way that no one can establish that fingerprints are unique. Instead, we may remark that this is probable and back it up with evidence from two Iannarelli investigations [1]. The first examines over 10,000 ears from a randomly selected location in California, while the second examines tested twin and sister twins with genetic factors that are regarded equal. Both studies reinforce the idea that the ear possesses precise physiological features, keep in mind that in both trials, all of the ears tested were shown to be distinctive, even though the twins were shown to be similar. The next stage is to investigate if the ear can offer fingerprints that are constant over time after exhibiting distinctiveness. The ear's structure does not appear to vary significantly over time. According to medical study [1,] ear growth after the first four days of pregnancy is proportionate. Despite the fact that ear boom is proportionate, gravity appears to be capable of vertically growing the ear. The ear lobe is the most affected by this lengthening, and measures reveal that the exchange is non-linear. The rate of elongation is around five times higher than usual from of the age of 4 months to the age of eight, after which it remains constant until around the age of 70, when it accelerates again. The ear fingerprints are plausible in the sense that ear construction is expected to be specific to each person and that features field measurements of that morphology are consistent across time. Ear biometrics authentication is fascinating if they are feasible since it is considerably more inactive than face detection, rather than challenging to separate face biometrics, robust such as certain extracted fingerprint scanners like those in fingerprints might be used.
A. Iannarelli's Ear Biometrics
Iannarelli [1] came up with an anthropometric recognition approach based on ear biometrics. Figure 1.1 shows the twelve estimations that make up the "Iannarelli System" (b). The coordinates shown were determined using aligned and normalised photographs of the proper ear. To normalise and align the images, they are shown onto a traditional "Iannarelli Inscribed" extending easel that is adjusted horizontally and vertically until the ear image falls into a present location on the easel. The mechanism, according to Iannarelli, needs careful positioning and levelling of the ear pox: After the ear is fixed and the image is limited within the easel boundaries, set the easel such that the indirectly guiding line is parallel to the outer strong spot of the tragus flesh line. The tragus's tip should be just touching the direct line. (For the 5 white lines in the left ear biometrics, see Figure 1.1(a).) Start moving the accessory gently until the top section of the oblique booklet line intersects with the point on the ear photo in which the internal helix rim surface shapes the top cucaracha flesh line underneath the depression or hollow known as the rectangle fossa, while preserving the ambiguous line touching the tragus.
When the indirectly guideline was utilised to align the ear shot, the ear photo was carefully synchronised. The size of the ear image must now be adjusted by the professionals. The ear picture is stretched or shrunk to its suitable size for comparison and classification using the easel's short vertical rule (the right white line in Figure 1). 83-84] 83-84] [1, pp. 83-84]
Figure 1.Ear Biometrics: (a) 1 Helix Rim, 2 Lobule, 3 Antihelix, 4 Concha, 5 Tragus, 6 Antitragus, 7 Crus of Helix, 8 Triangular Fossa, 9 IncisureIntertragica. (b) The locations of the anthropometric measurements used in the “Iannarelli System”
The final photos are standardised because each ear is matched and scaled during the design process, allowing for the retrieval of similar measurements from the pix all at once. The length between each of the named parts in Figure 1.(a) is considered in three devices and given an integer value. These twelve factors, as well as data on cohabitation and race, are then used to identify the person. Because an issue is classified into a single point in a 12 dimensional numeric space, where each unit on an axis represents a 3 mm height difference, inside each relationship and race group, the system allows for a too small class region, as previously stated. Using the 12 measures, presuming a four-unit average variability within the group 6, Burge and Burger get a space with less than 17 million great points (i.e., 12 mm). Though there are simple ways for raising the scale of the length (e.g., adding more measurements or adopting a smaller metric), the technique is inadequate for system vision owing to the difficulties in detecting the anatomical point that serves as the size gadget's starting point. Because all future variables are measured in respect to this origin, if it is not exactly determined, all subsequent measurements will be incorrect. On site page 83, Iannarelli states, "This is step one in synchronising the ear image....," indicating that he was aware of the problem. And it must be accurate; else, the entire kind of the ear may be erroneous."
II. OBJECTIVES
III. LITERATURE REVIEW
The notion of skin reflection within the safe range of the EM (Electromagnetic) part of the electromagnetic spectrum was used by Aakanksha and Setia, H, in their recommended painting [6], due to the fact that the skin has a 54 percent mirrored image and the fats under the skin have an 8 % reflected image. It was suggested that an ear imaging mechanism be used, as well as adequate monitoring of light reflected to use a photometer. Following that, the reflected spectrum is standardized, and the identical spectra are used to paint every ear separately. Codes and techniques for programming the spectrometer and determining the cutoff cost based on the most allowable error were demonstrated. How can the attractiveness correctness be changed indirectly when light reflection differs from ear to ear and each person's pores, skin, and fat intake is unique? In 2012, P. Ramesh Kumar et al. introduced the pixel-based feature identification technique [7] for the Ear biometric model. Offering a strategy based on ear shape photo pixels invites researchers to focus on ear identity verification since pixels are the essential components of an image.
In 2017, Mozammel Chowdhury et al. proposed an ear biometric attractiveness strategy based on the local capabilities of the human ear and a neural network algorithm [14]. The approach extracts the edge features from the recognized ear after guessing the position of the ear from the facial picture. They were chosen since portion local functions are unaffected by strobing lights and obstruction. By matching the user's collected ear functions to a feature database, a neural network is used to understand the client. The robustness and efficacy of the proposed strategy are excellent, but there is still potential for development. . In 2018, Salman Mohammed Jidda and colleagues [15] released research that employed textural and geometric skills to enhance ear identification. The AMI ear database was utilized in this study to try recovering Local Binary Patterns abilities and using the Laplacian filter on raw pictures one at a time to extract architectural skills. In order to find the region of intense interest in human ear photos, the ear database was analyzed by dividing ear images into four quarters and experimenting on each one independently. The textural and geometric variables were then fused, and tests were run to verify the contribution of the fused capability. While the top left zone ear image accuracy was finished, the upper left ear pictures with the Laplacian filter had a substantially lower average accuracy of 51.8 percent, which needed to be improved.
IV. METHODOLOGY
A sensor in an ear biometric device is frequently used to acquire the ear image (camera). The system does image pretreatment and feature extraction, as well as identifying features that may be utilized to distinguish between ear pictures. These features are stored in a feature database. During testing, the attributes produced from the test image are matched to those in the database. Finally, the classifier chooses from the database the best relevant ear image. Three main steps may be identified in a general ear biometric system, as shown in Fig.2. The three phases are preprocessing, feature extraction, and classification. Preprocessing and feature extraction are used in both the training and testing phases. This section examines the three phases as well as the major algorithms used in each.
A. Preprocessing
Prior to processing, preprocessing is necessary to transform the photographs to a standard and suitable format. Data preparation ensures that the image's most important attributes are enhanced while unwanted data such as noise is removed. The two primary types of preprocessing approaches are filter-based and vividness processes.
B. Feature Extraction
Simply put, feature extraction is the process of determining the main characteristics of a photograph that enable it to be recognized. The goal of feature extraction is to reduce dimensionality by using a small number of features to describe the entire image. The aspects of a photograph that can be used to features extracted are listed below.:
Many scholars have contributed minutiae extraction algorithms based on the foregoing qualities from the geographical and magically morph realms of image processing.
a. Canny edge detection: The edges of the ear picture are located using this procedure. A Gaussian filter is used to reduce noise. The slope is then calculated using any gradient operator, such as Sobel, Roberts, or Prewitt. Pixels that exceed a specified limit are selected and preserved as edge elements. These segments' boundaries are then stitched together to form unbroken segments [22, 29].
b. Contours of ear images: Several geometrical properties of the ear are identified using the ear contour. Only a few of these features are the curve start and finish locations, curve path points, and the gradient of the line between the start and end points of the contour. They're employed to create the feature database that may be used to verify the ears' authenticity [10].
c. Statistical feature based method: Ear image statistics such as ear height, ear width, and angles produced here between external borders of the ear and the half of the line forming the ear's height are included in this method's feature database [11, 30]..
2. Appearance based techniques: Using image techniques loudness and roughness, feature data is retrieved. These algorithms employ shape features and matrix factorization techniques to extract individual ear features. The specifics are as follows:
a. Feature descriptors: A feature extraction module is a method that returns an image as its input and outputs a feature vector. Local Binary Patterns (LBP) and Local Phase Quantization (LPQ) are the most extensively used pattern identifiers for storing local texture data in the spatial and spectral [20,21,31]. The Histogram of Oriented Gradients (HOG) is another classification method that counts the occurrences of distinct gradient directions within a small region of an image [21, 24, 31, and 33]. Binarized Statistical Images Features (BSIF) [20, 21, 31], Patterns of Oriented Edge Magnitudes (POEM) [21], and Scale-Invariant Feature Transform (SIFT) [32] are other possible feature descriptors used in ear biometrics.
b. Dimensionality Reduction Techniques: PCA [21, 24, and 34] is a matrix factorization technique that takes a higher-dimensional feature map and projects it to a lesser input vector while preserving its unique properties. Linear Discriminant Analysis (LDA), which emphasizes on preserving class-discriminatory information, is another method for image compression [31]. Reduced-dimensionality approaches are typically used with other methods, such as local features.
c. Force Field Transformations: It's an intensity-based approach that treats each pixel in the image as a source of force fields. Each pixel generates a symmetrical circular force field. They help to reduce energy wells, lines, and routes by using the energy properties of the ear image. The noise sensitivity of this approach is great [15].
d. Wavelets-based techniques: Haar wavelets are one of the most well-known and fundamental wavelets. The textural features of the ear are extracted using the Wavelet packet transform. The feature vectors are then constructed by quantizing the wavelet. Because of its inherent property of robust image fragmentation and rebuilding, this technique is beneficial in the field of biometrics [16, 17 and 37]..
C. Classification and Authentication
After feature extraction and classification vector construction, the program must compare the characteristics of the test picture to the elements in the database. Finally, the picture has been validated. The below are some of the several methods for doing so:
V. RESULTS AND OBSERVATIONS
Based on the above discussion, our methodology will comprise three steps
A. Data-Set Collection:
We'll use the consistently identifies IITD, that is used by the overwhelming bulk of researchers in this area, as well as create our unique dataset by collecting ear images from some of our agency's pupils..
B. General Procedure of Proposed Technique
The method for biometrics will be carried out in three steps:
1. The Discreet Wavelet Transform (DWT) will be used since it offers the following pre-processing: When contrasted to other pre-processing methods, this technique yields better results.
2. Feature Extraction: For feature extraction, we will utilise the PCA matrix factorization technique, which is computationally efficient and yields excellent results
3.Classification algorithm: For classification, we'll utilise the KNN Classifier, which outperforms other. classification techniques and is very scalable
C. Evaluation Procedure
The following parameters will be taken into account throughout the evaluation:
4.Samples
We will use these parameters to compare with the already existing ear biometrics technique.
Table 1 Performance matrix
Decision threshold |
Recall TRP |
Precision |
FAR |
FFR |
0.961 |
0.91 |
0.70 |
0.11 |
0.11 |
0.962 |
0.98 |
0.71 |
0.12 |
0.11 |
0.963 |
0.88 |
0.76 |
0.14 |
0.12 |
0.964 |
0.81 |
0.75 |
0.10 |
0.10 |
|
|
|
|
|
|
|
|
|
|
The results of the experiments reveal some intriguing details insight into the underlying geometric properties of the ear. The common challenges observed during feature extraction and subsequent derivation of inter feature mathematical hash values from the tests opened up new opportunities for future research in this sector. Common metrics were used to validate the facial similarity predictions. Table I lists the system's performance metrics derived from the numerous simulations.
The threshold values are based on a Bayesian comparison of two pictures of ears. The False positive rate (FAR) and False rejection rate (FRR) drop when the threshold is raised, but the Recall or True positive rate (TPR) begins to fall. As can be seen, the average degree of connection between two images of the same topic is around 0.96. Meanwhile, images with varied topics have an average score of 0.85 to 0.93. Because the human side profile changes only little, comparing side faces using geometric computations needs a high degree of accuracy (considering the experimental setup)
In this study, we developed a simple deep CNN architecture for recognising people from ear images. Classical computer vision systems use hand-crafted features, but CNNs learn the features from the input image right away, resulting in better performance. We explored the potential efficiency of the built Deep CNN on the IITD-II Ear dataset by modifying factors such as kernel size, learning rate, epochs, and activated functions, and we achieved a 97.36 overall accuracy. In both a controlled and uncontrolled context, the Deep CNN is evaluated against the AMI ear dataset, and it obtains a respectable recognition rate. Because the model requires extremely minimal RAM, it may be transformed into any embedded/handheld device. When this model is combined with an adequate monitoring program, automatic human recognition in densely crowded areas such as malls, railway stations, and banks is possible
[1] Iannarelli, Ear Identification. Forensic Identification Series. Paramont Publishing Company, Fremont, California, 1989. [2] M. Rahman, M. R. Islam, N. I. Bhuiyan, B. Ahmed, A. Islam, Person identification using ear biometrics, International Journal of The Computer, the Internet and Management 15 (2) (2007) 1–8. [3] Yan P and Bowyer KW (2007) Biometric recognition using 3D ear shape. IEEE Trans Pattern Anal Mach Intell, vol. 29, no. 8: 1297-1308 [4] D. Watabe, H. Sai, K. Sakai, and O. Nakamura, “Ear biometrics using jet space similarity,” 2008 Canadian Conference on Electrical and Computer Engineering, 2008. [5] A. Tariq, M. A. Anjum, and M. U. Akram, “Personal identification using computerized human ear recognition system,” Proceedings of 2011 International Conference on Computer Science and Network Technology, 2011. [6] Aakanksha and H. Setia, “Ear as a platform for latest biometric solution,” 2011 International Conference on Emerging Trends in Networks and Computer Communications (ETNCC), 2011. [7] P. R. Kumar and S. Dhenakaran, “Pixel based feature extraction for ear biometrics,” 2012 International Conference on Machine Vision and Image Processing (MVIP), 2012. [8] D. Watabe, T. Minamidani, W. Zhao, H. Sai, and J. Cao, “Effect of Barrel Distortion and Super-Resolution for Single-View-Based Ear Biometrics Rotated in Depth,” 2013 International Conference on Biometrics and Kansei Engineering, 2013 [9] G. S. Kamatchi and R. Gnanajeyaraman, “Ear as a Raised area for Hottest Biometric Solution Using Universality, Distinctiveness, Permanence and Measurability Properties -- PURS Using UDPM,” 2014 World Congress on Computing and Communication Technologies, 2014. [10] S. Khobragade, D. D. Mor, and A. Chhabra, “A method of ear feature extraction for ear biometrics using MATLAB,” 2015 Annual IEEE India Conference (INDICON), 2015. [11] U. Kacar, M. Kirci, E. O. Gunes, and T. Inan, “A comparison of PCA, LDA and DCVA in ear biometrics classification using SVM,” 2015 23nd Signal Processing and Communications Applications Conference (SIU), 2015. [12] B. S. Fabito, “Asynchronous examination security system using Ear Visual Biometric for college students,” TENCON 2015 - 2015 IEEE Region 10 Conference, 2015. [13] E. J. Ragan, C. Johnson, J. N. Milton, and C. J. Gill, “Ear biometrics for patient identification in global health: a cross-sectional study to test the feasibility of a simplified algorithm,” BMC Research Notes, vol. 9, no. 1, Feb. 2016
Copyright © 2022 Rahila Ayoub, Dr. Ashish Oberoi. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Paper Id : IJRASET39245
Publish Date : 2021-12-03
ISSN : 2321-9653
Publisher Name : IJRASET
DOI Link : Click Here