Authentication Performance Analysis of Acoustic-Based Ear Biometrics System Using Audible Signals
##plugins.themes.bootstrap3.article.main##
A study of an acoustic-based ear biometrics system for individual identification using audible signals was undertaken. In this era of too much network connectivity, an increase in cybercrimes has led to the loss of key information or leakage of it. Experts have therefore continued to work hard to protect and secure important data from being stolen, lost, leaked, or tempered with. The use of biometric systems is one of the methods that many have adopted to provide the needed security and protection of data. The authors in this study proposed an acoustics-based ear biometric system for the identification of individuals while taking into consideration the dynamics of real-time data capture and using real persons for user authentication applications. The system has been developed in the MATLAB/SIMULINK language which supports dynamic real-time data capture. The results of simulation experiments showed that with proper experimentation and threshold calibration, it is possible to develop acoustics-based authentication systems that can identify individuals correctly and with 100% recognition accuracy. Depending on the human subjects under study, the threshold cosine similarity setting may vary between 0.2 and 0.4. However, this variation is offset by the enrollment procedure deployed in practice. Furthermore, trend analysis using moving average analysis revealed the possibility that a false acceptance is equally likely even though 100% recognition accuracy was attained.
Downloads
Introduction
In this era of many networks’ connectivity, an increase in cybercrimes has led to loss of key information or leakage of it. Hearable devices like The Dash or Xperia Ear are also rapidly increasing. Experts have therefore continued to work hard to protect and secure important data from being stolen, lost, leaked, or tempered with. The use of biometric authentication is one of the methods that many have adopted to provide the needed security and protection of data [1]. Authentication of users comes in different forms such as voice recognition, use of fingerprint, use of iris or retina, etc., [2], [3]. To capture the user’s identity, each user is expected to expose the relevant body part (the eye, finger, palm, etc.) to a camera or place the palm on a scanner for proper capturing.
Biometrics authentication has been extended to the ear as a part of the human body, and much progress has been made as far as ear biometrics is concerned but different kinds of issues are readily encountered [4]. Besides ear biometrics based on ear recognition, acoustic ear recognition is another powerful or useful authentication method that is based on the exclusive acoustic features of different ear canals [5]. Interestingly, [6]–[9] carried out research that used the acoustic properties of the ear canal and the pinna for the purpose of user recognition.
In this paper, we intend to carry out quiet authentication of users using inaudible sound signals. The reason for the quiet authentication is to avoid interrupting the users. The efficiency of the method will be analyzed. The sections of this work are Section I which is the introduction, Section II which presents the literature review, Section III is the methodology adopted in the work, Section IV is the result presentation, and finally, Section V is the summary and conclusion.
Literature Review
The authors in [1] presented an identification system that used ear acoustic biometrics. The research was aimed at investigating how feasible the use of ear acoustic biometrics that takes advantage of the ear canal’s reflective properties as a Head Related Transfer Function (HRTF) can be when it comes to the recognition of individuals in real-time within dynamic environments. It was shown that with good calibration of threshold and adequate experimentation, the development of an authentication system based on ear acoustics capable of identifying people correctly is possible.
An investigation of fast individual authentication simulator system that considers Ear Canal Impulse Response, known as ECIR for short, and Ear Canal Transfer Function, known as ECTF for short, together with an earbud for the probing of microphone signal was carried out by [6]. The research used the acoustic properties of the ear canal and the pinna for the purpose of user recognition. [10] used certain audio signals that were audible and had a range of frequency of about 1500 Hz for capturing a person’s ear acoustics. It was mentioned that based on the kind of device used for capturing, error rates achieved ranged between 0.8 percent and 18 percent.
Mahto et al. [11] presented a detailed study on ear acoustics biometrics based on in-audible signals. The authors emphasized that even though the user is busy discussing, he can still be authenticated without any negative effect due to background noise. Zhong et al. [12] gave research work on ear acoustics biometric systems for the recognition of individuals through key features of the ear. In the study, an Equal Error Rate of 14.9 percent was obtained at the best performance. Since the error rate was that high, the adopted methods will have few applications in real-world situations.
Materials and Method
Materials
The materials considered necessary for this work include a personal computer, loudspeaker, earphones, sound card, etc.
Method
Based on Fig. 1, there are two signals, the echo signal represented by and the probe signal represented by . There is a transmission of through the ear canal while the produced echo signal is captured and recorded.
For a pair of echo signals and , cross-spectral density CSD from Fig. 1 is used to extract ear acoustic features. Probe signal and echo signals are fully subjected to Fast Fourier Transform (FFT) so as to get the Discreet Fourier Transform of the input signals. These outputs from the FFT are taken to the cross-spectral density where the CSD between the two the inputs is determined to obtain the transfer function of the ear canal. The ear canal transfer function is given by: where represents the complex conjugate of . Based on the output from the CSD which is the ear canal transfer function, Mel-Frequency Cepstral Coefficients (MFCCs) are extracted and subjected to filtering. Then after applying the Discreet Fourier Transform, the Mel-Frequency Cepstral Coefficients are post-processed. The purpose of the post-processing is to get normalized Mel-Frequency Cepstral Coefficients [13]–[15]. Finally, there is a computation of the similarity between ear acoustic features that have been extracted. The equation for cosine similarity is given by:
If is greater than a certain threshold already fixed, then one person has the two features.
The operation of Fig. 2 is that signals coming from a device like the Loudspeaker positioned near the ear are captured by the ear. The captured signals are pre-amplified by means of a Preamplifier. The sound card parses the pre-amplified signals to a personal computer for signal processing. In addition, the Sound Card provides a probing sound signal which the power amplifier amplifies and sends to the Loudspeaker. Fig. 3 shows a flowchart illustrating the data capture process.
Feature Representation Computation and Head-Related Transfer Function
To find the exclusivity of a particular human subject during the analysis of EBS and bearing in mind the modifying effect of the environment on sound pressure waves affecting the human ear, many compute the HRTF, (Head Related Transfer Function) of the considered subjects. With this, it is possible to evaluate the response of the human ear to changes in the excitation of stimulus probes as stated in [12].
where
PL represents left ear sound pressure,
PR represents right ear sound pressure,
P0 represents free-field sound pressure,
r represents source distance related to head center,
θ – azimuth,
ϕ represents head elevation relative to center,
f represents source distance related to head center,
a – human subject.
The length of the ear canal, the longest and shortest diameters, and the volume of the ear canal are as shown in Table I.
Parameters | Dimensions |
---|---|
Ear canal’s length | 25 (±2.5 mm) |
Longest diameter | 9.4 (±1.5 mm) |
Shortest diameter | 4.8 (±0.5 mm) |
Ear canal volume | 1014.0 (±15.4 mm3) |
Procedures for Requirements and Experiment
Preliminary live-test measurements were based on a whole of 4 subjects carried out within a closed room in the Agricultural Engineering Laboratory, Rivers State University (RSU) and Ignatius Ajuru University, all in Rivers State, Nigeria. The test subjects included two females and two males, and some consecutive measurements were taken with each participant reciting the word “Greeting” for 3 different simulation trials.
Earlier there was an implementation live, to test the technique that was developed. In most cases, the proposed protocols’ testing, and evaluation may not be practically achievable by means of real experiments because that would be far more complex, time-consuming, and highly costly. Therefore, to address this problem, “TESTBEDS and SIMULATORs” are effective tools relied upon for testing and analyzing the protocols’ performance [16].
Test Requirements for Dynamic Environment
a)Here, some small rules are made to limit the volume of ear canal changes. These rules are No gum-chewing, minimize laughing, yawning, smiling, etc. b)Avoiding abrupt head movements to keep the headphones intact.
Test Requirements for Static Environment
a)No movement of subjects for at least 10 seconds. b)Any obstruction to ear measurements must be removed.
Gathering of Dataset
To authenticate the concepts proposed in this research, the HRTF dataset obtained from the IRCAM laboratories was utilized. The dataset is composed of a set compensated Windows WAV file and a mat file which includes the structured data file format for easy analysis and comprehension. There are 187 sample points constituting azimuth, elevation, and signal content data captured at a sampling frequency of 44.1 kHz.
Results and Discussion
In an ideal situation, it should be expected that the moving average trend should move downwards from points 4 to 5 as seen in both Figs. 4 and 5, depicting a low or reducing false acceptance rate because of comparison to other subjects different from the base subject. As seen in the trend plots of Fig. 4, it is obvious that there is a somewhat useful indication in subject 1 of the likelihood that a deviation from the expected values is minimal as the cosine values decrease from the 4th and 5th comparison samples implying the less likelihood of a false acceptance. However, this is not the case when considering subject 2 as there is a tendency towards false acceptance from the 4th and 5th comparison sample trials as seen in Fig. 5.
Table II shows that the cosine similarity values of comparison between counts 2 and 1 are higher than the one of count 3. This can be explained because comparison count 3 is an intersubject recognition of the ear test. Therefore, as we can expect, it should give very low cos values.
Comparison count | Cos (a, b) |
---|---|
1 | 0.2893 |
2 | 0.2255 |
3 | 0.0573 |
Based on Table III, counts 1 and 2 comparison represents the acoustics-based intra-subject ear biometric recognition cases for subject 1. It should be noted that the cosine similarity values of comparison between counts 2 and 1 are higher than the ones of count 3. Again, this can be explained because comparison count 3 is an inter-subject recognition of the ear test. As a result, it should give very low cos values.
Comparison count | Cos (a, b) |
---|---|
1 | 0.3782 |
2 | 0.4820 |
3 | 0.0573 |
Conclusion
The authors in this study proposed an acoustics-based ear biometric system for identifying individuals while considering the dynamics of real-time data capture and using real persons for user authentication applications. The system has been developed in the MATLAB/SIMULINK language which supports dynamic real-time data capture.
Extensive experiments considering the Head Related Transfer Functions (HTRF) of several human subjects in the Rivers State University and the Ignatius Ajuru University campus environs have been conducted. In particular, the aspect of enrolling person data and considering comparative inter/intra-subject analysis were performed on subjects.
The results of simulation experiments showed that with proper experimentation and threshold calibration, it is possible to develop acoustics-based authentication systems that can identify individuals correctly and with 100% recognition accuracy. Depending on the human subjects under study, the threshold cosine similarity setting may vary between 0.2 and 0.4. However, this variation is offset by the enrollment procedure deployed in practice.
Furthermore, trend analysis using moving average analysis revealed the possibility that a false acceptance is equally likely even though 100% recognition accuracy was attained. Thus, it becomes necessary to conduct extensive simulation experiments to further reveal likely failures in the model of the proposed solution and minimize the problem of over-fitting on datasets.
References
-
Orike S, Bakare BI, Iyoloma CI. A unique identification system using ear acoustics biometrics. IOSR J Electr Commun Engr (IOSR-JECE). 2022;17(6):1–8.
Google Scholar
1
-
Mizoguchi M, Hara M. Fingerprint/palmprintmatching identification technology. NEC Techn J. June, 2010;5:18–22.
Google Scholar
2
-
Imaoka H. Face recognition research: beyond the limit of accuracy. The IAPR 2014 Biomet. Lect. The 2014 International Joint Conference on Biometrics.
Google Scholar
3
-
Abaza A, Ross A, Hebert C, Harrison MA, Nixon MS. A survey on ear biometrics. ACM Comput Surv (CSUR). 2013;45(2):22.
DOI | Google Scholar
4
-
Hurley DJ, Arbab-Zavar B, Nixon MS. The ear as a biometric. In Handbook of Biometrics. Ear Identification. Paramont Publishing Company: Springer, Boston, MA. A. Iannarelli, 2008, pp. 131–50.
DOI | Google Scholar
5
-
Arakawa T,Koshinaka T, Yano S, Irisawa H, Miyahara R, Imaoka H. Fast and accurate personal authentication using acoustics. Signal and Information Processing Association Annual Summit and Conference (APSIPA), pp. 1–4, Asia-Pacific: IEEE; 2016.
DOI | Google Scholar
6
-
Yano S, Hokari H, Shimada S. A study on personal difference in the transfer functions of sound localization using stereo earphones. IEICE Trans Fundam Electron Commun Comput Sci. 2000;83(5):877–87.
Google Scholar
7
-
Hiipakka M, Tikander M, Karjalainen M. Modeling the external ear acoustics for insert headphone usage. J Aud Engr Soc. 2010;58(4):269–81.
Google Scholar
8
-
Deng H, Yang J. Modeling and estimating acoustic transfer functions of external ears with or without headphones. J Acou Soc Amer. May, 2015;138(2):694–707.
DOI | Google Scholar
9
-
Akkermans AH, Kevenaar TA, Schobben DW. Acoustics ear recognition for person identification. Automatic Identification Advanced Technologies Fourth IEEE Workshop, pp. 219–22. 2010.
Google Scholar
10
-
Mahto S, Arakawa T, Koshinak T. Ear acoustics biometrics using inaudible signals and its application to continuous user authentication. 26th European Signal Processing Conference (EUSIPCO), pp. 1407–11. IEEE; 2018.
DOI | Google Scholar
11
-
Zhong XL, Xie BS, Glotin H. Head-Related Transfer Functions and Virtual Auditory Display. Soundscape Semiotics-Localization and Categorization; 2014.
DOI | Google Scholar
12
-
Patel VM, Chellappa R, Chandra D, Barbello B. Continuous user authentication on mobile devices: recent progress and remaining challenges. IEEE Sig Proc Mag. 2016;33(4):49–61.
DOI | Google Scholar
13
-
Mermelstein P. Distance measures for speech recognition, psychological and instrumental. Pattern Recognit Artif Intell 1976. 1976;116:374–88.
Google Scholar
14
-
Rajan P, Afanasyev A, Hautamäki V, Kinnunen T. From single to multiple enrollment I-vectors: practical plda scoring variants for speaker verification. Dig Sig Proc. 2014;31:93–101.
DOI | Google Scholar
15
-
Bakare BI, Enoch JD. Investigating some simulation techniques for wireless communication system. IOSR J Electr Commun Engr. 2018;13(3):37–42.
Google Scholar
16
Most read articles by the same author(s)
-
Bodunrin Isa Bakare,
Akujobi Ekeocha,
A Critical Review of Digital Terrestrial Television Broadcasting in Nigeria , European Journal of Electrical Engineering and Computer Science: Vol. 2 No. 7 (2018) -
Dikio C. Idoniboyeobu,
Sunny Orike,
Peace B. Biragbara,
Optimization of a Grid Connected Photovoltaic System using Fuzzy Logic Control , European Journal of Electrical Engineering and Computer Science: Vol. 1 No. 2: NOVEMBER 2017 -
Solomon M. Ekolama,
Sunny Orike,
Promise Elechi,
Preventing Cyber-Fraud in Nigeria’s Banking System Using Fraudaeck-AI , European Journal of Electrical Engineering and Computer Science: Vol. 6 No. 6 (2022) -
Bodunrin Isa Bakare,
T. E. Ngeri,
Sunny Orike,
Comparative Analysis of Signal Strength of Some Cellular Networks in Port Harcourt, Nigeria. , European Journal of Electrical Engineering and Computer Science: Vol. 2 No. 6 (2018) -
Christopher O. Ahiakwo,
Sunny Orike,
Ahuruezemma O. Obioma,
A Computationally Intelligent Power Transmission Expansion Strategy in a Deregulated Energy System , European Journal of Electrical Engineering and Computer Science: Vol. 2 No. 4 (2018) -
B. I. Bakare,
K. J. Biebuma,
Sunny Orike,
Analysis of Signal Attenuation in Digital Terrestrial Television (DTTV) Broadcasting , European Journal of Electrical Engineering and Computer Science: Vol. 4 No. 4 (2020) -
Promise Elechi,
Sunny Orike,
Determination of GSM Signal Loss in Multi-Partitioned Buildings , European Journal of Electrical Engineering and Computer Science: Vol. 1 No. 3: DECEMBER 2017 -
Bodunrin Isa Bakare,
Sunny Orike,
D. I. Oko,
Evaluation and Analysis of the Deployment of Green Communication Technology in GSM , European Journal of Electrical Engineering and Computer Science: Vol. 4 No. 1 (2020) -
Bodunrin Isa Bakare,
Victor T. Taol,
Sunny Orike,
Comparative Analysis of Emitted Magnetic Fields from Liquid Crystal Display (LCD) and Light Emitting Diode (LED) System , European Journal of Electrical Engineering and Computer Science: Vol. 4 No. 2 (2020)