Introduction
Internet of Things (IoT) has come in order to demonstrate the widespread of sensors and actuators to create a ubiquitous embedded architecture where machine-to-machine, human-to-machine, human-to-thing, etc. interactions are commonplace. Nowadays, IoT technology has pervaded into our everyday life through the smartphone, smart office, autonomous vehicle, smart homes, smart transportation and potentially cost-effective approach to the Internet in order to monitor physical systems and critical infrastructure such as remote healthcare application, airplane, etc. Our healthcare is increasingly being made smarter by the patient needs through embedded sensors which integrated into the wearable medical sensors and modern communication. This device can enable continuous and remote monitoring of people health at a low cost. Wearable technologies such as Fitbit, Apple Watch, and Microsoft Band are becoming part of everyday life and they have been integrated into a person’s daily routine which is making it possible to remotely monitor a patient’s health with low-cost sensors. For example, the Microsoft Band can be used to monitor heart rate, 3-axis accelerometer, skin temperature, and galvanic skin response that can be useful for people who are actively involved in data collection in diagnosing to finding the best feasible treatment solution. Apple watch series 4 have added electrodes on the digital crown and on the electrical heart rate sensor on the back of the watch that can measure heart activity (ECG signal) to detect abnormalities in the rhythm. In recent studies [1], [3] on the effectiveness of telehealth, they found that telehealth can be useful to track patients with higher exposure to prevent demanding conditions to save their lives. Moreover, they can reduce the hospitalization cost, the length of time that patients need to stay in the hospital, and mortality due to heart failure [1].
As IoT based Healthcare devices and applications are expected to monitor patient by gathering the vital data and transmitting it to other sources for remote monitoring. If they fail, the patient’s life is at risk. Examples include pacemakers. Thus, it is expected that security and patient privacy have higher priority in driving such technologies. In addition, such IoT medical devices may be joined to global information networks for their access anytime, anywhere. Therefore, the IoT healthcare may be a target of adversarial and can put the patient’s life at risk. Moreover, the National Healthcare Anti-Fraud Association estimates the loss to health care to be about billions of dollars annually in many countries [4], [5] based on fraud, waste, and abuse in healthcare. In order to assure the healthy operation of a Healthcare insurance system, patient verification can prevent false data injection for access control.
One important issue related to IoT-based healthcare application is an incorrect patient identification that leading to inaccurate treatment solution on the patient (such as wrong drug/dose/time/procedure). Therefore, the process of avoiding this happen is costly and have a significant impact on the healthcare provider [7], [14]. In addition, one of the crucial demands in healthcare that is how we can transmit data in a safe way. As it turns out, the existing access control structure relies heavily on traditional authentication, a new range of security attacks becomes feasible. Biometric technology-based patient recognition can significantly provide the desired levels of speed and reliability. Biometrics can also be considered more seamless and convenient, especially for continuous authentication. That being said, it has already been demonstrated that many of the most popular biometric modalities (iris, face, fingerprint, and speech) [12], [13], [17] has been used to overcome some of the aforementioned limitations.
Therefore, there is a need to provide a balance between medical data security and resource consumption of IoT wearable device. In recent years, the objectives of electrocardiogram (ECG) and photoplethysmogram (PPG) monitoring have gone beyond more heart rate and rhythm measurement to the analysis of chronic illnesses including diabetes, heart disease, and sleep disorders among others [18], [19]. Recently, it has been shown that the cardiovascular signal from person-to-person is unique and may be distinctive enough for biometric applications [17], [20]. In order to minimize product costs, remote monitoring of IoT based healthcare is likely to come equipped with low-power computing devices. This implies that biometric and security technologies should consider the constraints of IoT healthcare device. However, biometric protection becomes challenging since the template is frequently compromised, therefore, key derivation functions are used to derive cryptographic keys to protect sensitive information. In addition, IoT device is resource constrained, thus reducing storage complexity are needed. As a result, embedding identity information into a binary template can be applied to reduce the storage and matching complexity. However, biometric key generation may suffer from environment/measurement noise, internal instability, and so on which could result in errors during key generation process and serious security and privacy threat such as intrusion attack and cross-matching the template [48], [27] and make it impossible for even the system owner to access/use the system. Due to a large intra-subject variance during multiple acquisitions of the same biometric trait, it is a challenging task in order to overcome this issue for biometric template protection. To mitigate this challenging task, we propose a novel biometric key generation from noisy data to mitigate or eliminate the intra-subject variance while preserving privacy and generating long keys by developing a statistical approach.
There has been a number of technique deployed for biometrics key generation using in the literature. Monrose et al. [42] and Teoh et al. [43] applied a scheme that quantizes each background feature space into two intervals where each interval is labeled with bit “0” or “1” based on a fixed threshold. A feature value that falls into an interval is mapped to the corresponding 1-bit output label. Kelkboom et al. [2] developed a framework to estimate the genuine and imposter bit error probability mass functions by measuring hamming distance. Drozdowsk et al. [44], the authors applied deep learning for feature extraction followed by encoding schemes (Linearly Separable Subcode) which exhibit full-ideal and near-ideal separability capabilities, respectively. Chen et al. [36], [45] proposed a generic bit allocation algorithm scheme using pairwise adaptive phase quantization and long-short pairing strategy. Lim et al. [37], [46] developed a DROBA-based approach, to improve the biometric performance of the binarized representation of facial features based on bit statistics (reliability and discriminability). Osadchy and Dunkelman [31], explore the existing security and privacy of feature extraction and binarization processes where they found that the most important part of biometric protection is how to extract features with high accuracy. Kaur and Khanna [40] proposed cancelable biometric template protection to address security and privacy. In 2019, cancelable ECG biometric has been proposed by Wu et al. [9] to address the biometric privacy properties such as revocability, unlinkability, and irreversibility.
One of the most important requirements in bio key is the entropy [8] in order to guarantee resistance against attacks. However, it is a challenging task to simultaneously achieve high key entropy and high key stability. Usually, error correction code (ECC), and fuzzy extraction will be applied [9], [11]. However, fuzzy extraction has potential drawbacks including high computational cost, increased area overhead, and vulnerability to side channel attacks [10] which again it is not feasible for IoT application. The best method to protect user biometrics is a hash function. However, since the biometric is noisy, therefore, the hash function cannot be directly used. To address this issue, the quantization technique where it can map noisy samples to a unique vector then protected by hash function. There is limitation associated with existing quantization algorithm, where the formulation is a generic characteristic and because of that, the attacker has appropriate knowledge of transformation algorithm and parameters, which resulted in linkability of the system. In fact, most of the limitation due to reliability and entropy of the system. In addition, the binarization technique is generic, however, our approaches adapt to the population itself (to preserve entropy) while also being adaptable to individual users (to improve reliability), without revealing any information about the system’s users.
The difference between this work and the previous work is that the tunable parameters which allow an amount of intra-user variability among multiple acquisitions of the same biometric trait are dynamically updated by optimizing local minimum of the proposed algorithm. The major goal of this work is to demonstrate how the proposed algorithm can address the intra-variability of ECG-based biometric while multiple biometric modalities such as PPG, iris, and fingerprint is conducted. We also conducted the proposed algorithm with the challenging ECG database with very noisy signal under multiple data acquisition session. Moreover, none of the existing work implements their proposed algorithm, in this work, our algorithm for the first time has been implemented to evaluate it in terms of the feasibility of the IoT application.
In this paper, we focus on how to incorporate a new biometric key generation/authentication framework that provides secure and reliable access control in resource-constrained environments for remote healthcare application. Our main contributions can be summarized as follows:
Noise-aware Feature Quantization: We propose noise aware interval optimized mapping bit allocation (NA-IOMBA) to select and quantize biometric features on a user-to-user basis. In other words, only the most robust features are selected for each user, thus avoiding unnecessary post-processing costs. Noise models are used to predict the impact of noise and further reduce error correction costs and enrollment times.
Multiple Modalities: We perform experiments on four biometrics modalities (ECG, PPG, fingerprint and iris scan) to demonstrate the effectiveness of NA-IOMBA. Experimental results show that NA-IOMBA increases the length, entropy, and robustness of keys generated from multiple biometric modalities. Further, ECG-based authentication performs the best among the candidates. Thus, we take a closer look at its performance throughout the rest of the paper.
Synthetic ECG signals: One of the biggest challenges in electrocardiogram (ECG) based authentication and/or key generation lies in how to obtain noisy ECG data. Put simply, it is impossible to take raw ECG measurements from a large population under all possible test conditions. In this paper, we discuss how dynamical models can be used to generate synthetic ECG signals for common noise sources as well as activity/stress.
Biometric Authentication Cost and Enrollment Time: By incorporating noise models into feature selection, certain denoising/filtering steps can be avoided by NA-IOMBA. Implementation results from Xilinx Zynq-7000 show that, even with less resources (65% power and 62% utilization), NA-IOMBA offers higher reliability. In addition, the number of enrollment samples and enrollment time can be reduced since biometric noise and different conditions are modeled in NA-IOMBA. This behavior is confirmed with experiments using multi-session data and training data from non-ideal conditions.
Machine Learning vs. Quantization: Machine learning results are compared to our quantization schemes on multiple biometric modalities. We demonstrate that the accuracy of the proposed approach performs just as good, and in most cases better than machine learning based authentication.
Security Analysis based on Entropy and Key Randomness: To evaluate the correctness and security of our proposed biometric key generation, we inspect it using a set of standard statistical suite test (NIST). The results of these test on our proposed method passes the randomness test.
The rest of the paper is structured as follows. In Section II, the noise-aware quantization framework is described. Denoising overhead reduction, ECG synthetic model, ECG noise modeling, and ECG stress modeling, are described in Section III. Section IV presents NA-IOMBA case studies including the comparison of multiple biometric modalities, ECG long-term feasibility, the impact of heartbeats on training, a fixed threshold in IOMBA, and machine learning vs quantization approach. The analysis of noisy ECG signal, FPGA implementation of our approach, and reduction in enrollment times are presented in Section IV. Finally, the conclusion is drawn in Section VI.
Noise-Aware Quantization Framework
A. IOMBA Based Biometric Key Generation
Biometric systems that operate using any single biometric characteristic have several limitations such as noise in sensed data, intra-class variations (the biometric data acquired from an individual during authentication may be very different from the data that was used to generate the template during enrollment), and distinctiveness (biometric trait is expected to vary significantly across individuals). Noise in biometric authentication and recognition applications can be tolerated to some extent; however, not a single error can be tolerated in key generation applications. To address this limitation, we have previously developed the interval optimized mapping bit allocation (IOMBA) scheme for biometric key generation [21].
In short, IOMBA tunes the biometric key generation process to each user rather than relying on a generic approach for all users. The performance is controlled by two parameters,
The major steps of the IOMBA are described as follows.
1) Data Pre-Processing
The signals from the population are pre-processed to remove noise followed by feature extraction. The feature elements from the same location are extracted from the population and normalized into a standard normal distribution. The same normalization parameters are later exploited to normalize the corresponding feature elements of each subject. Further, a decorrelation step can be applied to the distributions as well. Note that our approach can work with any biometric provided it produces statistically independent and Gaussian features in some representation - any feature that does not meet these requirements will be discarded.
2) IOMBA Margin Calculation from Population Statistics
IOMBA quantizes each feature into a different number of bits. 2 bit quantization is illustrated in Fig. 1. The population probability density function (\begin{align*}&\int _{T}^{\infty } PDF_{\text {pop},f} \le \beta, \quad \text {if} ~ \mu _{f,i} < T \tag{1}\\&\int _{-\infty }^{T} PDF_{\text {pop},f} \le \beta \cap \int _{0 }^{\infty } PDF_{f,i} \le \beta, \quad \text {if} ~ T < \mu _{f,i} < 0 \\{}\tag{2}\end{align*}
\begin{align*}&\mu _{f,i} \leq \mu _{00}, \quad \text {if} ~\mu _{f,i} < T \tag{3}\\&\mu _{\text {left},01} \leq \mu _{f,i} \leq \mu _{\text {right},01}, \quad \text {if} ~ T < \mu _{f,i} < 0 \tag{4}\end{align*}
Illustration showing how IOMBA optimizes quantization of a single feature into two bits.
Moreover, the entropy should be large enough to guarantee resistance against attacks. To address this problem,\begin{align*} P_{00}=&\int _{-\infty }^{\mu _{00}} PDF_{\text {pop},f}dx \tag{5}\\ P_{01}=&\int _{\mu _{left,01}}^{\mu _{right,01}} PDF_{\text {pop},f}dx \tag{6}\end{align*}
\begin{equation*} \frac {P_{01}}{P_{00}}\le \alpha \tag{7}\end{equation*}
3) Enrollment for Key and Helper Data Generation
Essentially, IOMBA personalizes a biometric system to each user. For each subject, the key generation framework utilizes the above boundaries to determine whether each feature element is good for generating key bits or not. The least reliable features of a user which do not fulfill the above constraints are discarded. The helper data for each subject consists of the following: (i) the index of the reliable features selected for the user, (ii) the number of bits each feature can be quantized into, and (iii) the normalization parameters for each feature. If an error still exists, then error correction based helper data and an error correction module can also be added in the next step.
4) Key Regeneration
The user presents his/her biometric, features are extracted, and the helper data stored on the system is used to eliminate the unreliable features and then quantize the reliable features to regenerate the key.
In IOMBA, feature extraction, helper data, and key regeneration steps are different for each user. Reconfigurable hardware is, therefore, a natural candidate for implementing IOMBA because each step can be tailored uniquely to the user, thereby avoiding certain processing costs. Results show that the proposed approach for ECG key generation achieves 28% improvement of reliability and four times longer key size in the worst case scenario compare to our previous work [21].
B. Noise Aware IOMBA (NA-IOMBA)
In the original IOMBA, the standard deviation in user PDFs was fixed for each feature in a worst-case manner. Feature selection was therefore pessimistic and resulted in shorter keys. In fact, when the biometric data is not noisy, IOMBA approach works quite well. Estimating the standard deviation, especially for continuous biometrics (keystroke dynamics, ECG, etc.), is nontrivial since the biometric would need to be collected at all conditions and types of noise. In most applications, such as IoT, the enrollment process would be too long for users to tolerate. In addition, the impact of noise is substantially affected by the type and amount of pre-processing. In resource-constrained scenarios, it would be better to eliminate pre-processing steps which are costly and energy consuming to perform. To accommodate these issues, we propose the noise-aware IOMBA framework in this section and demonstrate its benefits in Section IV. NA-IOMBA is a variant of IOMBA that incorporates models to predict the impact of different noise sources, noise scales, and pre-processing steps on biometric key generation technique. In short, all IOMBA steps are performed with one major change; Margins and boundaries are recomputed based on more accurate estimates of user feature standard deviation. Basically, features that are modeled as less (more) susceptible to noise will, therefore, be given smaller (larger) margins than IOMBA. To better understand NA-IOMBA, we have illustrated it in Fig. 2. First, optimal margins and thresholds are calculated based on inputs
Block diagram show for NA-IOMBA margin reconstruction key calculation from noisy biometric.
In order to determine a noise model in NA-IOMBA (see Fig. 2),
The following are the impacts and benefits of this approach:
Impact on key length: If a margin for a feature increases, it could result in the feature being selected and/or longer bit lengths compared to IOMBA. If a margin for a feature shrinks, it could result in the feature no longer being selected and/or shorter bit lengths compared to IOMBA.
Impact on key reliability: If noise samples or models are accurate, the reliability of key should improve regardless of whether or not the key length shrinks/grows.
Impact on cost and enrollment time: There are three ways that overheads can be reduced. First, error correction costs tend to increase nonlinearly. By improving key reliability, error correction hardware can be substantially reduced. Second, certain denoising/filtering steps can be removed provided that NA-IOMBA accurately estimates the noise appearing in features without them. Third, the number of enrollment samples and enrollment time can be reduced if noise over time and different conditions can be modeled. This can be particularly important for continuous physiological signals, like ECG and PPG, which can be impacted by so many different conditions, e.g., exercise, stress, and food/drink/drug consumption.
Incorporation of Real Noise Models
In this next section, we discuss how NA-IOMBA can be further improved by incorporating actual noise models. We take ECG as an example since there are a variety of models available in the literature for it. In order to determine the sensitivity of ECG key generation based on these feature extraction, the noisy ECG signal with different variances (SNR) is applied. To view the impact of each noise source, synthetic ECGs are generated and not pre-processed to remove the noise.
A. ECG Synthetic Model
We adopt the non-linear dynamical model proposed by McSharry et al. [22] to extract parameters from an ECG and generate synthetic ECGs. The aim of this approach is to provide a standard realistic ECG signal with known characteristics from ECG (Fig. 3 (a)), which can be generated with specific statistics thereby facilitating the performance evaluation of a given technique. McSharry et al.’s model uses three ordinary differential equations. It consists of a circular limit cycle of unit radius in the \begin{equation*} \begin{cases} \dfrac {dr}{dt}=r(1-r) \\ \dfrac {d\theta }{dt}=\omega \\ \dfrac {dx}{dt}=-\sum _{i\in {P,Q,R,S,T}}a_{i}\Delta \theta _{i} exp\left[{-\dfrac {\Delta \theta _{i}^{2}}{2b_{i}^{2}}}\right] -(z-z_{0})\\ \end{cases} \tag{8}\end{equation*}
Typical normal ECG signal. (a) One beat normal ECG signal with fiducial point, (b) trajectories several cycles of the ECG phase-wrapped in the Cartesian coordinates.
B. ECG Noise Modeling
The three main types of noise sources in raw ECG signals are (1) muscle artifacts (MA) which occur due to electrical activity of muscles; (2) baseline wander (BW) caused by body movement; and (3) electrode movement (EM) due to poor contact to the sensor. We adopt the noise model from [24], [25].
Fig. 4 presents the experimental protocol of NA-IOMBA scheme for noisy ECG signals. In noisy ECG key margin reconstruction, dynamic model parameters (
Block diagram show for NA-IOMBA margin reconstruction key calculation from stressed ECG.
Time-varying auto-regressive (AR) parametric models are applied to generate realistic ECG noise which follows the non-stationary characteristics and the spectral shape of real noise. The parameters of this model are trained by using real noises such as NSTDB [26]. To estimate the time-varying AR parameters, a standard Kalman Filter (KF) is used [28]. For the time series of \begin{equation*} y_{n}=-\displaystyle \sum _{i}^{p} a_{n}(i)y(n-i)+v_{n} \tag{9}\end{equation*}
C. ECG Stress Modeling and Reducing Enrollment Times
Another concern that restricts the use of ECG for biometric authentication is the variability of heart signal within the subjects. Heart rate varies with an individual’s physiological and mental conditions. Stress, excitement, exercise, and other working activities may have an impact on the heart rate and can elevate it. These variations are likely to affect the reliability of ECG based key generation or authentication. A previous study [29] about the influences of physical exercise indicates that the ECG morphology is affected by exercise/stress. In other words, each peak (P, QRS, T) in the ECG may increase/decrease in amplitude, temporal location, etc. To cover the impact of stress/exercise on the reliability with different scenarios, we vary the dynamical model parameters by type (
Each dynamical model parameter is scaled by a factor (0.9–0.5). Fig. 5(a-c) show how the ECG changes when scaling
Impact of ECG signal by decreasing dynamical model parameters: (a)
NA-IOMBA Case Studies
The improvements gained by the proposed approach will be initially demonstrated in this section.
A. Comparison of Multiple Modalities Using IOMBA and NA-IOMBA
In this section, we present a comprehensive performance evaluation of biometric-based key generation. We apply our approaches (IOMBA & NA-IOMBA) on four biometric modalities: ECG, PPG, iris, and fingerprint. Table 1 shows the methodologies, databases, and train/test sizes, that have been employed for multiple biometric modalities.
1) Electrocardiogram (ECG)
ECG is a recording of the electric potential, generated by the electric activity of the heart. The ECG recordings of 52 subjects from the PTB database [15] are used in this paper. We employ low and high pass finite impulse response (FIR) filters with cut off frequencies 1Hz-40Hz to eliminate noise associated with an ECG signal (see Fig. 6). Normalize-Convoluted Normalize (NCN) is used as the feature extraction technique [24].
ECG Pre-processing. Raw data is acquired from database and then butterworth filtered; the individual heartbeat waveforms are segmented by their R peaks.
2) Multi-ECG
The ECG-ID database at the PhysioNet [15] was used as a multi-ECG in our experiments. Each raw ECG record was acquired for about 20 seconds with a sampling rate of 500 Hz and 12-bit resolution. First two records acquired from the same day were used for each subject. The database consists of 310 one-lead ECG recording sessions obtained from 90 volunteers during a resting state. The number of sessions for each volunteer varied from 2 to 20 with a time span of 1-day to 6-months between the initial and last recordings. This study utilizes the same pre-processing aforementioned for ECG PTB database.
3) Photoplethysmogram (PPG)
The photoplethysmogram (PPG) is a biomedical signal that estimates volumetric blood flow changes in peripheral circulation using low-cost and simple LED-based devices typically placed on the fingertips. In order to evaluate the efficiency of the PPG biometric authentication based on IOMBA and NA-IOMBA, a publicly available Capnobase dataset [30] with 42 subjects was used.
Pre-processing: Typically PPG signal is interfered by several noise sources including baseline wander (BW), motion artifact (MA), and respiration. To remove this artifact, we applied a third order Butterworth band pass filter with cutoff frequency 1Hz-5Hz as can be seen in Figure 7. Then, filtred PPG signal is passed by Pan Tompkins algorithm to extract systolic and diastolic peaks.
Feature Extraction In this section the feature extraction methods based on non-fiducial approach which will be discussed below.
PPG Non-Fiducial Features: Since the PPG signal is effected by noise, we applied non-fiducial feature extraction based on overall morphology of waveform rather than specific fiducial points. To this end, wavelet transform technique which is a very popular technique for biomedical signal processing due to the fact that it is lightweight and capable of providing time and frequency information simultaneously. The PPG signal will be passed through a series of low and high pass filters by decomposing it into various scales.


Plots illustrating PPG signal from the database for (a) filtered PPG signal where the baseline and other artificial noise sources have been removed. (b) Extracted systolic and diastolic peaks by using pan tompkins algorithm.
The wavelet coefficients can be obtained by taking the inner product:\begin{align*} V_{\phi }[j_{0},k]=&\frac {1}{\sqrt {M}}\displaystyle \sum _{n}~PPG[n]\phi _{j_{0},k}[n] \tag{12}\\ W_{\psi }[j_{0},k]=&\frac {1}{\sqrt {M}}\displaystyle \sum _{n}~PPG[n]\psi _{j,k}[n]\quad j_{0} \le k \tag{13}\end{align*}
4) Iris Scan
The iris is called the colored ring around the eye pupil. According to research, the human eye is one of its most unique characteristic that can be used for biometric recognition. To evaluate the iris key generation based on IOMBA and NA-IOMBA, we first take the iris images from available CASIAv1-Interval iris database [16]. In the pre-processing stage, Canny edge detection is used to enhance the iris outer boundary due to the eyelid or eyelashes. Detecting the boundary of iris and sclera is applied for segmentation. Then, the iris is converted to a 2D matrix represented by (
5) Fingerprint
Fingerprint is a unique pattern of ridges and valleys that have been used widely in biometric application. The fingerprint used for biometric key generation in the study is taken from FVC2004 database [34]. This data set, containing 8 images of 110 users. Images are aligned according to a standard core point position, in order to avoid a one-to-one alignment. In this paper, the Gabor filter is used to directly extract fingerprint features from gray level images [35]. The raw measurements contain two categories: the squared directional field in both
The quality of generated keys is compared by four evaluation criteria: reliability, entropy, key length, and equal error rate (EER). The metrics used for each are discussed below and a brief comparison of the above biometric modalities based on IOMBA and NA-IOMBA is provided in Tables 2. In the table, ‘max’, ‘ave’, and ‘min’ columns correspond to the highest value (best case) achieved among all users, the average of keys across the users, and the lowest value obtained among all users. Note that for this initial comparison, the noise model in NA-IOMBA is adopted from the standard deviation of enrollment measurements and the standard deviation is adjusted on a per feature basis. A more elaborate model will be used for ECGs in the next section.
6) Reliability
Reliability of key generation represents the stability of keys over time. If all bits generated by the biometric of an individual are equal to the key produced in enrollment, it can be considered reliable. We adopt the reliability metrics from [21]. As can be seen in Table 2, improvements in reliability are achieved by applying the NA-IOMBA technique. Average and worst cases improve by 2% and 9.7% on average for all modalities compared to IOMBA. Among all modalities, fingerprint attains the largest percentage of improvements (3.8% and 26.9% on average and worst cases). However, ECG has the best performance for both NA-IOMBA and IOMBA.
7) Entropy
To measure key randomness, we calculate the min-entropy. We adopt the min-entropy metrics from [21]. As shown in Tables 2, the min-entropy of ECG signal is higher than iris, PPG and fingerprint, however, under NA-IOMBA technique, there is a huge entropy improvement for iris, fingerprint and PPG compared to IOMBA results. For example, the min-entropy is not only improved by 35% at the minimum case for fingerprint based on NA-IOMBA but also increased by 8% on average.
8) Key Length
Since certain features may be reliable for some users and unreliable for others, our approach will only use reliable features from each individual. Thus, the key length per person may change. As can be indicated in Table 2, the key length of ECG, PPG, iris, and fingerprint based on IOMBA are 668, 114, 66, and 835, respectively. When NA-IOMBA is applied, the average key length for ECG, PPG, iris, and fingerprint increases by 30%, 6%, 88%, and 27%. Fingerprint obtains the largest key for both IOMBA and NA-IOMBA while PPG obtains the smallest.
a: Equal Error Rate (EER)
EER is a step in the biometric security system that determined by the threshold values for its false reject rate (FRR) and false accept rate (FAR). FAR refers to the rate at which an impostor user incorrectly identified as a genuine user, while FRR refers to the rate at which a genuine user incorrectly identified as an impostor user. An ideal biometric system, the EER is close to zero. To better understand how we can calculate EER from our IOMBA/NA-IOMBA, Fig. 8 is demonstrated. As can be seen in this figure, the probability density function (PDF) of the impostor is plotted with the red solid line and PDF of genuine is plotted with the blue solid line. Minimizing the EER is equivalent to minimizing two areas shaded as indicated in Fig. 8 corresponds to the FRR and FAR. Improving the biometric system authentication performance requires diminishing in this area.
According to Poh and Bengio [38], we can consider \begin{align*} FRR(\theta)=&\int _{-\infty }^{\theta } P(y| x \in x_{G}) dy \\=&\int _{-\infty }^{\theta } \frac {1}{\sigma _{G}\sqrt {2\pi }}\exp \frac {-(y-y_{G})^{2}}{2\sigma _{G}^{2}} dy \\=&\frac {1}{2}+\frac {1}{2}erf\frac {\theta -\mu _{G}}{\sigma _{G}\sqrt {2}}\tag{14}\\ FAR(\theta)=&\int _{\theta }^{\infty } P(y| x \in x_{I}) dy \\=&1- \int _{-\infty }^{\theta } P(y| x \in x_{I}) dy \\=&1-\left({\frac {1}{2}+\frac {1}{2}erf\frac {\theta -\mu _{I}}{\sigma _{I}\sqrt {2}}}\right) \\=&\frac {1}{2}-\frac {1}{2}erf\frac {\theta -\mu _{I}}{\sigma _{I}\sqrt {2}}\tag{15}\end{align*}
\begin{equation*} EER= \frac {1}{2}-\frac {1}{2}erf\frac {\mu _{G}-\mu _{I}}{(\sigma _{I}+\sigma _{G})\sqrt {2}}\tag{16}\end{equation*}
For one session ECG signal (PTB database), we have constructed an average of 668 and 953 key bits with 5.08% and 1.25% EER based on IOMBA and NA-IOMBA. We achieve an EER around 15.46% based on IOMBA for PPG database while by incorporating NA-IOMBA model, the EER is decreased by 93%. In addition, the EER is decreased by 81% and 89% for fingerprint and iris database after employing NA-IOMBA approach.
9) Machine Learning vs Quantization
The key element of traditional biometrics authentication system is driven by machine learning, deep learning that makes it possible to drive the decision making processes based on given input data from original biometric templates or features that has been extracted during pre-processing to deal with the intra-class variation of biometric measurement. As indicated earlier, we propose a noise aware quantization approach to enhance the system security and user privacy on the resource-constrained IoT application. Existing quantization techniques invariably result in loss of some discriminatory information leading to lower recognition performance. Moreover, we compare the performance of IOMBA/NA-IOMBA with other state-of-art methods for biometric authentication. In order to compare our approach with machine learning, we apply a one-class support vector machine (SVM) technique for classification as a biometric authentication matching. Similar to most of the existing works for biometric authentication, we evaluated the accuracy and EER by applying machine learning techniques. Receiver operating characteristic (ROC) curves are shown in Fig. 9 for the biometric modalities considered in this paper. As shown in this figure, the average accuracy of ECG, multi-ECG, PPG, Fingerprint, and Iris are 98.35%, 94.09%, 98.26%, 92.63%, and 86.90%; the average EERs are 1.75%, 7.99%, 1.81%, 8.83%, and 20.68%, respectively; while performing five-fold cross-validation.
Interestingly, we were able to approach a performance of 98.76%, 98.47%, 99.04%, 98.36%, and 98.76% in accuracy and 1.25%, 1.43%,.97%, 1.7%, and 1.96% in EER based on NA-IOMBA method for ECG, Multi-ECG, PPG, fingerprint, and iris, respectively, which is significantly better than machine learning results based on Table 2. In other words, we perform as well and in many cases better than off-the-shelf machine learning. In addition, quantization is a key step in signal/image system, especially in the big data age, because quantization not only can decrease storage costs but also can accelerate the detection speed in a large-scale database. Finally, machine learning techniques require many training samples per subjects in order to determine user-specific feature while in NA-IOMBA, we have shown that it is independent of training samples.
10) Cost
The effectiveness of biometric technology is dependent on how and where it is used. Each biometric modality has its own strengths and weaknesses. Today, an ECG or a PPG sensor costs around 20 when ordered in large quantities, thus has a marginal cost of embedding into a biometric system. However, fingerprint and iris scan costs about 70 and $280, respectively. Note that the hardware cost is normalized into 1 in order to make it simpler to consider as metrics. In that case, if the value is lower than 1; meaning a more expensive sensor.
Figure 10 ranks four common technologies (ECG, PPG, iris scan, finger scan) according to four criteria: reliability, entropy, key length, and hardware cost. The maximum point in each length indicates the best candidate for that specific criteria. As can be seen in Figure 10, the average reliability of ECG is 99.76% belong to the maximum point of the plot while the average reliability of iris is 98.36% which belongs to the minimum point of the plot. In addition, the entropy of ECG is on the maximum point of the plot. For the cost, PPG is the best choice among all biometric modalities. Furthermore, the key length of the fingerprint is higher than other biometric modalities which made it become at top of the plot, although the ECG signal is following this with a small margin. ECG appears to be the best candidate for all the criteria when applying NA-IOMBA. However, it is worth noting that ECG still suffers from several other issues (impact of noises, stress condition, and aging) that need to be tackled in order to make this candidate an even stronger selection. The rest of this paper more closely examines the impact of noise on ECG with and without incorporation of noise models.
11) ECG Long-Term Feasibility
We studied the impact of long-term variability of ECG signal called multi-session ECG signal from various days and obtained the number of training and testing heartbeats on ECG biometric performance. To this end, we have considered the case whereas the training feature sets contain subsets of ECG beats that has been extracted from a session, while the testing data come from another session. In this paper, the effect of changes on the ECG signal over time has been investigated. As can be seen in Table II, the performance of ECG based on IOMBA technique has degraded while NA-IOMBA were incorporated to reduce the loss of performance. These results demonstrate that the evolution of the multi-session ECG signal is generated in a long time interval based on NA-IOMBA approach. All in all, results seem to confirm the long-term stability of reliability and EER. This is an essential condition since the threshold and boundaries in NA-IOMBA and IOMBA are the pre-configured parameter of the system, so if these results vary with time, the system will be incompatible and less useful for IoT healthcare applications.
12) Impact of Training Set Size
To understand the utility of ECG as a biometric, we examine the ECG biometric performances using a different number of training heartbeats. This helps to understand the sample size effect on ECG biometric performance for both IOMBA and NA-IOMBA. Fig. 11 shows the key reliability for testing on a session, when training is performed on a different number of heartbeats. In general, the variation in key reliability increases as the number of samples in the training increases. This is in agreement with our expectation. The average key reliability is increased when the number of training heartbeats or segments increases for IOMBA technique. However, the average key reliability for NA-IOMBA is already saturated at 5 samples. We observe the NA-IOMBA is independent of training heartbeats; hence enrollment time is reduced.
Reliability of IOMBA vs NA-IOMBA with different numbers of heart beats used during training.
The maximum, average, and minimum accuracy value for a training sample size
Case Studies on ECG Using Noise Models
In this section, we follow the approach described in Section III be incorporating ECG noise models. Figure 12 is a box plot showing the reliability rate versus input signal-to-noise ratio (SNR) for ECG based on IOMBA without denoising filters, IOMBA with denoising filters, and NA-IOMBA without denoising filters. The SNR during the noisy segments was set to 30dB, 20dB, 10dB, 5dB, 0dB, and −5dB separately. Figures 12 (a-d) indicate the impact of each noise source (BW, EM, and MA, and mixed of them) on the reliability. In the context of ECG noise levels, the lower SNR provides more fluctuation on ECGs (higher intra-class variation). Intuitively, there appears to be an inverse relationship between the level of generated noise and reliability. Among all these noise sources, MA and EM are the strongest noise sources and have an enormous impact on key reliability when IOMBA is applied with and without filtering. As one would expect, IOMBA with filtering obtains better key reliability than without. However, for NA-IOMBA, the reliability is never less than 96.7% even at worst case (mixed noise with -5dB). In contrast, there is considerable degradation beyond 20dB by using IOMBA with/without filtering (63% reliability). As mentioned earlier, ECC increases nonlinearly with the number of errors. NA-IOMBA has very high reliability compared to IOMBA, and therefore ECC will inherently consume less overhead. The cost reduction is discussed further below.
IOMBA, filtered IOMBA, and NA-IOMBA Keys reliability rate vs input SNR; impact of (a) BW noise, (b) EM noise, (c) MA noise, and (d) mixed noises, on the reliability.
A. FPGA Implementation of an IOMBA
In this paper, finite impulse response (FIR) is designed using Simulink in the Xilinx System Generator. The Xilinx System Generator tool is a high-level tool for designing high-performance DSP systems and enables us to integrate Xilinx with Simulink. To implement noise reduction using FIR filter, an FDA tool has been applied to design a filter for required specifications. Pan Tompkins algorithm is applied for detecting ECG R peak and segmentation. Finally, the NCN feature extraction technique has been considered for key generation. Table 3 shows implementation of the ECG key generation using IOMBA with filtering and NA-IOMBA without filtering on the Xilinx Zynq-7000. In IOMBA case, 11% of total flip-flops (FF), 20% of all available Look-up tables (LUTs), and 71% of the DSP slice are used while in NA-IOMBA consumes only 1% of FFs, 4% of LUTs, and 10% of DSP are utilized. In addition, IOMBA consumes 113 mW power while NA-IOMBA consumes only 39 mW power. As a result, by saving overall overhead while applying NA-IOMBA, we are able to add ECC in the IoT devices to reconstruct the errors. In fact, NA-IOMBA allows hardware to adapt pre-processing, feature extraction, post-processing, and error correction overheads on a user-to-user basis.
B. Stress/Exercise Results
Fig. 13((a-c) indicate the reliability of each dynamic parameters on the P, QRS, and T waves based on IOMBA, respectively. The reliability of NA-IOMBA is shown in Fig. 13(d-f) where IOMBA margins are re-optimized assuming dynamic parameters scaled to 0.7. For IOMBA, the T wave is impacted by all parameters (
Impact of stress/exercise on the reliability of; IOMBA for changing (a) P wave, (b) QRS wave, (c) T wave; NA-IOMBA for changing, (d) P wave, (e) QRS wave, and (f) T wave of ECG.
Security Analysis
Frankly speaking, measuring the entropy of key bits is one measures the strength of a cryptographic, which quantifies the amount of uncertainty in the key from an adversary’s standpoint [39]. We also evaluate the security requirements for biometric key generation based on key randomness.
A. Key Randomness Based on NIST Test
The Bio-key generated through our scheme need appear as random to an adversary which has access to the auxiliary information. To evaluate the randomness, the NIST statistical test suite is applied [41]. The NIST Test Suite (NTS) is a statistical test consisting of different types of tests to evaluate the randomness of binary sequences. Each statistical test is employed to calculate a p-value that shows the randomness of the given sequences based on that test. If a p-value for a test is determined to be equal to 1, then the sequence appears to have perfect randomness. A p-value ≥ 0.01 (normally 1%) means that sequence would be considered to be random with confidence of 99%. The results of 15 performed NIST tests on our proposed key generation are shown in Fig. 14 that showing proposed key generation passes the randomness tests.
NIST statistical tests Suite results for the randomness tests of proposed biometric key generation.
B. Entropy Analysis
For the highest level of security, the keys used should be drawn from a uniform distribution since this will result in the greatest number of brute force attempts needed to identify the correct key by an attacker. Entropy is one of the most widely used measures of such randomness or unpredictability in keys. The entropy should be large enough to guarantee resistance against attacks. In this paper, each key has been generated from our proposed algorithm and the randomness of each bit is estimated by calculating the min-entropy, which is the most conservative way of estimating entropy or unpredictability. In this paper, the min-entropy of a feature \begin{equation*} H_{\infty }(k)= -\xi \log _{2}(\underset {i}{max}\{P_{i}(k)\})\tag{17}\end{equation*}
C. Template Irreversible
The requirement of a template being irreversible is that the dimensionality of feature set should be smaller than dimensional subspace [6]. In our approach, each user has its own template due to a user-specific algorithm. In other words, none of all feature vector will be selected for key generation. As can be seen in Fig 15, the number of feature sets that have been accepted in our algorithm versus the number of total feature sets is different. In addition, if a feature
Conclusions
In this paper, the interval optimized mapping bit allocation (IOMBA) scheme for the key generation was improved by incorporating noise models. It was demonstrated that keys generated from ECG, PPG, iris, and fingerprint by noise-aware IOMBA are more reliable, longer, and higher entropy than noise-free IOMBA. Furthermore, by using more advanced noise models for ECG, overhead from denoising filters and error correction could be further reduced by 62% without additional enrollment measurements. Moreover, we analyze our model under the multiple-session ECG signal where a single session is used to train our model and testing data come from different sessions. Besides, we also compared the performance of our noise-aware biometric quantization framework with other state-of-art machine learning techniques. In the future, we intend to develop an end-to-end framework ensuring to protect our model from vulnerabilities to attacks. In addition, revocability, unlinkability, and irreversibility will be an interesting and challenging work that deserves our collective efforts in the future.