Keywords

1 Introduction

EEG signals have been widely studied since the beginning of the last century, mainly for clinical applications, to investigate brain diseases like Alzheimer, epilepsy, Parkinson and many others. Specifically, EEG signals, acquired by means of scalp electrodes, sense the electrical activity related to the firing of specific collections of neurons during a variety of cognitive tasks such as response to audio or visual stimuli, real or imagined body movements, imagined speech, etc. The most relevant cerebral activity falls in the range of \([0.5,40]\) Hz. In details five wave categories have been identified, each associated to a specific bandwidth and to specific cognitive tasks: delta waves (\(\delta )\) [0.5,4] \(\text {Hz}\) which are primarily associated with deep sleep, loss of body awareness, and may be present in the waking state; theta waves \((\theta )\ [4,8]\,\text {Hz}\) which are associated with deep meditation and creative inspiration; alpha waves (\(\alpha )\ [8,13]\,\text {Hz}\) which indicate either a relaxed awareness without any attention or concentration; beta waves (\(\beta ) \ [13,30]\,\text {Hz}\) usually associated to active thinking; gamma waves (\(\gamma ) \ [30,40]\,\text {Hz}\) usually used to locate right and left side movements.

In the last decades, the brain activity, registered by means of EEG, has been heavily employed in brain computer interfaces (BCI) [1] and more recently in brain machine interface (BMI) [2] for prosthetic devices. In the last few years EEG signals have also been proposed to be used in biometric based recognition systems [3].

EEG signals present some peculiarities, which are not shared by the most commonly used biometrics, like face, iris, and fingerprints. Specifically, brain signals generated on the cortex are not exposed like face, iris, and fingerprints, therefore they are more privacy compliant than other biometrics since they are “secret” by their nature, being impossible to capture them at a distance. This property makes EEG biometrics also robust against the spoofing attack at the sensor since an attacker would not be able to collect and feed the EEG signals. Moreover, being brain signals the result of cognitive processes, they cannot be synthetically generated and fed to a sensor, which also addresses the problem of liveness detection. Also, the level of universality of brain signals is very high. In fact people with some physical disabilities, preventing the use of biometrics like fingerprint or iris, would be able to get access to the required service using EEG biometrics.

Table 1. State of the art on EEG biometrics using a resting state protocol.

However, the level of understanding of the physiological mechanisms behind the generation of electric currents in the brain, not yet fully got, makes EEG a biometrics at its embryonic stage. Nevertheless, some preliminary, but promising, results have already been obtained in the recent literature, see for example [4, 5, 1214] where a review on the state of the art of EEG biometrics is also given, and [15]. Due to the early stage of research dealing with EEG as biometrics, currently, the deployment of convenient and accurate EEG based applications in real world are limited with respect to well established biometrics like fingerprints, iris, and face. However the brain electrical activity has already shown some potentials to allow automatic user recognition. Answers to practical and theoretical questions addressed for the development of a usable system can be found in [16] where promising results are obtained from the implementation of a portable EEG biometric framework for applications in real world scenarios. Improvements in EEG signal acquisition and technological advances in the use of wireless and dry sensors, easy to wear and robust with respect to noise [17] could represent the cue for outlining guidelines for practical systems implementation.

In Table 1, an extensive although not exhaustive list of research studies which have already been published using a resting state acquisition protocol, either closed eyes (CE) or open eyes (OE), is provided. It is evident that the database dimension is quite limited in almost all of these contributions. This is also due to the lack of a public EEG database suitably collected for the biometric recognition purpose, where acquisitions and protocols would be designed according to the specific requirements. In fact most of the works in this field test the implemented techniques on datasets recorded in BCI contexts. Moreover the issue of the repeatability of EEG biometrics in different acquisition sessions has never been systematically addressed in any of the aforementioned contributions and it has never received the required attention from the scientific community. Nevertheless, its understanding is propaedeutic towards the deployment of EEG biometrics in real life. Although in some referred works different acquisition sessions have been provided, they were considered to assort a single dataset where randomly selected EEG segments were used for training or testing a classification algorithm for the recognition purpose. On the other hand, in [12] the session-to-session variability was tested on a dataset of 6 subjects performing imagined speech. The entire set of \(128\) channels was used to extract features, and results show a decreasing performance when considering sessions temporally apart, which led to assess that the imagined speech EEG does not show to have a reliable degree of repeatability. Therefore, in this paper we further speculate on the use of EEG as a biometric characteristic by focusing on the analysis of repeatability of its features, thus starting filling a gap in the existing literature. More in details, we rely on two simple acquisition protocols, namely “resting states with eyes open” and “resting states with eyes closed” to acquire data from nine healthy subjects in two acquisition sessions separated in time. Different configurations for the number of electrodes employed and for their spatial placement have been taken into account. Specifically, sets of three and five electrodes have been considered to acquire the signals, and several frequency bands have been analyzes. The so acquired signals, after proper preprocessing, are then modeled using autoregressive stochastic modeling in the feature extraction stage. Linear classification is then performed. The paper is organized as follows. The acquisition protocol is detailed in Sect. 2.1, and in Sect. 2.2 the template extraction procedure is described. In Sect. 3 classification is performed, while experimental results are given in Sect. 4. Finally conclusions are drawn in Sect. 5.

2 Experimental Setup

2.1 EEG Data

Nine healthy volunteers have been recruited for this experiment. Informed consent was obtained from each subject after the explanation of the study, which was approved by the local institutional ethical committee. During the experiment, the participants were comfortably seated in a reclining chair with both arms resting on a pillow in a dimly lit room properly designed minimizing external sounds and noise in order not to interfere with the attention and the relaxed state of subjects. The subjects were asked to perform one minute of “resting state with eyes open” and one minute of “resting state with eyes closed” [18] in two temporally separated sessions, from 1 to 3 weeks distant from each other, depending on the subject.

Fig. 1.
figure 1

Scalp electrodes positioning in the employed protocol according to an extension of the standard 10-20 montage.

Brain activity has been recorded using a BrainAmp EEG recording system operating at a sampling rate of 200 Hz. The EEG was continuously recorder from 54 sites positioned according to the International 10-20 system as shown in Fig. 1. Such configuration is not meant to be a user convenient solution, but allows to investigate about a proper electrode placement, able to catch distinctive features according to the employed protocol. Before starting the recording session, the electrical impedance of each electrode was kept lower than 10 kOhm through a dedicated gel maximizing the skin contact and allowing for a low-resistance recording through the skin. After the EEG recording sessions, the EEG signals have been band pass filtered in the band \([0.5,30]\) Hz, before further analysis.

2.2 Methods

The template is generated by considering the signals acquired by a properly chosen set of electrodes. We have tested different acquisition configurations. Specifically, sets of three and five electrodes have been employed in our tests to understand, at a first stage, the proper number of electrodes to employ, limiting it, and at a later stage, the proper electrodes positioning to use in order to capture repeatable and stable features, if present. The signals so acquired are preprocessed as described in Sect. 2.2 in order to perform denoising and to select the proper subbands. Then, the EEG signals in the selected subbands are AR modeled as described in Sect. 2.2. The template is obtained by concatenating the reflection coefficients vectors related to the different channels in the sets under analysis. Specific brain rhythms are mainly predominant in certain scalp regions during different mental states. Therefore, we expected a certain variability of recognition performance spanning the entire scalp through specific configurations of electrodes, and considering the closed or open eyes condition, being different the capability to catch distinctive features.

Preprocessing. Before performing feature extraction, each acquired raw EEG signal has been processed as described in the following. Neural activity reflected in resting state EEG signals shows to contain frequency elements mainly below \(30\) Hz. Hence, a decimation factor has been applied to the collected raw signals, after filtering them through an anti-aliasing FIR filter. A sampling rate of \(S_r=60\) Hz was selected to retain spectral information present in the four major EEG subbands referring to the resting state (\(\delta \) \([0.5, 4]\) Hz, \(\theta \) \([4, 8]\) Hz, \(\alpha \) \([8, 13]\) Hz and \(\beta \) \([13, 30]\) Hz). The \(\gamma \) subband \([30, 40]\) Hz is not considered, given that it is known not to be relevant in a resting condition. A further stage of zero-phase frequency filtering was applied to discriminate the different EEG rhythms. The single \(\delta \), \(\theta \), \(\alpha \) and \(\beta \) subbands and their combinations (frequency components from \(0.5\) Hz up to \(30\) Hz, and from \(0.5\) Hz up to \(14\) Hz) have been considered in our experiments.

A spatial filter has been then applied to the acquired signals. When sufficiently large numbers of electrodes are employed, potential at each location may be measured with respect to the average of all potentials, approximating an inactive reference. Specifically, a common average referencing (CAR) filter has been employed in the herein proposed analysis by subtracting the mean of the entire \(C_T=54\) electrodes montage (i.e. the common average) from the channel \(c\) of interest, with \(c=1,2,\cdots , C_T\), at any one instant, according to the formula:

$$\begin{aligned} ^{CAR}V_u^{c}[n]=V_u^{c}[n]-\frac{1}{C_T}\sum _{j=1}^{C_T}V_u^{j}[n], \end{aligned}$$
(1)

where \(V_u^{j}[n]\) is the potential between the \(j\)-th electrode and the reference electrode, for the user \(u\), with \(u=1,2,\cdots , U\). CAR filtering has been employed to reduce artifacts related to inappropriate reference choices in monopolar recordings [19] or not expected reference variations, as well as to provide measures as independent as possible from the recording session. This results in an increased signal-to-noise ratio, since artifacts related to a single reference electrode are better controlled, as showed in [20], where authors compared spatial filter methods with a conventional ear reference in an EEG-based system.

Fig. 2.
figure 2

Classification results in % obtained for the best performing set of three (P7-Pz-P8) and five channels (Cz-CP5-CPz-CP6-Pz), considering AR order \(Q=10\) and different values of frame length (\(T_f\)).

A set of instances to be used for the training and the testing stages has been obtained from the signal segmentation. A range from 1 up to 3 seconds of EEG frame length has been spanned stepwise, in order to best characterize each user brain signal for the identification purpose. The one second frame length has been experimentally selected as it has shown to best catch distinctive features of users’ EEG segments for the recognition purpose. This can be observed in Fig. 2, where best performance is achieved both for sets of three and five electrodes, considering one second EEG segments, in the band \(\delta \cup \theta \cup \alpha \cup \beta =[0.5,30]\,\text {Hz}\) shown to be the best performing. These results refer to \(10\) order AR modeling and best sets of three and five channels, and show averaged performance obtained training the classifier on each acquisition session and testing it on the other one. Such framework has been employed to increase the number of trials used to study the repeatability of EEG biometrics in terms of recognition performance over the investigated period.

In this stage an overlap interval between adjacent frames was set to increase the sample size. Overlapping percentages of 25 %, 50 % and 75 % have been tested. Subsequently the DC component jointly to the linear trend has been removed from each EEG segment. The so obtained data-set have been further processed to extract the distinctive features from each user brain signal, as described below.

Modeling and Feature Extraction. After the preprocessing stage, detailed in Sect. 2.2, each acquired signal is modeled as a realization of an AR stochastic process. A realization \(x[n]\) of an AR process, of order Q, can be expressed as:

$$\begin{aligned} x[n] = - \sum _{q=1}^{Q} a_{{\scriptscriptstyle Q},q}\cdot x[n-q] + w[n] \end{aligned}$$
(2)

where \(w[n]\) is a realization of a white noise process of variance \(\sigma ^2_{{\scriptscriptstyle Q}}\), and \(a_{{\scriptscriptstyle Q},q}\) are the autoregressive coefficients. The well known Yule-Walker equations [21], which allow calculating the \(Q\) coefficients, can be solved recursively, employing the Levinson algorithm and introducing the concept of reflection coefficients. Specifically:

$$\begin{aligned} \left\{ \begin{array}{ll} a_{{\scriptscriptstyle Q},q}&{}= a_{{\scriptscriptstyle Q-1},q} + K_{{\scriptscriptstyle Q}}\cdot a_{{\scriptscriptstyle Q-1},{\scriptscriptstyle Q}-q}, \quad q=1,\cdots , Q-1\\ \ \sigma ^2_{{\scriptscriptstyle Q}}&{} = \sigma ^2_{{\scriptscriptstyle Q-1}} \left( 1-K^2_{{\scriptscriptstyle Q}}\right) , \end{array} \right. \end{aligned}$$
(3)

where the factor \(K_{{\scriptscriptstyle Q}}\) is the so-called reflection coefficient of order \(Q\) which is calculated as follows [21]:

$$\begin{aligned} K_{{\scriptscriptstyle Q}}= - \left( {\displaystyle R_x[Q] + \sum _{q=1}^{Q-1} R_x[q]\cdot a_{{\scriptscriptstyle Q-1},{\scriptscriptstyle Q}-q}}\right) /{ \sigma ^2_{{\scriptscriptstyle Q-1}}} \end{aligned}$$
(4)

where the generic \(R_x[m]\) is the signal autocorrelation function, defined as \(R_x[m]=\mathrm {E}_{}\left\{ {x[n]\,x[n-m]}\right\} \), for all \(m\ge 0\).

Among the possible estimation approaches, the Burg method [21] estimates the reflection coefficients \(K_{{q}}\), for \(q=1,\cdots ,Q\), operating directly on the observed data \(x[n]\) rather than estimating the autocorrelation samples \(R_x[m]\). Therefore, the Burg’s reflection coefficients, which have been shown in [5] to achieve better performance than the most commonly employed AR coefficients, are here employed.

Given the generic user \(u\), and the generic channel \(c\), let us indicate with \({\varvec{\zeta }}^{(u,c)}\) the vector, of length \(Q\), composed by the AR model reflection coefficients \(K_{{q}}\), for \(q=1,\cdots ,Q\), using the Burg method:

$$\begin{aligned} \varvec{\zeta }^{(u,c)} = [K^{(u,c)}_1, K^{(u,c)}_2, \cdots , K^{(u,c)}_Q]^T. \end{aligned}$$
(5)

The model order \(Q\) has been selected according to the Akaike Information Criterion (AIC) to minimize the information loss in fitting the data. It can be observed in Fig. 3(a), that the \(\text {AIC}(Q)\) function, averaged among subjects and channels, reaches minimum plateau zone for values of \(Q\) from \(6\) to \(12\). The feature vector \(\mathbf {x}\) for the user \(u\) is obtained by concatenating the AR coefficients vectors related to the signals obtained from the channels in the set under analysis. The 10 AR order has been experimentally selected since it has shown to best fit the EEG data for the recognition purpose, as it can be observed in Fig. 3(b), where correct classification percentage is reported considering one second EEG segmentation. Averaged results are shown, obtained training on each session and testing on the remaining one, and considering the best performing sets of three and five channels.

3 Classification

The classifier we propose estimates the class (user identity) to which the observed feature vector \(\mathbf {x}\) belongs to by means of a linear transformation \( \hat{\mathbf {y}_{}}^{\scriptscriptstyle \mathrm {T}}(\mathbf {G}) = \mathbf {x}^{\scriptscriptstyle \mathrm {T}}\mathbf {G}\), where the transformation matrix \(\mathbf {G}\) is obtained by minimizing the mean square error (MMSE) thus obtaining:

$$\begin{aligned} \mathbf {G}= \arg \min _{\varvec{\varGamma }} \sum _{i=1}^{N} \mathcal {P}_{i}\!\cdot \mathrm {E}_{\mathrm {x}|\mathcal {H}_{i}\!\!}\left\{ {\left[ \mathbf {y}_{i}\! -\!\hat{\mathbf {y}_{}}(\varvec{\varGamma })\right] ^{\scriptscriptstyle \mathrm {T}}\!\left[ \mathbf {y}_{i}\! -\!\hat{\mathbf {y}_{}}(\varvec{\varGamma })\right] }\right\} \end{aligned}$$
(6)

where \(\mathcal {H}_{i}\) indicates the hypothesis \(\mathbf {x}\) belongs to the \(i\)-th class, with \(i=1,2,\cdots ,N\). Here, assuming the hypothesis \(\mathcal {H}_{i}\) holds, the vector \(\mathbf {y}_{i}=[0,\cdots ,0,1,0,\cdots ,0]\) with the unique one in the \(i\)-th position, indicates the class \(i\) \(\mathbf {x}\) belongs to, while \(\hat{\mathbf {y}_{}}^{\scriptscriptstyle \mathrm {T}}(\mathbf {G})\) represents its estimation. \(\mathcal {P}_{i}\) denotes the prior probability that \(\mathbf {x}\) belongs to the \(i\)-th class. It can be easily shown that the employed optimization criterion given in (6) brings to the normal equations:

$$\begin{aligned} \mathbf {R}_{\mathrm {x}}\cdot \mathbf {G}= \mathbf {R}_{\mathrm {x}\mathrm {y}} \end{aligned}$$
(7)

where \(\mathbf {R}_{\mathrm {x}}= \mathrm {E}_{}\left\{ {\mathbf {x}\mathbf {x}^{\scriptscriptstyle \mathrm {T}}}\right\} \) is the auto-correlation matrix for the elements of the feature vector \(\mathbf {x}\), while \(\mathbf {R}_{\mathrm {x}\mathrm {y}}=\sum _{i=1}^{N}\mathcal {P}_{i}\cdot \mathrm {E}_{\mathrm {x}|\mathcal {H}_{i}\!\!}\left\{ {\mathbf {x}}\right\} \cdot \mathbf {y}_{i}^{\scriptscriptstyle \mathrm {T}}\) turns out to be the matrix whose columns are the probabilistically averaged conditional mean values of the observations \(\mathbf {x}\).

Fig. 3.
figure 3

(a) AIC function, averaged on all subjects and channels, for the frequency band \([0.5,30]\) Hz. (b) Classification results in % obtained for the best performing set of three (P7-Pz-P8) and five channels (Cz-CP5-CPz-CP6-Pz), considering \(T_f=1s\) and different AR orders \(Q\).

Dataset. As pointed out in Sect. 2.2, different sets of \(N_c = 3,5\) channels, to acquire the signals from which the feature vectors \(\mathbf {x}\) is extracted, have been considered for both the employed protocols. Given a chosen set of channels, each of the signals so acquired has been pre-processed, as described in Sect. 2.2, segmented into \(N_f\) frames, and modeled by resorting to the reflection coefficients of an AR model of order \(Q\). Therefore, considering, for each user, EEG signals of duration of \(60\,s\), segmented into frames of \(1, 2\) and \(3\,s\), with an overlap factor of 75 %, a number of \(N_f = 237, 117\) and \(77\) frames has been obtained respectively, each of which is represented by the feature vector \(\mathbf {x}\) of \(Q \times N_c\) elements.

Such a set of feature vectors has been collected for each of the two temporally separated recording sessions, 1-3 weeks distant from each other, and each protocol, i.e. closed and open eyes resting conditions. It is worth pointing out that the vectors used in the training stage and in the recognition stage have been obtained from the two different acquisition sessions in order to infer about the repeatability over the considered interval of the EEG features for the acquired dataset and the employed acquisition protocols. Hence, we applied the classification algorithm selecting the train and the test datasets without shuffling the EEG frames belonging to different sessions, as performed in other works with user recognition aims. In order to achieve our goal, each one of the two sessions has been sequentially considered for the training dataset while the remaining session has been used to obtain the test dataset, thus obtaining two couples of temporally separated datasets, (training set, recognition set) to train and test the classifier. This kind of validation framework has been provided just to encrease the statistical significance of the results. They show that we can’t assess a perfect symmetry of changes over time, but that the features keep stable over the considered interval (1-3 weeks). Some of the results of each test are shown in subsequent columns of Table 2 where each set has been acquired at a different time.

3.1 Training

The training stage consists in the estimation of the matrix \(\mathbf {G}\) in (7) computed as \(\mathbf {G}=\widehat{\mathbf {R}}_{\mathrm {x}}^{-1}\cdot \widehat{\mathbf {R}}_{\mathrm {x}\mathrm {y}}\), where the matrices \(\mathbf {R}_{\mathrm {x}}\) and \(\mathbf {R}_{\mathrm {x}\mathrm {y}}\) are estimated through MonteCarlo runs, considering equal prior probabilities \(\mathcal {P}_{i}\) for all the classes (users identities) to distinguish between. The estimation was obtained performing the following two sample averages:

$$\begin{aligned} \begin{array}{ll} \widehat{\mathbf {R}}_{\mathrm {x}}&{}=\displaystyle \frac{1}{NM}\sum _{i=1}^{N}\sum _{m=1}^{M}\mathbf {x}_{m,i}\mathbf {x}_{m,i}^T\\ \widehat{\mathbf {R}}_{\mathrm {x}\mathrm {y}}&{}=\displaystyle \frac{1}{NM}\sum _{i=1}^{N}\sum _{m=1}^{M}\mathbf {x}_{m,i}\mathbf {y}_{i}^T, \end{array} \end{aligned}$$
(8)

where \(\mathbf {x}_{m,i}\) is the \(m\)-th observed feature vector belonging to the \(i\)-th class, with \(M\) being the number of instances of \(\mathbf {x}\) for each class, and \(\mathbf {y}_{i}=[0,\cdots ,0,1,0,\cdots ,0]^T\), with the unique \(1\) in the \(i\)-th position. The considered matrices can be simply upgraded in case of enrollment of \(N'\) new users, summing the related matrices \(\mathbf {x}_{m,i}\mathbf {x}_{m,i}^T\) to \(\mathbf {R}_{\mathrm {x}}\), and adding new columns \(i\) to \(\mathbf {R}_{\mathrm {x}\mathrm {y}}\) given by \(\frac{N}{M(N+N')}\sum _{m=1}^{M}\mathbf {x}_{m,i}\), where \(i=N+1,\cdots , N+N'\). To avoid failures and to control accuracy in the estimation of \(\mathbf {R}_{\mathrm {x}}^{-1}\), the singular value decomposition based pseudoinversion has been used for the matrix inversion.

Table 2. Classification results in % for CE protocol, obtained using the acquisition session \(t\) for training and the acquisition session \(r\) for recognition, with \(t,r=S_1,S_2\) and \(t\ne r\), for the subband \(\delta \cup \theta \cup \alpha \cup \beta =[0.5,30]\) Hz, for sets of three electrodes. For each test \(t\rightarrow r\) \(2\) results are provided, considering \(75\,\%\) of the training dataset while \(25\,\%\) (first column) and \(75\,\%\) (second column) of the test dataset.

3.2 Recognition

In the recognition stage, a linear transformation is applied to each of the \(M \times N\) observations from the test dataset. For the \(i\)-th user a score vector \(\hat{\mathbf {y}}_{i}\) was obtained for each instance of \(\mathbf {x}_i\) in the dataset applying the discrimination matrix \(\mathbf {G}\) to \(\mathbf {x}_{m,i}\):

$$\begin{aligned} \hat{\mathbf {y}_{}}_{m,i}=\mathbf {G}\cdot \mathbf {x}_{m,i} \end{aligned}$$
(9)

with \(m=1,\cdots , M\). Subsequently, the \(M\) score vectors related to each tested user were summed together to reduce the misclassification error, obtaining

$$\begin{aligned} \hat{\mathbf {y}}_{i}=\sum _{m=1}^{M}\hat{\mathbf {y}_{}}_{m,i}. \end{aligned}$$
(10)

Finally the estimation of the index representing the user identity is obtained locating the maximum of the score vector \(\hat{\mathbf {y}}_{i}= [\hat{y}_i(1),\cdots , \hat{y}_i(N)]^T\) according to the criterion

$$\begin{aligned} \hat{i}=\arg _l\max y_i(l). \end{aligned}$$
(11)

As previously pointed out, to solve the classification problem we separately considered different symmetrical sets of sensors and different brain rhythms for the template extraction. In order to improve accuracy, an information fusion integrating multiple sensors distributions and brain rhythms was then performed at the match score level, which is the most common approach in multibiometric systems [22]. The aim was to determine the best sets of channels configurations and frequency bands that could optimally combine the decisions rendered individually by each of them. The basic hypothesis was that different representations of brain rhythms show different distinctive traits, which may increase variability between-subjects if efficiently combined together. In this regard the proper level for information fusion in the multibiometric approach is an important issue to dramatically improve the classification performance. We observed that different subjects showed advantagious scores for different sets of channels and different rhythms, that is each of them presented particular spactral distribution and topography of distinctive traits (Table 4). Hence, for each tested user the proposed score fusion was obtained through the sum

$$\begin{aligned} \frac{1}{N_B N_{Ch}}\sum _{b}^{B}\sum _{Ch}^{S} { }_{Ch}^b\hat{\mathbf {y}}_{i}\end{aligned}$$
(12)

of scores vectors \(\hat{\mathbf {y}}_{i}\) related to specific \(N_B\) bands \(b\in B\) and \(N_S\) selected sets \(Ch\in S\) composed of three electrodes. All tests performed and obtained results are reported in the next Section.

4 Experimental Results

The results of the performed analysis are reported for all the experiments carried out. The aim of the study was to test the repeatability of the considered EEG features, needed to recognize users previously enrolled in a biometric system. Repeatability and stability represent properties of paramount importance for the use of EEG biometrics in real life systems. For this purpose we have selected two simple tasks to be performed by a set of users, and a classification problem has been set up, where the training set and the one to be used in the recognition stage have been chosen belonging to temporally separated sessions 1-3 weeks distant from each other.

More in detail, given the “resting state” acquisition protocols here considered and the \(54\) employed channels shown in Fig. 1, we have selected different subsets of them in order to find the best performing spatial arrangements of the electrodes while minimizing their number. Although the considered acquisition technique doesn’t result user convenient, not being this the focus of the paper, in a preliminary study, as this is, it allows to detect on the scalp the brain rhythms which provide the best distinctive features, according to the employed protocol. In order to achieve this goal we have considered sets of three and five electrodes, the former listed in Table 2. An information fusion approach combining match scores obtained for the selected distributions of sensors and the selected brain rhythms is also proposed to improve performance.

Table 3. Classification results in % for OE protocol. See caption of Table 2 for description.
Table 4. Classification results in % of correct identification reported for each subject \(I_i\) in the cross validation framework (averages over 237 runs). Results refer to CE condition, when training on session \(S_1\) (\(75\,\%\) of frames) and performing identification tests on \(S_2\) (\(25\,\%\) of frames).
Fig. 4.
figure 4

Improvement of the correct recognition rate obtained performing subsequent score fusions (see Sect. 4), within the CE condition. Curves refer to the combination of different electrodes sets (x-axis). Labels in the x-axes refer to the score added at each subsequent step. Results refer to the training on \(S_1\) and test on \(S_2\).

Template extraction has been performed as described in Sect. 2.2, by first preprocessing the EEG signals, which includes decimation with sampling rates \(S_r= 60 \) Hz, CAR spatial filtering, segmentation into frames of \(T_f\) s with an overlapping factor \(O_f\) between consecutive frames, and eventually band pass filtering in order to analyze the subbands \(\delta , \theta , \alpha \) and \(\beta \), which are the ones interested by the “resting state” protocols, and some of their combinations. A value of \(O_f=75\,\%\) has been here employed since we have experimentally proven it is able to guarantee good performance as it provides an adequate sample size to assort the training and recognition datasets. Then the so obtained frames are modeled using an AR model, whose tested orders \(Q=\{6, 8, 10, 12\} \) have been estimated by means of the AIC function (see Fig. 3(a)). Value of \(T_f=1\) s and \(Q= 10 \) have been selected as they showed to best characterize the users’ EEG for the recognition task (see Fig. 2).

The template is then obtained by concatenating the reflection coefficients of the signals acquired by means of the electrode set under analysis, thus generating feature vectors of length \(3Q, 5Q\) for the sets of three, and five electrodes respectively.

In Tables 2 and 3 the results obtained for sets of three electrodes when using the MMSE classifier, described in Sect. 3, are given for both employed protocols CE and OE. It is worth pointing out that the signals employed to obtain the templates to be used in both the training and the recognition stage are disjoint in time. Therefore two different combinations of training \((t)\) and recognition \((r)\) sessions, \((t,r)\) with \(t,r\in \{S_1,S_2\}\) and \(t\ne r\), have been tested. Such kind of tests varying the sequence of sessions in the recognition framework are provided to validate the results about repeatability of the considered EEG features over the interval under analysis, for a real usability of an EEG-based biometric system. The results for the different tests performed are reported separately, not expecting to make assumptions on symmetry of changes over time. Moreover, from the analysis of different spatial configurations of electrodes we could observe that triplets of channels allow achieving about same performance then configurations employing sets of five channels. This is due to a good spatial localization achieved by configurations of only three sensors, which allow to well capture the underlying phenomena, reducing the problem dimensionality. Moreover, in Tables 2 and 3 results are shown considering the band \(F=\delta \cup \theta \cup \alpha \cup \beta =[0.5, 30]\) Hz which is the one that allows obtaining the best results, and considering a preprocessing including CAR filtering or not.

Table 5. Classification results in % for both CE and OE protocols, obtained using the same acquisition session \(S\) for training and for recognition, with \(S=S_1,S_2\), for the subband \(\delta \cup \theta \cup \alpha \cup \beta =[0.5,30]\) Hz, for the best performing sets of three and five electrodes. Results are provided considering \(75\,\%\) of the dataset for training while \(25\,\%\) of the dataset for recognition.

Provided performance refers to a cross-validation framework, obtained selecting for each user \(75\,\%\) of feature vectors \(\mathbf {x}\) related to cyclically subsequent frames from the training dataset, while \(25\,\%\) and \(75\,\%\) from the test dataset, as reported in subsequent columns of Tables 2 and 3. Numerical results are obtained averaging over \(237\) independent cross-validation runs to improve the statistical analysis. As previously pointed out, recognition tests have been carried out keeping independency between the training and the test datasets, acquired in different sessions, for the classification purpose. This aspect is highlighted in Tables 2 and 3 denoting with \(S_i\rightarrow S_j\) the result achieved training on \(S_i\) and testing on \(S_j\).

It should be noticed that applying the CAR filter in the preprocessing stage doesn’t yield a general improvement in the performance for all employed sets of channels and protocols, while it appears to provide best results for some selected channels (FC3-FCz-FC4, C3-Cz-C4, T7-Cz-T8) for sets of 3 channels in the OE protocol. This is likely to be due to artifacts which more affected the open-eyes condition, removed by the spatial filtering. As regards differences between the two employed protocols it is evident, by observing the reported results, that in this experiment the CE protocol provides best performance considering the adopted EEG feature extraction for the recognition task. In fact, within the CE condition \(100\,\%\) of correct classification is achieved for instance employing channels P7-Pz-P8 and \(75\,\%\) of the test feature vectors for each user in the cross-validation framework. This has been observed to be due both to being the open-eyes signal more affected by the eyes movement artifacts, and to distinctive traits contained in the \(\alpha \) rhythm which is mainly detected on the posterior head when resting with eyes closed. In this regard is was noticed that the combination of channels affected in a different way the recognition results for CE and OE protocols. In fact, referring to sets of three channels the parietal region has proven to best perform in CE condition, while the central region achieved best results in OE condition. Moreover it has been observed, individually analyzing the extracted brain rhythms, that in CE the \(\alpha \) band most contributed to the best performance obtained combining all bands (\([0.5,30]\,\text {Hz}\)). The results just pointed out are in agreement with the fact that in resting state with eyes closed the dominant brain rhythm \(\alpha \) can be detected mainly in the posterior area of the scalp, while it is attenuated when opening eyes.

Repeatability over the considered interval of the analyzed EEG features can be inferred by observing that users enrolled in a session have been recognized in a different one, disjoint in time from 1 to 3 weeks. Besides, it is also evident that by swapping the training and recognition roles of the session datasets, that is by considering \((t,r)\) or \((r,t)\), quite coherent performance are obtained.

Table 5 shows results obtained training and testing the classifier on the same session. It should be noticed that very high correct recognition rate is achieved considering just \(25\,\%\) of the test dataset (\(100\,\%\) for CE and \(S_2\)), while a greater number of feature vectors for each user are needed in the inter-sessions framework. This evidence proves the importance of speculating about the stability and repeatability over time of EEG features for biometric systems. The performance significantly decreases for the case of disjoint training and test datasets when considering just few frames for the identification test (\(25\,\%\) of the test dataset). On the other hand, the match score fusion obtained as discussed in Sect. 3.2 has led to a dramatic increase of the recognition accuracy, especially for the otherwise poorly performing case just mentioned, as observed in Fig. 4 where the CE condition is analyzed. Same improvements are observed within the OE condition, not reported in here due to space limitations, still remaining the CE condition the best performant one. For the selection of the electrodes configurations to combine, the best combination of rhythms was considered, and subsequent score fusions were performed. To this aim the electrode sets were sorted in descending order of performance achieved individually (see Table 2), and sequentially combined within a forward-backward stepwise approach, retaining in the information fusion only those sets which improved the correct classification. Results reported in Fig. 4 showed that a significant improvement was obtained combining rhythms and sets of channels. Within the multi-session framework a perfect recognition percentage of \(100\,\%\) could be achieved when the sets of three inter-hemispheric channels \( P7-Pz-P8, C1-Cz-C2, C3-Cz-C4, FC3-FCz-FC4, AF3-AFz-AF4\), and the rhythms \([0.5,30], \delta , \theta , \alpha \) containing most information, were combined into the match score fusion. It should be noticed that the selected channels result located all over the head, showing that antero-posterior differences could be distinctive as well as iter-hemispheric asymmetry. Figure 4 reports the improvements obtained across the subsequent steps of the information fusion, when considering \(75\,\%\) of training frames from \(S_1\) and just \(25\,\%\) of test frames from \(S_2\), for the CE condition. Accordingly, this approach allows to obtain high accuracy while significantly reducing the recording time needed for the recognition tests.

5 Conclusions

In this paper the problem of repeatability over time of EEG biometrics, for the same user, within the framework of EEG based recognition, has been addressed. Simple “resting state” protocols have been employed to acquire a database of nine people in two different sessions separated in time from 1 to 3 weeks, depending on the user. Although the dimension of the database employed is contained, we would like to stress out that this contribution represents the first systematic analysis on the repeatability issue in EEG biometrics. As such, this contribution paves the road to more refined analysis which would include more sessions separated in time as well as different acquisition protocols. Extensive simulations have been performed by considering different sets of electrodes both with respect to their positioning and number. A combination of match scores obtained from different analysis have shown to significantly reduce the frames needed for test, still maintaining high recognition accuracy. In summary in our analysis a very high degree of repeatability over the considered interval has been achieved with a proper number of electrodes, their adequate positioning and by considering appropriate subband related to the employed acquisition protocol.