Keywords

1 Introduction

Electroencephalogram (EEG) are widely used techniques in cognitive neuroscience and clinical surgeries due to their non-invasiveness and high temporal resolution. The task to reconstruct the corresponding cortical neural activities from EEG signals on the scalp surface is termed as EEG source imaging. Using EEG source imaging, we can estimate the extents and locations of brain active regions, which provide important auxiliary information for neuroscience research and medical treatments.

To reconstruct brain activities, the distributed source model divides the cerebral cortex into several triangular grids, each of which represents a dipole source [8]. The amplitude of each diploe source can be calculated by solving a linear inverse problem. However, the number of the candidate dipole sources largely outnumbered that of EEG sensors. Hence, the linear inverse problem is highly ill-posed, i.e., there are infinite source solutions to satisfy EEG signals. To obtain a unique solution, appropriate regularization operator is needed to narrow the solution space [8, 9, 16].

Typical EEG source imaging methods employ the \({L_2}\)-norm regularizer, such as the weighted minimum norm estimate (wMNE) [4] and the low-resolution electromagnetic tomography (LORETA) [11]. However, the estimations of \({L_2}\)-norm based methods are overly diffused and spread widely [8]. To improve the spatial resolution, sparse constrained methods based on \(L_p\)-norm (\(0<p\le 1\)) regularizer [1] and sparse Bayesian learning (SBL) [15] framework are developed. However, the solutions of conventional sparse constrained methods are too sparse and provide little information about source extents. To estimate the locations and extents of brain activities, several studies have proposed to implement sparse constraint on the transformed domains, such as the variation, wavelet and Laplacian domains [3, 5, 10, 16], which can provide more information about the source extents than the sparse constraints in the original source domains.

Most of the aforementioned methods assume that the measurement noise satisfies Gaussian distribution and employ the \(L_2\)-norm to measure the residual error. Nevertheless, the EEG signals are inevitably contaminated by the strong background activities and outliers caused by ocular or head movements during recording procedure. The \(L_2\)-norm may not be suitable to represent these outliers. To handle these outliers in the EEG signals and estimate the extended cortical activities, as in [1], we propose a robust EEG source imaging method, namely \({L_1}\)-norm Residual Variation Sparse Source Imaging (\(L_1\)R-VSSI). \(L_1\)R-VSSI employs the \(L_1\)-norm to measure the residual error. In addition, to estimate the locations and extents of sources, \(L_1\)R-VSSI attains sparseness of sources on the variation domain by incorporating the \(L_1\)-norm constraint. The cortical sources of \(L_1\)R-VSSI are efficiently obtained by the alternating direction method of multipliers (ADMM) algorithm [2].

The remaining of the paper is organized as follows. In Sect. 2, we present details of the proposed EEG source imaging algorithm. In Sect. 3, the simulation design and evaluation metrics are presented. Section 4 presents the results on simulated and real EEG data, followed by a brief conclusion in Sect. 5.

2 Methods

2.1 Background

The relationship between the EEG recordings and cortical sources can be expressed as [8]

$$\begin{aligned} \varvec{b}=\varvec{Ls}+\varvec{\varepsilon } \end{aligned}$$
(1)

where \(\varvec{b}\in \mathbb {R}^{d_b \times 1}\) is EEG recording, \(\varvec{s}\in \mathbb {R}^{d_s \times 1}\) is the unknown source vector. \(\varvec{L}\in \mathbb {R}^{d_b \times d_s}\) is the lead-field matrix that describes how a unit source of a certain candidate location is related to the EEG measurements. \(d_b\) and \(d_s\) denote the number of sensors and candidate sources respectively. \(\varvec{\varepsilon }\) denotes the measurement noise vector.

To estimate the extents and locations of sources, as in [5, 10], we assume that the sources are sparse in the variation domain. Each element in the variation domain is the difference of amplitudes between two adjacent dipole sources. When the amplitudes of sources are uniform, the non-zero vales of the variation sources are largely expected to occur on the boundaries between the active and inactive areas. Hence, we can estimate the sparseness on the variation domain to obtain the extents and locations of cortical activities. To obtain the variation sources, a variation operator \(\varvec{V}\in \mathbb {R}^{P \times d_s}\) is defined as [5, 10]

(2)

where P denotes the number of edges of triangular grids, which are defined by the source model. The pth row of \(\varvec{V}\) corresponds to edge p. For each row of \(\varvec{V}\), only two elements corresponding to the two dipoles that share edge p are non-zero (i.e., 1, −1). The pth row of the variation sources \(\varvec{u} = \varvec{V}\varvec{s} \in \mathbb {R}^{P \times 1}\), \(\varvec{u_p}\), indicates the difference of dipoles that share the pth edge.

2.2 \(L_1\)R-VSSI: \(L_1\)-norm Residual Variation Sparse Source Imaging

Usually, the EEG measurement noise \(\varvec{\varepsilon }\) is assumed to satisfy Gaussian distribution. And the fitting error is measured by the \(L_2\)-norm for most EEG source imaging methods. However, the EEG signals are inevitably influenced by the artifacts induced by the head movement or eye blinks. The \(L_2\)-norm for the fitting error will exaggerates the effect of these outliers. To handle these outliers or artifacts in EEG, our proposed EEG source imaging method, \(L_1\)R-VSSI, employs the \(L_1\)-norm to measure the fitting error. Combining the variation sparse constraint, the solution of \(L_1\)R-VSSI is obtained as

$$\begin{aligned} \hat{\varvec{s}}= \arg \min _{\varvec{s}} \Vert \varvec{Ls}-\varvec{b}\Vert _1+\lambda \Vert \varvec{Vs}\Vert _1 \end{aligned}$$
(3)

where \(\lambda >0\) is the regularization parameter. In this work, the ADMM algorithm is employed to solve formulation (3).

We first rewrite Eq. (3) as

$$\begin{aligned} \hat{\varvec{s}}&= \arg \min _{\varvec{s}} \Vert \varvec{e}\Vert _1+\lambda \Vert \varvec{u}\Vert _1 \nonumber \\ s.t.,&\varvec{e}=\varvec{Ls}-\varvec{b},\varvec{u}=\varvec{Vs} \end{aligned}$$
(4)

Then the augmented Lagrangian function of the constrained optimization problem (4) is

$$\begin{aligned} \begin{aligned} \mathcal {L}(\varvec{s},\varvec{e},\varvec{u},\varvec{y},\varvec{z})=\,&\Vert \varvec{e}\Vert _1+\lambda \Vert \varvec{u}\Vert _1+\varvec{y}^\top (\varvec{Ls}-\varvec{b}-\varvec{e})+\frac{\rho _1}{2}\Vert \varvec{Ls}-\varvec{b}-\varvec{e}\Vert _2^2\\&+\,\varvec{z}^\top (\varvec{Vs}-\varvec{u})+\frac{\rho _2}{2} \Vert \varvec{Vs}-\varvec{u}\Vert _2^2 \end{aligned} \end{aligned}$$
(5)

where \(\rho _1>0\) and \(\rho _2>0\) are penalty parameters, \(\varvec{y}\in \mathbb {R}^{d_b\times 1}\) and \(\varvec{z}\in \mathbb {R}^{P\times 1}\) are the Lagrangian multipliers. Since \(\mathcal {L}\) is separable with respect the variables \((\varvec{s},\varvec{e},\varvec{u})\), we can obtain \((\varvec{s},\varvec{e},\varvec{u})\) with three subproblems.

$$\begin{aligned} \hat{\varvec{s}}^{(k+1)}= & {} \arg \min _{\varvec{s}}\mathcal {L} \left( \varvec{s},\varvec{e}^{(k)},\varvec{u}^{(k)},\varvec{y}^{(k)},\varvec{z}^{(k)} \right) \nonumber \\= & {} \Big (\rho _1 \varvec{L}^\top \varvec{L}+\rho _2 \varvec{V}^\top \varvec{V} \Big )^{-1} \Big [\rho _1 \varvec{L}^\top \left( \varvec{b}+\varvec{e}^{(k)}\right) \nonumber \\+ & {} \rho _2 \varvec{V}^\top \varvec{u}^{(k)}-\varvec{L}^\top \varvec{y}^{(k)}-\varvec{V}^\top \varvec{z}^{(k)} \Big ] \end{aligned}$$
(6)
$$\begin{aligned} \hat{\varvec{e}}^{(k+1)}= & {} \arg \min _{\varvec{e}}\mathcal {L} \left( \varvec{s}^{(k+1)},\varvec{e},\varvec{u}^{(k)},\varvec{y}^{(k)},\varvec{z}^{(k)} \right) \nonumber \\= & {} \mathcal {S}_\frac{1}{\rho _1} \left( \varvec{L}\varvec{s}^{(k+1)} -\varvec{b}+\frac{1}{\rho _1}\varvec{y}^{(k)} \right) \end{aligned}$$
(7)
$$\begin{aligned} \hat{\varvec{u}}^{(k+1)}= & {} \arg \min _{\varvec{u}}\mathcal {L} \left( \varvec{s}^{(k+1)},\varvec{e}^{(k+1)},\varvec{u},\varvec{y}^{(k)},\varvec{z}^{(k)} \right) \nonumber \\= & {} \mathcal {S}_\frac{\lambda }{\rho _2} \left( \varvec{V}\varvec{s}^{(k+1)}+\frac{1}{\rho _2}\varvec{z}^{(k)} \right) \end{aligned}$$
(8)

where \(\mathcal {S}_\kappa (a)\) is

$$\begin{aligned} \mathcal {S}_\kappa (a)=\left\{ \begin{aligned}&\ a-\kappa ,&a>\kappa \\&\ 0,&|a|<\kappa \\&\ a+\kappa ,&a<-\kappa \end{aligned} \right. \end{aligned}$$
(9)

The Lagrangian multipliers \(\varvec{y}\), \(\varvec{z}\) are updated using the dual ascent method [2]

$$\begin{aligned}&\hat{\varvec{y}}^{(k+1)} = \varvec{y}^{(k)}+\rho _1 \left( \varvec{L}\varvec{s}^{(k+1)}-\varvec{b}-\varvec{e}^{(k+1)} \right)&\end{aligned}$$
(10)
$$\begin{aligned}&\hat{\varvec{z}}^{(k+1)} = \varvec{z}^{(k)}+\rho _2 \left( \varvec{V}\varvec{s}^{(k+1)}-\varvec{u}^{(k+1)} \right)&\end{aligned}$$
(11)

where \(\varvec{x}^{(k)}\) is the value of x at kth iteration.

In summary, to obtain the source estimations, \(L_1\)R-VSSI alternatively updates the variables \((\varvec{s},\varvec{e},\varvec{u})\) and the Lagrangian multipliers \((\varvec{y},\varvec{z})\), until the relative change of \(\varvec{s}\) reaches a user-specified tolerance (e.g., \(10^{-4}\)). Through our simulations, \(\varvec{e}, \varvec{u}, \varvec{y}, \varvec{z}\) are all initialized to \(\varvec{0}\). The penalty parameters \(\rho _1\), \(\rho _2\) are determined by cross-validation [14]. For the simulation experiment of \(d_s=6002\), \(d_b=62\), \(P = 17977\), it reaches convergence after approximately 200 iterations, which takes about 2 min.

3 Simulation Design and Evaluation Metrics

To test the performance of \(L_1\)R-VSSI, a series of Monte-Carlo numerical simulations were set up to compare \(L_1\)R-VSSI and four imaging methods: two widely used \({L_2}\)-norm based methods, i.e., wMNE [4] and LORETA [11], and sparse constraint methods in the original domain (SBL [15]) and in the variation domain (variation-based sparse cortical current density (VB-SCCD) [5]). Additionally, \(L_1\)R-VSSI was also applied to one real clinical EEG data.

3.1 Simulation Design

In the numerical simulations, a three-shell head model is obtained using Brainstorm [13]. There are 6002 triangle grids evenly distributed on the cortical surface, each of which is a dipole source. The dipole orientation is fixed to be normal to the cortical surface. The lead-field matrix \(\varvec{L}\) was constructed using Brainstorm with the sensor configuration of the 64-channel Neuroscan Quik-cap system (two channels are not EEG electrodes, hence, \(\varvec{L} \in \mathbb {R}^{62 \times 6002}\)).

A triangle grid is randomly selected as a seed point. The adjacent triangles are gradually added to form an extended source. After multiplying the source vector with the lead-field matrix, the cortical activities are transformed into EEG electrode recordings. To simulate the actual EEG signals, both Gaussian noise and outliers are added to the EEG electrode recordings. The noise level is described by signal-to-noise ratio (SNR), which is defined as \( \text {SNR}=10 \log _{10}\left[ \frac{\sigma ^2(\varvec{Ls})}{\sigma ^2(\varvec{\varepsilon })}\right] \). where \(\varvec{Ls}\) is raw EEG data without noise and \(\varvec{\varepsilon }\) is mixture measurement noise (Gaussian and outliers). \(\sigma ^2(\varvec{x})\) is the variance of \(\varvec{x}\).

As in [1], the mixture measurement noise is defined as

$$\begin{aligned} \varvec{\varepsilon }=\frac{\varvec{\varepsilon }_1}{\sigma (\varvec{\varepsilon }_1)} \left[ 10^{(-\frac{\text {SNR}}{20})} \right] \sigma (\varvec{Ls}) \end{aligned}$$
(12)

where \(\varvec{\varepsilon }_1\) denote the outliers and \(\sigma \) is the standard deviation.

To generate the mixture noise, we first generate the outliers \(\varvec{\varepsilon }_1 \in \mathbb {R}^{d_b \times 1}\), where each element of \(\varvec{\varepsilon }_1\) obeys Gaussian distribution \(\mathcal {N}(\mu +10 \sigma ^2,\sigma ^2)\) [1]. \(\mu \) and \(\sigma ^2\) are the mean and variance of the true EEG \(\varvec{Ls}\) across all channels at the single time slice, respectively. Then, given a specified SNR, the mixture noise is generated by Eq. (12) and added to the true EEG recordings \(\varvec{Ls}\).

In the numerical Monte-Carlo experiments, the performance of the proposed method are verified in two cases: (1) One extended source with different extents (i.e., 0.8, 4, 8, 12, 18 cm\(^2\)); (2) EEG signals with different SNRs (i.e., 0, 5, 10, 15 dB). For each case, 100 simulation experiments were conducted to ensure that the simulated extended sources could cover most areas of the brain.

3.2 Evaluation Metrics

The performance of imaging algorithms are evaluated by four evaluation metrics. (1) The area under the receiver operating characteristic (ROC) curve (AUC), which assesses the sensitivity and specificity of imaging algorithms [7]; (2) Spatial dispersion (SD), which depicts that the level of the spatial dispersion of reconstructed source [3]; (3) Distance of localization error (DLE), which describes the location error of the estimated source activities [16]; (4) Relative mean square error (RMSE), which estimates the amplitudes accuracy of the reconstructed sources [9]. Larger AUC, smaller SD, DLE and RMSE values indicate that the method has better performance. The detailed computation of these metrics can refer to the supplementary document in [10]. To image the estimated sources, the absolute value of the recovered sources at specified time points are shown, thresholded by the Otsu’s method [7, 10].

4 Results

4.1 Results for Various Extents

Figure 1 shows performance metrics under different extents. As the source extents increases, the AUC values of wMNE, SBL, VB-SCCD, \(L_1\)R-VSSI gradually decrease. In contrast, LORETA’s AUC value increases, indicating the potential of LORETA to estimate large extents’ sources. The SD, DLE, RMSE values of wMNE, LORETA, VB-SCCD and \(L_1\)R-VSSI gradually reduced, but \(L_1\)R-VSSI get the smallest values, which means that \(L_1\)R-VSSI have more accurate reconstructed solutions, smaller spatial dispersion and location errors. Due to the sparse constraint of SBL, although it gets the smallest SD, DLE values, its overly focal results also lead to a gradually increasing RMSE value. Comprehensively comparing the results of four evaluation metrics, the \(L_1\)R-VSSI method shows better performance than the other four methods under different source extents.

Fig. 1.
figure 1

Evaluation metrics of different source extents. The data is the results of 100 Monte Carlo simulations and is described as Mean ± SEM (SEM: standard error of mean). The SNR is 5 dB.

Figure 2 depicts the imaging maps of reconstructed sources under different extents. The first column in Fig. 2 shows the simulated extended sources, and the remaining columns show the reconstructed sources estimated by the five imaging methods respectively. From the Fig. 2, we can see that SBL can reconstruct the focal sources perfectly. However, for large extents’ sources, it can only accurately locate the center of the sources, but the estimated extents are overly focused. In contrast, the solutions of wMNE and LORETA are too blurred compared to the simulated sources. Because the addition of variation operator, the blurriness of the imaging results of VB-SCCD solution is greatly reduced and more sensitive to the extents of the simulated sources. Nevertheless, compared to VB-SCCD, \(L_1\)R-VSSI approach’s recovered solutions have more precise description for extents and strengths. This is because the use of \(L_1\)-loss of the residual error reduces the influence of artificial noise. Combining the results of Figs. 1 and 2, \(L_1\)R-VSSI performs better than wMNE, LORETA, SBL and VB-SCCD.

Fig. 2.
figure 2

Imaging results for different extents of extended sources. Source activity maps show the absolute value of the extended sources’ activities. The threshold is determined using Otsu’s method. The SNR is 5 dB. Some sources are circled for illustration purposes.

4.2 Results for Different SNRs

Figure 3 shows the performance evaluation metrics under different SNRs (0, 5, 10, 15 dB). All algorithms are significantly affected by the noise level. As the SNR increases, these algorithms’ AUC value gradually increases, and the RMSE, SD, and DLE value decrease. Under comprehensive comparison, \({{L}_{1}}\)R-VSSI has a larger AUC value, smaller SD, RMSE, DLE value, so the performance is better than other approaches.

Fig. 3.
figure 3

Evaluation metrics of various SNRs. The data is the results of 100 Monte Carlo simulations and is described as Mean ± SEM (SEM: standard error of mean). The extents of source is about 8 cm\(^2\).

4.3 Application of Epilepsy EEG Data

To show the practical performance of \({{L}_{1}}\)R-VSSI, we further apply the proposed method to the Brainstorm public epilepsy dataFootnote 1. The average spikes across 58 trials is used to estimate sources, which are presented in Fig. 4(a). We calculate the lead-field matrix based head model obtained from the head anatomical structure of the subject. The EEG data was collected from a patient who suffered from focal epilepsy with focal sensory, dyscognitive and secondarily generalized seizures since the age of eight years. After the resection of the epileptogenic area, postoperative investigation showed the patient no epilepsy in 5 years. In [6], it has located that the patient’s epileptogenic area is left frontal, which was estimated by invasive EEG. Figure 4(b) shows the imaging results of various algorithms at the peak time point (i.e., 0 s). Comparing the results in Fig. 4(b) with the analysis in [6], we can find that these algorithms all locate the epileptogenic area. However, the epileptogenic area identified by LORETA and wMNE are overly diffused, even covering many other areas of the brain scalp. Conversely, SBL can only obtain some point sources in the epileptogenic area. For the variation sparse constraint based methods, the estimations by \(L_1\)R-VSSI and VB-SCCD is well accordant with previous reports [6, 12]. However, the solution of \(L_1\)R-VSSI shows more clear boundaries than that of VB-SCCD.

Fig. 4.
figure 4

Application of the public epilepsy data. (a) is averaged EEG signals across 58 trials. (b) is the imaging maps, showing the absolute value of the source activities at the peak time point. The imaging threshold was determined by Otsu’s method. SBL’s estimation is circled for illustration purposes.

5 Conclusions

We have proposed a robust EEG source estimation algorithm, \({{L}_{1}}\)R-VSSI, which reconstructs extended sources using \({{L}_{1}}\)-loss of the residual error and sparse constraints in the variation domains. The solution of \({{L}_{1}}\)R-VSSI is efficiently obtained by ADMM. Numerical results indicate that \({{L}_{1}}\)R-VSSI can effectively alleviate the influence of outliers and achieves better performance (larger AUC value and smaller SD, DLE, RMSE value) than wMNE, LORETA, SBL and VB-SCCD in terms of reconstructing extended sources. However, due to page limit, for the experimental data analysis, we only applied \({{L}_{1}}\)R-VSSI to analyze the epilepsy EEG data from Brainstorm. In our future work, we will also apply the proposed method to more clinical data to verify the performance, which will promote the development of neuroscience research and clinical medicine. Additionally, since the minimization of sources in the variation domain does not limit the global energy, \(L_1\)R-VSSI tends to underestimate the amplitude of sources [10, 16]. Our future work will also consider adding regularizers in other transform domains to improve the accuracy of amplitude.