Introduction

Atrial fibrillation (AF) is currently one of the major cardiovascular challenges in the developed world.44 It is the most common supra-ventricular arrhythmia in clinical practice, affecting approximately 1.5–2% of the general population.60 More than 33 million individuals worldwide have AF, and 5 million of new cases are approximately diagnosed each year.12 Moreover, its incidence and prevalence increase with age, thus around 15% of people older than 80 years suffer from this arrhythmia.60 However, the reasons for the increasing impact of AF on aged population49 as well as the underlying electrophysiological mechanisms of this arrhythmia50 are still topics of intense discussion due to their incomplete understanding. Indeed, this aspect hampers effective diagnosis and treatment of the arrhythmia, thus causing that 15% of the healthcare budget in cardiac diseases has to be spent on the management of AF.57

Within this context, the analysis of atrial electrical activity is of paramount relevance to improve current knowledge on the mechanisms responsible for initiation, maintenance and perpetuation of AF.23 This information can be analyzed either from surface recordings, i.e., from the electrocardiogram (ECG),3,22 or from intracardiac recordings, i.e., from electrograms (EGM).24 Whereas the ECG records the superposition of the heart’s electrical activity from the patient’s thorax, the EGM directly captures local electrical activity on the cardiac tissue. Therefore, this latter recording provides more detailed information about the electrical status of any particular area of the myocardium than the ECG, thus being nowadays the hallmark of cardiac electrophysiology.15 In fact, the EGM can provide accurate information about the time, direction and complexity of atrial activations within the field of view of the recording electrodes.26 As a result of this near-field information, current cardiac mapping systems are able to provide nonfluoroscopic electroanatomic maps which help in determining the site of arrhythmia origin, its mechanism as well as the particular cardiac structures sustaining AF.29 Knowledge of the arrhythmia mechanism and circuit is critical in planning arrhythmia ablation, which is currently considered as the first-line therapy for the treatment of AF.35

Every single EGM responds to the potential differences recorded at two separated electrodes. In this regard, two different signals are widely used in clinical electrophysiology, namely unipolar and bipolar recordings.53 Unipolar EGMs are obtained by placing an electrode in the heart’s surface and the second one distant from it to serve as a reference. This recording provides information about the impulse propagation direction, but substantial far-field signals, generated by depolarization of remote tissues, are also acquired.53 On the contrary, bipolar EGMs are obtained by attaching closely two electrodes in a specific area of the cardiac tissue, thus mostly providing information about local electrical activity.53 The main disadvantage of this recording is the impossibility of quantifying wavefront propagation direction.26 Hence, both types of recordings contain complementary information and, usually, both are simultaneously acquired to assist cardiac mapping techniques.26

Although to a lesser extent than for the surface ECG, noise and other nuisance signals also disturb EGM acquisition.55 From an electromagnetical point of view, the electrophysiology laboratory is an extraordinarily noisy environment,55 where a wide conglomerate of electrical equipments, some of them connected to the patient’s body, are simultaneously operating. These equipments provoke high-frequency and broadband noises, which are unavoidably recorded during EGM acquisition.55 Another important source of interference is the electrical mains, which often contaminates EGM recordings with a component of 50 or 60 Hz.56 Finally, the EGM can also be corrupted by high-frequency noise generated from non-cardiac muscle activity of the patient as well as by a low-frequency component, such as the baseline wander, resulting from respiratory or catheter movements and an unstable catheter contact.56

Considering that a signal-to-noise ratio (SNR) of 20 dB has been proposed as a recommended minimum recording quality to discern small changes in any EGM,55 the application of an adequate preprocessing method able to reduce noise and preserve waveform pattern is of paramount relevance for successful and reliable further analysis.38 However, this aspect has been the forgotten issue of cardiac electrophysiology studies, with very reduced attention in the literature, thus lacking of thorough validation analysis.38 In fact, the EGM is regularly filtered during its acquisition to reduce power-line interference, far-field signals as well as low- and high-frequency noises, but no alternative preprocessing is often considered.37,41 Most of the time, a filtering-based strategy, proposed by Botteron and Smith more than 20 years ago,8 is used to highlight atrial activations, thus making their detection easier. Although this approach has proven to be useful to facilitate activation detection8 as well as to quantify the dominant frequency (DF) of AF from the EGM,9 the resulting waveform morphology as well as the DF estimation can sometimes be significantly altered, as will be quantitatively demonstrated in the present work. Therefore, regular EGM preprocessing should be used carefully, or even avoided, over applications relaying on EGM analysis that go beyond considering exclusively activation times.

On the other hand, although filtering can alter EGM amplitude, timing and morphology,53 its effect on parameters widely used to characterize both in time and frequency the EGM has not been quantified yet. The goal of the present work is therefore to analyze the ability of regular preprocessing approaches to preserve waveform integrity and reduce noise both from unipolar and bipolar EGM recordings. In addition, alternative and efficient denoising methods successfully applied to the ECG will be explored and proposed.28 The manuscript will prove that methods based on the Wavelet transform (WT) and the empirical mode decomposition (EMD) are able to provide cleaner and less morphologically altered EGM recordings, thus resulting in an optimal starting point for the proper interpretation of invasive AF recordings and a more truthful and accurate quantification of atrial activity dynamics.

The remainder manuscript is organized as follows. “Methods” section describes the analyzed databases containing synthesized and real EGMs as well as the proposed denoising algorithms. In this section the parameters used to quantify noise reduction and EGM morphology preservation reached by each algorithm are also presented briefly. “Results” section summarizes the obtained outcomes, which are then commented in “Discussion” section. Finally, “Conclusions” section presents the concluding remarks of this study.

Methods

Study Population

The proposed denoising algorithms have been first validated using a database of synthesized unipolar and bipolar EGM recordings, since these signals allow to quantify accurately the achieved noise reduction as well as waveform integrity preservation. To mimic EGMs as realistic as possible, an approach similar to the one presented by Oesterlin et al. 42 was used. Thus, atrial activations from a dataset of 40 real EGMs (20 unipolar and 20 bipolar) were manually annotated by two experts. These signals were acquired with a sampling frequency of 1000 Hz after a band-pass filtering between 0.5 and 500 Hz and a band-stop filtering around 50 Hz. Moreover, to avoid far-field ventricular contamination, they were also recorded under adenosine administration.5 This drug caused a transient atrioventricular block with a duration between 5.5 and 8.6 s and, therefore, only atrial activations were recorded during this interval. A window of 90 ms in length was then centered on each atrial activation, such that all were aligned in time. Next, sets of 10 single activations were averaged to generate noise-free templates. Every activation entered the template just in case its cross-correlation with the other template members was higher than 95%. A total of 18 and 26 templates for unipolar and bipolar recordings were respectively collected and normalized in amplitude. To synthesize an artificial EGM, a template was randomly selected and the probability of including others in the EGM was weighted with a gaussian distribution adapted to their cross-correlation. The DF for each recording was randomly chosen between 3 and 12 Hz and the distance between successive atrial activations was generated by a normal distribution centered on that frequency with a standard deviation of 25 ms. Finally, the amplitude of the atrial activations was also modulated according to real signals. For that purpose, the upper and lower envelopes of real EGMs were extracted by cubic splines interpolation. Sets of 100 unipolar and 100 bipolar realistic, completely clean and 7 second-length EGMs were generated. This clean signal will be referred to as x(n) from now on.

To assess the performance of the proposed denoising algorithms, standard EGM noise had to be added, thus obtaining noisy recordings referred to as \(\tilde{x}(n).\) The noise was generated following the approach described in Ref. 14. Briefly, it was modeled as an autoregressive (AR) system, whose coefficients were derived by curve fitting from the gaps between atrial activations of real EGMs. The optimal order for the AR model was estimated by the Akaike information criterion, ranging from 10 to 35 for each analyzed EGM interval.7 Thus, choosing randomly an order within this range and the mean coefficients for all the real signals, different noise sets were synthesized and normalized in amplitude to generate EGMs with values of SNR of 30, 25, 20, 15, 10 and 5 dB. Figure 1 shows an example of unipolar and bipolar recordings synthesized with different levels of noise.

Figure 1
figure 1

Example of two second-length synthesized recordings of unipolar (a) and bipolar (b) EGMs in AF with increasing levels of noise. The first row corresponds to clean EGMs, x(n). The amount of added noise is different for unipolar than for bipolar synthesized recordings just to provide a better graphical idea on the noisy recordings, \(\tilde{x}(n)\).

For the evaluation of the proposed algorithms over real scenarios, the aforementioned set of 40 real EGMs was doubled, so that a total of 80 real EGMs (40 unipolar and 40 bipolar) were analyzed. The study protocol was approved by the Institution’s Research and Ethics Committee and all the enrolled patients provided informed written consent. All signals were acquired from 50 different patients undergoing catheter ablation of AF. The expansion with the second group of real recordings was included with the aim to minimize any bias that could arise due to the set of synthesized recordings.

Studied Denoising Methods for EGM Acquisition

Next subsections describe the filtering approach regularly applied to EGM recordings. Next, alternative efficient methods applied to ECG noise reduction will be described and adapted to the case of intracardiac unipolar and bipolar EGM recordings.

Regular EGM Denoising

Some atrial structures, such as the pulmonary veins, can produce very rapid deflections with a substantial high-frequency content.55 Thus, as an optimal trade-off between noise reduction and clinical information preservation, both unipolar and bipolar EGMs are regularly low-pass filtered with a cut-off frequency around 300 Hz.38,53 Contrarily, both kinds of recordings are often high-pass filtered with different cut-off frequencies. Indeed, this lower corner frequency is rarely higher than 3 Hz for unipolar EGMs.26 In this way baseline wander is removed and some clinical information is preserved. Moreover, filtering at higher corner frequencies has proven to alter recording morphology, thus obscuring the proper detection of the wavefront propagation direction, which is the typical target in unipolar EGMs.53 Surprisingly, although bipolar EGMs are more robust to noise than unipolar ones, since the two electrodes capture a similar far-field contamination which can be easily reduced by subtraction, a more aggressive high-pass filtering with a higher cut-off frequency of about 30 Hz has been regularly applied to these recordings.26 In fact, the main goal of this preprocessing step is to reduce far-field ventricular contamination. However, the main atrial component, normally between 3 and 12 Hz in AF,13 is also strongly reduced. Obviously, this operation distorts substantially the EGM morphology and reduce its amplitude, thus turning the recording more complex and fractionated due to the severe reduction of its main atrial component.26 Nonetheless, a variety of clinical studies have reported that some useful information still remains in the signal.53 With the aim to reproduce the same steps of regular filtering, the synthesized recordings were preprocessed by second-order digital Butterworth filters following a forward/backward filtering strategy, which avoided the introduction of any phase delay.33 The bandwidth of the applied filters was 3–300 and 30–300 Hz for unipolar and bipolar EGM recordings, respectively.

Wavelet Transform Denoising

Given its interesting ability to deal with non-stationary signals containing sharp spikes and discontinuous intervals, the WT has been widely used for ECG denoising.51 This transformation allows to decompose a signal at different time and frequency scales. Indeed, it can be defined for the continuos signal x(t) as

$$C_{a}(b)=\int _{-\infty }^{\infty }x(t)\psi_{a,b}^{\ast}(t)\text{d}t,$$
(1)

where the asterisk stands for the complex conjugate operation and \(\psi _{a,b}(t)\) represents scaled and translated versions of the basis function \(\psi (t),\) named mother wavelet, such that

$$\psi _{a,b}(t)=|a|^{-\frac{1}{2}}\psi \left(\frac{t-b}{a}\right).$$
(2)

The factors of scale a and translation b have to meet two conditions, i.e., \(a,b \in \mathbb{R}\) and \(a \ne 0.\) Obviously, the resulting wavelet coefficients \(C_a(b)\) depend on these factors, such that if they are discretized for only taking integer values power of two (i.e., \(a=2^j\) and \(b=k\cdot 2^j\)), non-redundant information is obtained.34 This discrete WT (DWT) has been widely used in biomedical applications,1 since it can be easily implemented by a bank of low-pass and high-pass finite impulse response filters followed by decimation stages, such as Fig. 2 displays. The signals obtained from the high-pass and low-pass filtering processes consist of the detail \(\text{DC}_j(k)\) and approximation wavelet coefficients \(\text{AC}_j(k),\) respectively.

Figure 2
figure 2

Detail of the filter bank structure for DWT computation.

A generalization of this transformation able to offer additional information is the Wavelet packet transform (WPT).34 In contrast to the DWT, where only approximation coefficients are low- and high-pass filtered, both detail and approximation coefficients are split at each level into finer components. In this case optimal decomposition can be obtained by minimizing the entropy during the process.54 Thus, a vector of wavelet coefficients is only split when the entropy is reduced. Although several approaches have been proposed to compute entropy from wavelet coefficients, the common Shannon Entropy (ShEn) has been used in this work.34 Both DWT and WPT have been applied to the recordings, so that once the noisy EGM \(\tilde{x}(n)\) was decomposed into a specific number of scales, the wavelet coefficients were thresholded and used to reconstruct back a denoised signal \(\hat{x}(n).\) 51

Two thresholding rules, i.e., hard and soft, have been widely used for ECG denoising.30 In hard thresholding the wavelet coefficients below a threshold \(\lambda\) are set to zero, whereas in soft thresholding coefficients lower than \(\lambda\) are set to zero and the remaining ones are reduced by that value.54 Obviously, the thresholding rule together with the value of \(\lambda\) determine the resulting signal. Hence, all possible combinations between hard and soft thresholding rules and the most typical approaches to compute \(\lambda\) have been analyzed both for DWT and WPT.

To this last respect, considering jointly all the wavelet coefficients from a decomposition, thus producing a vector of S samples in length, two different values of \(\lambda\) have been first defined following16

$$\lambda_{\rm FIX}=\sqrt{2\cdot \ln (S)},$$
(3)

which stands for a fixed thresholding rule for all the wavelet scales, and

$$\lambda_{\rm MMX}=0.3936+0.1829\cdot \log (S),$$
(4)

which performs a minimax estimation of the thresholding rule for every decomposition level. In addition, an adaptive threshold for each vector of wavelet coefficients has also been computed from the Stein’s unbiased risk estimate (SURE), i.e.,

$$\lambda_{\rm SURE}=\sqrt{2\ln (S_{\rm C}\cdot \log _2(S_{\rm C}))},$$
(5)

where \(S_{\rm C}\) is the length of the analyzed vector.17 Finally, a variant of this approach (named heuristic SURE, \(\lambda_{\rm HSURE}\)) where \(\lambda_{\rm SURE}\) was replaced by \(\lambda_{\rm FIX}\) in very noisy conditions, has also been considered following the recommendations in Ref. 17. For a more effective denoising, these thresholds were rescaled by the standard deviation of the noise estimated from the highest frequency vector of wavelet coefficients.16 The standard deviation was computed as

$$\hat{\sigma }=\text{median}_{k=1\ldots ,S_{C_1}}\{|C_1(k)|\}/0.6745.$$
(6)

Finally, it has to be noted that no systematic approach exists for optimal selection of the mother wavelet, even though the use of different functions to analyze the same signal can sometimes produce notably different results.45 In denoising applications, the mother wavelet is normally determined by its similarity with the fundamental signal pattern under analysis. In this way the highest wavelet coefficients are obtained in the decomposition, thus turning thresholding of contaminant signals easier.51 In the present study, all the functions from the most common orthogonal and biorthogonal wavelet families, including Daubechies, Biorthogonal, Coiflets, Symlets, Reverse Biorthogonal and Discrete Approximation of Meyer, were tested. A total of 54 functions were then studied. Note that, only when orthogonal functions are used, the information stored in each vector of wavelet coefficients is not repeated elsewhere and the original signal can be accurately reconstructed. On the other hand, the number of decomposition levels is also a key parameter to achieve satisfactory results. The maximum level depends on the length of both the signal and the decomposition filter.45 In this case, that value was 8 and, hence, decomposition levels between 3 and 8 were analyzed.

Empirical Mode Decomposition Denoising

In contrast to WT, EMD does not require any a priori known basis to decompose a signal into a set of oscillatory components ranging from high to low frequencies, which are named intrinsic mode functions (IMFs).25 Hence, this transform also behaves as a filter bank, but without a predefined cut-off frequency.20 This property makes EMD especially useful for the processing of nonlinear and non-stationary signals. Indeed, it has been widely used for denoising of ECG as well as other physiological signals.6,32

Every IMF must have the same number of extrema and zero-crossings or differ at most by one and, additionally, it must be symmetric with respect to local zero mean.25 Thus, for a successful decomposition into IMFs, the signal must have at least two extrema, i.e., one maximum and one minimum. Given that any noisy recording \(\tilde{x}(n)\) in the present study met this condition, the decomposition process started with the identification of all local maxima and minima in the signal. Next, these points were interpolated by a cubic spline curve to reconstruct the upper \(e_{\rm u}(n)\) and lower \(e_{\rm l}(n)\) envelopes, respectively. The mean envelope was obtained as a simple average, i.e.,

$$m(n)=\frac{e_{\rm u}(n)+e_{\rm l}(n)}{2},$$
(7)

and, it was then subtracted from the contaminated signal to obtain

$$h_1(n)=\tilde{x}(n)-m(n).$$
(8)

The same process was iteratively repeated on the resulting signal \(h_k(n)\) until the following condition was satisfied

$$\frac{|h_{k-1}(n)-h_k(n)|^2}{h_{k-1}(n)^2}\le 0.3,$$
(9)

thus \(h_k(n)\) being the first IMF, named \(c_1(n).\) The residue \(r_1(n)=\tilde{x}(n)-c_1(n)\) still contained useful information and, hence, the process was repeated on this signal to obtain \(c_2(n).\) In the same way \(c_3(n),\ldots ,c_{10}(n)\) were computed, such that the original signal could be reconstructed as

$$\tilde{x}(n)=\sum _{k=1}^{10}c_k(n)+r_{10}(n).$$
(10)

Bearing in mind that IMFs from the lower levels present higher frequency components and vice versa, previous works have proven that high-frequency noise can be successfully removed from some physiological signals by reconstructing the recording without the first functions.10 Consequently, the present work has analyzed EMD-based denoising performance by discarding up to 5 IMFs, such that

$$\hat{x}(n)=\sum _{k=K}^{10}c_k(n)+r_{10}(n), \quad \text{for} \quad K=2,\ldots ,5.$$
(11)

Waveform Integrity and Noise Reduction Assessment

The ability to reduce noise and preserve waveform integrity of the proposed algorithms was validated in terms of time and frequency parameters widely accepted to characterize both unipolar and bipolar EGMs. Firstly, morphological alterations caused by the denoising methods were quantified by computing the cross-correlation and the root mean square error (RMSE) between every pair of denoised recording \(\hat{x}(n)\) and its corresponding original signal x(n). Since the common Pearson’s correlation coefficient does not account for amplitude differences between signals, an adaptive signed correlation index (ASCI) was used for a more accurate morphological comparison.31 Considering N sample-length signals, the ASCI was defined as

$$\text{ASCI}(x(n),\hat{x}(n))=\frac{1}{N}\sum _{k=1}^{N}x(k)\otimes \hat{x}(k),$$
(12)

where \(\otimes\) denotes the signed product of two dichotomized scalars as

$$x(n)\otimes \hat{x}(n)=\left\{ \begin{array}{rl} 1 &{} \quad \text{if}\;|x(n)-\hat{x}(n)| \le \beta ,\\ -1 &{}\quad \text{if}\;|x(n)-\hat{x}(n)| > \beta . \end{array} \right.$$
(13)

The threshold \(\beta\) was experimentally selected as 5% of the standard deviation of x(n). Also, note that the RMSE was normalized to the root mean square of x(n), i.e.,

$$RMSE=\sqrt{\frac{\sum _{k=1}^N(x(k)-\hat{x}(k))^2}{\sum _{k=1}^Nx(k)^2}}.$$
(14)

A metric proposed to quantify the whole EGM morphology and, thus, to discern between fractionated and non-fractionated signals, is ShEn.37 Hence, this parameter was also computed to assess morphological variations after denoising. In brief, the EGM was quantified using L uniform and equal-size amplitude levels. Then, the occurrence probability of the ith level (\(p_i\)) was estimated from its repetition rate and ShEn was computed as

$$\text{ShEn}=\frac{-1}{\log{L}}\sum _{k=1}^{L}p_k\cdot \log{p_k}.$$
(15)

Moreover, the morphological comparison between atrial activations has been widely used as a measure of AF organization, which has proven to be useful in a variety of clinical scenarios.46 In contrast to the above defined indices ASCI, RMSE and ShEn, this morphological comparison only considers atrial activations themselves, thus discarding the remaining EGM content in between.19 Given that this methodology requires an accurate identification of every single activation, the algorithm introduced by Ng et al. 40 has been used in this study. More precisely, the EGM was first processed with the approach proposed by Botteron and Smith,8 i.e., the signal was band-pass filtered between 40 and 250 Hz, then rectified and, finally, low-pass filtered with a cut-off frequency of 20 Hz. Zero-phase third-order Butterworth filters were used. The peak with the highest amplitude in the resulting signal was marked as the first atrial activation. Next, peaks with decreasing amplitudes were identified considering a blanking period of 50 ms around each one. The process finished when the mean distance between successive peaks was lower than 275 ms and, moreover, that distance was also lower than its median plus 5 ms or the detected peak magnitude was 20% less than the previous one.40 Finally, activations within intervals 1.5 times greater than the median distance between successive peaks were also searched. The peaks detected from clean and denoised signals, x(n) and \(\hat{x}(n),\) were compared in terms of sensitivity (Se), positive predictivity (\(\text{P}^+)\) and location error (LE). Thus, Se was defined as the rate of atrial activations properly identified, whereas \(\text{P}^+\) was the proportion of true activations from all detected. Lastly, the LE was defined as the time difference between activation locations in the clean signal with respect to the noisy one for every pair of x(n) and \(\hat{x}(n)\) considered.

Once the atrial activations were detected, morphological similarity among them was quantified through the algorithm presented by Faes et al. 19 In brief, the barycenter for every atrial activation was identified using a moving average noncausal filter with 90 coefficients. Centered on this point, which divides the activation into two equal parts, a window of 90 ms was defined. To prevent morphological alterations from factors unrelated to the arrhythmia, each activation was normalized by the Euclidean norm of all atrial activations in the recording.19 Finally, a repetition rate of M activations was defined by computing the relative number of similar atrial activation pairs, so that

$$\rho (\xi )=\frac{2}{M\cdot (M-1)}\sum _{i=1}^M\sum _{j=i+1}^M\Uptheta (\xi -d_{ij}),$$
(16)

where \(\Uptheta (x)\) is the Heaviside function and \(d_{ij}\) is the distance between ith and jth activations, defined as the arc cosine of their scalar product.19 The threshold \(\xi\) was chosen as \(\pi /3\) radians.19

The proposed denoising algorithms have been also validated from the spectral domain through the use of two indices. In fact, the goal of most frequency-domain techniques applied to EGMs in AF is to estimate the local atrial rate, i.e., the DF.39 For that purpose, the preprocessing introduced by Botteron and Smith has been widely used and is applied by default in many recording systems.39 Indeed, it has been proven that this approach removes detailed features of the atrial activations with the aim to emphasize their timing.9 In the present work, the resulting signal from every denoising method was spectrally characterized through the Welch Periodogram with a Hamming window of 4096 points in length, 50% overlapping and 10240-point FFT.18 These computational parameters provided a frequency resolution of 0.1 Hz. Next, the DF was computed as the peak with the highest spectral energy. Finally, to estimate spectral preservation, the regularity index (RI) was computed due to its wide use as an estimator of AF organization in clinical studies.48 The RI is defined as the power spectral density ratio of the 2 Hz band centered on the DF with respect to the physiological frequency range of AF, i.e., 3–20 Hz.

On the other hand, given that a completely clean signal is unavailable in real scenarios, the parameters introduced above cannot assess the performance of the denoising algorithms over real EGMs. For this case, the yielded noise reduction is assessed through the SNR estimated as the power ratio between the denoised signal and the corresponding extracted noise,52 i.e.,

$$\widehat{\text{SNR}}=10\cdot \log{\left( \frac{{\sum _{k=1}^{N}\hat{x}(k)^2}}{\sum _{k=1}^{N}\hat{n}(k)^2}\right) },$$
(17)

where the noise power is estimated as the power difference between the noisy and denoised recordings, so that

$$\sum _{k=1}^{N}\hat{n}(k)^2 = \sum _{k=1}^N(\tilde{x}(k)-\hat{x}(k))^2.$$
(18)

A perfect denoising stage will extract all the noise and only the noise. Thus, yielding the same result in estimated \(\widehat{\text{SNR}}\) than the real SNR of the recording. On the contrary, a poor denoising stage will not remove all the noise, or even potentially will consider part of the signal as noise. Hence, \(\widehat{\text{SNR}}\) will be far away from the real SNR, indicating that noise still remains or that part of the signal has been considered as noise and removed. As a consequence, the ratio between \(\widehat{\text{SNR}}\) and SNR will not be one to one. For synthesized EGM recordings and with the aim to validate the index, \(\widehat{\text{SNR}}\) has been directly compared to the generated SNR. For real EGM recordings and each denoising methodology, the estimated \(\widehat{\text{SNR}}\) has been compared with the optimally estimated \(\widehat{\text{SNR}}_{\rm O},\) that was provided by the denoising approach with best results in the synthesized signals.

Finally, note that all the described algorithms have been developed under Matlab 8.6, which is a popular mathematical programming platform from The MathWorks, Inc. Moreover, they have been run on a desktop computer with 3.2 GHz Intel Core i5 processor under Mac OS X 10.11.6, thus providing the results that will be presented in the forthcoming section.

Results

Synthesized EGM Recordings

The effect of noise on the synthesized recordings is quantified in Table 1. Some of the parameters are preceded by a Δ thus representing normalized relative variations between clean and noisy recordings. As can be observed, the defined indices are able to reflect accurately how the presence of increasing levels of noise degrade the EGM waveform morphology and alter its time and frequency features. Although this finding applies both to unipolar and bipolar EGM signals, a slightly different behavior between both kinds of recordings was observed in some cases, as for ΔShEn and ΔDF. Thus, the bipolar signals were more sensitive to noise in terms of ΔShEn, while an opposite trend was noticed for ΔDF. Another interesting finding is that \(\text{P}^+\) decreased notably as the noise increased, thus highlighting a substantial rise in the number of falsely detected atrial activations under noisy conditions.

Table 1 Morphological alteration results caused by noise addition on the synthesized recordings as a function of the signal-to-noise ratio (SNR).
Figure 3
figure 3

Waveform integrity preservation and time and frequency indices provided by the analyzed algorithms as a function of noise for synthesized unipolar EGMs. Note that second-order Coiflet and second-order Daubechies wavelet functions were used for DWT and PWT, respectively.

Regarding how atrial waveform integrity is preserved by the analyzed denoising methods, the main outcomes are summarized in Fig. 3 for synthesized unipolar EGMs. This figure only displays the best results for DWT-, WPT- and EMD-based denoising algorithms. Additional results obtained from a wider variety of tested parameters for each denoising strategy can be found in the online section of Supplementary Materials. No significant differences were noticed both for DWT and WPT from the analyzed wavelet functions, wavelet levels of decomposition, thresholds and thresholding rules. Nonetheless, for DWT the highest degree of waveform morphology preservation and noise reduction was achieved with the second-order Coiflet function with 8 decomposition levels and soft thresholding using \(\lambda_{\rm SURE}.\) With these parameters, WPT also provided good results, however its highest performance was obtained with the second-order Daubechies function selecting 5 decomposition levels and soft thresholding using \(\lambda _{HSURE}.\) Finally, the EMD-based technique provided the highest waveform integrity by discarding the first IMFs, i.e., for \(K=2.\)

Coherently, most of the tested methods provided more altered features of signal morphology and frequency content as the SNR decreased. However, the two WT-based algorithms provided the cleanest and most unaltered unipolar recordings. Their performance was superior to that of regular filtering in any analyzed case and they provided especially good results for moderate or high levels of noise. Observe in Fig. 3 how regular filtering alters waveform morphology and how EMD provided poorer results than WT-based methods. Thus, even for the most favorable case with a clean recording of 30 dB in SNR, the gap in waveform integrity preservation between regular filtering and WT-based methods was significantly large. Metrics such as ASCI or RMSE provided improvements in waveform preservation of about 20% (ASCI of 79.2% vs. 99.9%). Furthermore, in noisy cases, such as SNR of 5 dB, the waveform alteration introduced by regular filtering was extremely high compared to WT-based methods (ASCI of 26.8 vs. 60.4%). Other interesting observation refers to ΔDF because the detected dominant frequency variation has been always notably higher for regular filtering, thus involving deviations of up to 37.2%. Taking into consideration the relevance that tiny differences in dominant atrial frequency maps have reached in recent works as a tool for guiding catheter ablation,4 this finding is of great interest.

Figure 4
figure 4

Waveform integrity preservation and time and frequency indices provided by the analyzed algorithms as a function of noise for synthesized bipolar EGMs. Note that second-order Coiflet and second-order Daubechies wavelet functions were used for DWT and PWT, respectively.

Regarding synthesized bipolar recordings, Fig. 4 summarizes the obtained results. As before, only the best denoising results are included, so that the remaining additional outcomes can be found in the online section of Supplementary Materials. As in the unipolar case, most of the methods provided poorer results as noise increased. The exception to this rule was regular filtering, in which a significant high degree of waveform distortion was introduced for any SNR. More concretely, in the most favorable case with SNR of 30 dB, regular filtering distorted notably the atrial waveform in comparison to WT-based methods (ASCI of 39.6 vs. 99.8%). In fact, the presence of such a low level of noise did not distort the atrial waveform, but regular filtering provoked considerable distortion levels by itself. Bearing in mind that such a kind of filtering is the most extended procedure in routine analysis of bipolar EGMs, this result is also specially significant. Furthermore, it has to be highlighted the poor outcomes provided by EMD which, in some cases, were below regular filtering as Se, LE or ΔDF.

Finally, the relation between average values of estimated \(\widehat{\text{SNR}}\) and generated SNR for synthesized recordings is graphically sketched in Fig. 5. As has been detailed back in “Waveform Integrity and Noise Reduction Assessment” section, the \(\widehat{\text{SNR}}\) has been defined to estimate noise reduction performance on real EGMs. However, its application to synthesized EGMs is also of interest in order to verify its validity and robustness. Two main considerations can be done in view of the results for synthesized EGMs. Firstly, WT-based methods present a clearly linear and progressive relation between \(\widehat{\text{SNR}}\) and SNR, thus indicating that they are quite close to an ideal denoising approach. In fact, these strategies are able to remove most part of the noise in the recording and only the noise. Secondly, EMD and especially regular filtering do not present a one-to-one ratio between estimated and generated SNRs. Therefore, these methods are unable to remove a significant part of the noise in the signal or, even worse, they also remove part of the signal which is mistakenly considered as noise.

Figure 5
figure 5

Relation between estimated \(\widehat{\text{SNR}}\) and generated SNR achieved by every denoising algorithm in the case of (a) synthesized unipolar EGMs and (b) synthesized bipolar EGMs. Note that second-order Coiflet and second-order Daubechies wavelet functions were used for DWT and PWT, respectively.

Real EGM Recordings

The only parameter that can be used to assess noise reduction performance as well as waveform preservation when dealing with real recordings is the estimated \(\widehat{\text{SNR}}.\) To this respect, Fig. 6 shows the ratio between \(\widehat{\text{SNR}}\) and optimally estimated \(\widehat{\text{SNR}}_{\rm O},\) as described back in “Waveform Integrity and Noise Reduction Assessment” section. This last index was computed making use of the DWT-based denoising strategy to obtain results for regular filtering, PWT-based and EMD-based methods because DWT yielded the best noise reduction and waveform preservation results with synthesized recordings. In the case of DWT-based denoising, the \(\widehat{\text{SNR}}_{\rm O}\) was provided by the WPT-based approach, which also provided excellent results. In view of the results for real recordings in the figure, the WT-based methods offered a denoising and waveform preservation behavior with outstanding performance. Thus, the estimated \(\widehat{\text{SNR}}\) followed a clear line very close to the one-to-one ratio with \(\widehat{\text{SNR}}_{\rm O}.\) This is indicative of the great ability that these methods provide for removing most part of the noise and only the noise, both in unipolar and bipolar recordings. On the contrary, EMD and especially regular filtering, presented a significantly limited ability of noise reduction and waveform preservation even in the most favorable cases of very low noise with SNR of 30 dB.

Figure 6
figure 6

Relation between estimated \(\widehat{\text{SNR}}\) and optimally estimated \(\widehat{\text{SNR}}_{\rm O}\) achieved by every denoising algorithm in the case of (a) real unipolar EGMs and (b) real bipolar EGMs. Note that second-order Coiflet and second-order Daubechies wavelet functions were used for DWT and PWT, respectively.

Finally, with the aim to provide a wide overview on the performance of the analyzed denoising strategies over real EGMs, Figs. 7 and 8 show a set of examples for unipolar and bipolar real EGMs, respectively, with different morphologies and levels of noise, together with the corresponding result provided by each methodology. As can be observed, even for reduced noise (Figs. 7a, 7b and 8a, 8b), regular filtering performs extremely poor in EGM denoising compared to WT-based methods, whereas EMD performs quite better that regular filtering but poorer that DWT or WPT. Furthermore, waveform integrity is completely reduced and altered in bipolar EGMs processed by regular filtering. On the other hand, when the EGM is relatively noisy (Figs. 7c, 7d and 8c, 8d) WT-based methods still behave with outstanding performance, while EMD is unable to remove noise with similar efficiency but does not seem to provoke significant waveform distortion. In contrast, regular filtering performs extremely bad both in EGM denoising and waveform integrity preservation, especially in the bipolar case. Observe how resulting amplitudes are notably reduced, even provoking the masking of some atrial activations inside the noise. Obviously, amplitude reduction as well as activation timing lost are the potential source of significant variations in voltage cartography, activation maps, rotor maps or any other operation performed with bipolar EGMs in cardiac mapping systems.

Figure 7
figure 7

Example of waveform integrity preservation and denoising performance of the analyzed methods for real unipolar EGMs under different noisy conditions. The noisy recording \(\tilde{x}(n)\) is filtered out by the different methodologies to provide the denoised resulting EGM \(\hat{x}(n).\) The estimated \(\widehat{\text{SNR}}\) in every case is (a) 20 dB, (b) 17 dB, (c) 13 dB and (d) 10 dB, respectively.

Figure 8
figure 8

Example of waveform integrity preservation and denoising performance of the analyzed methods for real bipolar EGMs under different noisy conditions. The noisy recording \(\tilde{x}(n)\) is filtered out by the different methodologies to provide the denoised resulting EGM \(\hat{x}(n).\) The estimated \(\widehat{\text{SNR}}\) in every case is (a) 20 dB, (b) 18 dB, (c) 12 dB and (d) 9 dB, respectively.

Discussion

The surface ECG, although highly sensitive to noise, still is the most widely used recording for cardiac diagnosis.33 Indeed, noise reduction strategies applied to the ECG have been widely addressed in the last years to improve its characterization and interpretation.33 As a result, a broad variety of denoising methods are available in the literature, including filter banks, adaptive filtering, independent component analysis, neural networks, WT and EMD.28 Contrarily, noise reduction applied to EGMs has received poor attention. Although intracardiac recordings are often less disturbed by noise than the surface ECG, nuisance interferences are still present during its acquisition55 and, therefore, this issue has to be properly addressed. Thus, the exploration of new algorithms to reduce noise in the EGM but, at the same time, preserving waveform integrity still is an unsolved challenge. In fact, most of the EGM recording systems regularly filter quite aggressively the acquired signal,26 even though this preprocessing has proven to alter its morphology. Within this context, the present work has introduced the first systematic and thorough comparison between regular filtering and more refined algorithms, successfully used in ECG denoising, applied to the EGM. The coherence between the results obtained both from synthesized and real EGMs allows us to present some remarkable observations as well as useful recommendations about EGM preprocessing, which will be next picked out.

Comparison Among EGM Denoising Algorithms

To begin with, regular filtering has provided very poor performance in waveform integrity preservation as well as in noise reduction compared to WT-based methodologies. Thus, under low noise conditions, variations of about 20% in waveform preservation were noticed for unipolar EGM recordings between regular filtering and WT-based denoising (see Fig. 3). Furthermore, the bipolar case was even worse, and variations of about 60% in ASCI or RMSE were observed (see Fig. 4). In fact, parameters like ΔShEn and \(\Updelta \rho\) behaved very poorly with regular filtering regardless of the SNR. For noisy cases the situation remained the same for some indices and even worse for many others. Thus, for unipolar EGM recordings parameters like \(\Updelta \rho ,\) \(\text{P}^+,\) ΔDF and ΔRI increased the gap between regular filtering and WT-based methods, whereas in the case of bipolar recordings indices like \(\text{P}^+\) or ΔRI also performed poorer than for low noise conditions. The significant and high-amplitude components between 1 and 30 Hz eliminated by the aggressive regular filtering clearly justify these outcomes. Moreover, this fact also explains the notably low estimated \(\widehat{\text{SNR}}\) by this filtering than the remaining denoising algorithms, as Figs. 5 and 6 show for unipolar and bipolar EGM recordings, respectively. This indicates that, in addition to noise, regular filtering is mistakenly removing significant components of the EGM, as Figs. 7 and 8 display for several examples.

On the other hand, it is interesting to note that results for DWT- and PWT-based methods were quite similar, although both made use of different wavelet decomposition strategies.54 Slightly better outcomes were only observed for bipolar EGMs with the DWT and for unipolar recordings with the WPT-based approach (see Figs. 3 and 4). Moreover, the experimentally obtained optimal wavelet parameters were also greatly different in both methods, especially regarding decomposition levels and denoising thresholds. A similar behavior was also observed in most of the tested wavelet functions. This wide variety of results with respect to levels, thresholds and decomposition strategies could be justified by the fact that notably different morphologies can be found among the analyzed signals and even, sometimes, within the same signal. Nonetheless, the obtained results still confirm the high versatility of the WT. This is an important advantage, since its adaptability allows to remove efficiently noise in a wide variety of recordings, including other cardiac signals such as the ECG28 and the transmembrane fluorescence optical recording from voltage-dependent dyes.59 According to these previous works, WT-based methods have also proven here top performance in EGM waveform preservation and denoising, both under clean or very noisy conditions. To this respect, variations between the synthesized x(n) and denoised \(\hat{x}(n)\) signals lower than 1% were noticed under low noise conditions for the vast majority of indices related to time and frequency morphological preservation (see Figs. 3 and  4). Furthermore, the increasing presence of noise degraded WT-based method performance, but in a significantly lower amount than for regular filtering. Another interesting point was that the rate of falsely detected atrial activations from the denoised signals with these algorithms was drastically reduced even in the presence of high noise contamination. Thus, \(\text{P}^+\) still remained around 90% for a SNR of 5 dB, while this rate decreased around 70 and 65% for regular filtering and EMD, respectively.

From the noise reduction point of view, the superior performance of WT-based methods compared with the other ones has also been demonstrated in Figs. 5, 6, 7, and 8. WT-based algorithms provided recordings with the lowest levels of noise, also preserving substantially waveform morphology, high-frequencies and abrupt atrial activations, their only drawback being a slight reduction in signal amplitude provoked by the soft thresholding.51 This alteration could explain the variations of \(\Updelta \rho ,\) ΔDF and ΔRI higher than 10% observed for very noisy cases with SNR of 5 dB in Figs. 3 and 4. Nonetheless, given the promising results obtained for values of \(\text{SNR} > 5\) dB, the design of more advanced thresholding rules to mitigate this problem does not seem a prime issue. In any case, future works could address this topic in a similar way as with ECG denoising.51

As can also be observed in Figs. 7 and 8, the EMD-based approach resulted in a noisier signal than the WT-based algorithms. This outcome could justify the relatively high variations observed between synthesized x(n) and recovered \(\hat{x}(n)\) recordings in most of the analyzed parameters both for unipolar and bipolar EGM recordings with values of SNR ≤ 15 dB (see Figs. 3 and 4), thus provoking more distorted atrial activations in comparison with the WT-based methods. This poor performance could be explained by the fact that IMFs from lower levels can still contain high-frequency information from the atrial activations which is disregarded when these levels are considered as noise. Given that the present study made use of the most popular EMD-based denoising, which discards complete IMFs,10 the best trade-off between noise reduction and waveform preservation was reached by selecting \(K=2,\) thus the reconstructed signal \(\hat{x}(n)\) presented the noise mainly contained by \(c_2(n)\) and \(c_3(n).\) A similar finding has also been reported by previous works dealing with EMD-based ECG denoising, which have revealed that high-frequency IMFs contain information from the QRS complexes in addition to noise.28

Finally, the increase of computational load associated to the proposed EGM denoising algorithms deserves a comment. Thus, in average for the synthesized and real EGM recordings, the computational time per sample was about 50 ns, 0.8 μs, 5 μs and 1 ms for regular filtering, DWT-, WPT- and EMD-based approaches, respectively. Note that both WT-based algorithms provided a notably better EGM waveform preservation than the regular approach at the cost of increasing computational load. Nonetheless, bearing in mind that computational time limitations can be easily avoided by using specific digital signal processors or complex programmable logic devices, such as FPGAs,21 the WT-based algorithms could be easily run in real-time under current commercial devices, which typically use a comfortable EGM sampling rate of about 1 kHz. Indeed, some recent works have proven that the DWT-based denoising of different physiological signals, even with extensive additional processing, can be successfully run online.11,21 Contrarily, some extra difficulties could arise in the case of EMD due to its significantly high computational load. Nonetheless, in view of its poor results compared with the WT-based methods, its integration into commercial devices does not seem to deserve much attention.

Clinical Implications

Despite the fact that regular filtering provokes significant EGM waveform alteration and poor noise reduction, such aggressive filtering between 30 and 300 Hz is currently a routine practice in electrophysiology studies.26 Although previous clinical studies suggest that useful information still remains in the EGMs filtered by this approach,53 undoubtedly another not yet quantified amount of information is being disregarded when producing such a significantly different resulting signal. Indeed, when a signal is notably distorted, two relevant hazards occur. On the one hand, the association between altered morphological features in the resulting signal and their underlying clinical implications can be unclear. Thus, when morphological and spectral variations higher than 40 and 10%, respectively, between original and preprocessed bipolar EGMs are provoked (even for a SNR as high as 20 dB) on features such as ρ RI or DF (see Fig. 4), the clinical validity of these metrics is completely doubtful. On the other hand, when a preprocessing approach is too aggressive, interesting information from the original signal may be lost, as Figs. 7 and 8 have shown. That information may, or may not, be useful to improve clinical diagnosis or treatment of AF, but clearly it can lead to different clinical results in later stages. In fact, a less distorted signal will allow the obtention of a more realistic view about the atrial activity dynamics through different atrial mapping representations, such as voltage cartography, atrial activation maps, rotor maps or dominant frequency maps.36 Hence, only when preprocessing is able to preserve the original morphology of the EGM, one can be completely sure that the obtained results are totally reliable and in accordance with reality.

Precisely, the current widespread use of the regular approach for EGM filtering could be one of the many reasons explaining inconclusive results provided by a variety of electrophysiological studies. For instance, a standardized ablation protocol for persistent AF has not been achieved yet.58 To this respect, catheter ablation based on targeting areas with high DF has been proposed as an alternative for patients with persistent AF.4,48 Thus, in addition to the common pulmonary vein isolation, atrial sites showing a DF at least 20% higher than their surrounding points are targeted in this protocol.4 However, given the relevant differences observed between DF values obtained from the original and regularly preprocessed EGMs, it is impossible to assure that proper atrial sites are being currently ablated. This aspect may turn the validation of this protocol clinically questionable. The same comment also applies to ablation based on targeting atrial points with complex fractionated EGMs.2 In this case, the presence of high levels of noise after EGM preprocessing can make the successful identification of that fractionated areas very difficult. Obviously, a completely successful ablation guided procedure is unreachable through the sole use of improved EGM preprocessing, however, this stage might be undoubtedly helpful in providing truthful signals aimed at reaching novel insights about AF mechanisms which are unknown to date.

Limitations

Only mid and high-frequency noise reduction has been detachedly analyzed in the present work. Although this is not a realistic scenario, since EGMs also present baseline wandering, power-line interference and far-field ventricular contamination, a more accurate performance assessment of the proposed denoising algorithms is achieved in this way.10 Nonetheless, given that these nuisance interferences could also be removed by using WT-based approaches,27,43,47 the development of an unified methodology to obtain a compact, simple and computationally light preprocessing approach both for unipolar and bipolar EGMs will be explored in future works. Similarly, other alternative denoising algorithms, such as adaptive noise cancellation, will also be assessed within this context.

Conclusions

Regular filtering applied both to unipolar and bipolar EGMs has altered substantially atrial waveform morphology and spectral properties, thus providing distorted information to later analysis as well as stages of cardiac electrophysiology systems. Moreover, it has proven reduced ability to remove moderate levels of noise. Contrarily, denoising algorithms based on Wavelet transform have shown a great capability to reduce huge levels of noise preserving EGM time morphology and spectral distribution. Hence, the use of an adequate preprocessing is recommended in routine EGM acquisition. In this way, undistorted information may provide new insights about the mechanisms triggering and maintaining AF, thus improving current understanding, management and treatment of the arrhythmia.