1 Introduction

Coastal surges are natural hazards that can lead to considerable casualties and material losses (Baxter 2005; Dube et al. 2009; Melton et al. 2010; Vinet et al. 2013; Shepard et al. 2013). The sea level is a complex signal, with a deterministic part and a stochastic one. The deterministic term is the tidal component, which corresponds to the astronomical influence of the Earth–Moon–Sun system. It is classically modeled using harmonic analysis, as a sum of monochromatic trigonometric terms. The amplitudes of the first and main terms of this harmonic decomposition depend on the shape of the local oceanic basins; they are obtained using numerical methods taking into account the topography and bathymetry (Pugh and Woodworth 2014). The harmonic constants can be numerically estimated in all coastal locations, providing a tidal signal. In France, this is done by the SHOM (“Service hydrographique et océanographique de la marine”, hydrographic and oceanographic services of the French Navy): It provides official tidal predictions for all coastal positions of the French littoral. The stochastic component is mainly the surge, which corresponds to the influences of meteorological phenomena. It is the result of generally unknown nonlinear interactions between atmospheric pressure, water and atmospheric temperatures, winds, waves, and turbulence (Pugh 2004; Pugh and Woodworth 2014). For a positive surge, the measured sea level is larger than the predicted and calculated one (see Fig. 1, a typical example extracted from the Boulogne series). The most extreme surge levels offshore can reach 7–9 m (Pugh 2004), but are generally much lower along the coasts of France (Ullmann et al. 2007; Bertin et al. 2015).

Here, we consider data recorded in three ports of the North East France. It is a densely populated region with many sandy beaches and coastal dunes. The sea is macro-tidal, with an average tidal range varying from about 4.5–6.2 m along the coast of the region. In this region, storm surges can sometimes exceed 2 m (Chaverot et al. 2008). In such system, a positive surge occurring during a low tide is not dangerous, whereas when the positive surge occurs at or near high-level tide, there can be a dramatic coastal flooding since some parts of the coastal zone, behind protecting coastal dunes, are low-lying and hence quite vulnerable (Hequette 2010; Maspataud et al. 2013). Erosion is also linked to the water level, when waves can reach the coastal dunes, which in normal conditions are not in contact with the sea (Pye and Blott 2008). It has been shown that coastal dunes erosion is strongly dependent on the duration of the positive surge event during a flood (Ruz et al. 2009). The magnitude of the erosion process is strongly linked to the maximum height of the sea level reached during the surge event (Esteves et al. 2012). There are potentially high sea levels for each spring tide, but the largest morphological modifications of the littoral zone, and material damages occur when there is a strong meteorological surge during a high spring tide (Betts et al. 2004; Jones and Davies 2007). This was the case during Xynthia storm that hit the French Atlantic coast in February 2010 (Bertin et al. 2014; Vinet et al. 2013).

A good understanding of the statistical and dynamical properties of the sea level time series is hence needed, in order to better know the frequency and the probability of occurrence of extreme sea levels and to prevent coastal flooding hazards. This is particularly critical for low-lying coastal regions, threatened by marine flooding and changing sea levels (Haigh et al. 2011; Weisse et al. 2012; Maspataud et al. 2013).

Here, we analyze sea level time series from the ports of Boulogne-sur-Mer, Calais, and Dunkirk, both theoretically predicted (tidal model) and measured water levels from tide gauges. The present paper is mainly a methodological contribution, where new methods, borrowed from the fields of turbulence and time series analysis, are applied to water level time series. Below in the first section, the data are presented. Then, the main methods which are used here are presented in the next section. Dynamical and intermittent properties of tidal and measurements time series are presented in the following section: The analyses of power spectra are provided, and then return times of both series are considered. The last section is devoted to the analysis of the surge: the probability density of the surge times series, dynamical properties of the intermittency of surge time series, using Hilbert–Huang transform, and finally tide–surge interactions through the probability of surge data conditioned by the tidal water level.

Fig. 1
figure 1

A typical 3 days portion of the sea level in the eastern English Channel (extract from the Boulogne series). a superposition of the theoretical curve and the measured time series. b The difference, measured-theory, showing the non-tidal residual component, reaching here values from \(-1\) to +1.5 m

2 Data sets

The data consist of hourly tide gauge measurements recorded using tidal gauges in the three ports of Boulogne-sur-Mer, Calais, and Dunkirk (Fig. 2). Table 1 provides the dates of beginning and end of the measurements used in the present study. There are missing data, and the last column of the table provides the percentage of present data. Since the predicted series are obtained from a numerical program, they are available for the same period, without missing data. For each series, several hundred thousands of data points are available. For some statistical analyses for which time is not a relevant parameter, such as mean, variance, quantiles, probability laws, the fact that there are missing data is not a problem. For dynamical analyses, such as Fourier spectra and return times, analyses are performed for subsets for which the sampling is regular.

As classically considered, the sea level series X(t) can be represented in general as the sum of three terms (Pugh 2004):

$$\begin{aligned} X(t) = Z_0(t)+T(t)+S(t), \end{aligned}$$
(1)

where \(Z_0\) is the mean sea level, which increases (or in some locations decreases) slowly with time, T is the tidal part, and S is the non-tidal residual, which corresponds mainly to the surge. A previous study has estimated the mean sea level change in the Dunkirk location as 1.8 mm/year, as an averaged trend given by measurements from 1942 to 2006 (Haigh et al. 2009). At the scales we consider here (from 2 to 5 decades being the largest scale), this makes a signal of amplitude from 3.5 to 9 cm, which is weak compared to surge events. Hence, we do not consider this trend here, and we consider that the theoretical, predicted signal corresponds to T(t), while the measurements are X(t). The difference \(X-T = S\) is taken as the non-tidal residual.

Table 1 Information about water level measurement time series
Fig. 2
figure 2

Map showing the measurements locations in the ports of Boulogne-sur-Mer, Calais, and Dunkirk

3 Methods

In this section, different methods which will be used in the paper are presented: Fourier spectral analysis, return times estimation using the exceedance function, and empirical mode decomposition associated with Hilbert spectral analysis.

3.1 Fourier spectral analysis

Fourier spectral analysis is useful to provide the energy associated with each frequency (Bendat and Piersol 2000). More precisely, E(f) is an energy density where f is the frequency, and E(f)df is the part of the variance of the original signal, whose frequency range belongs to the frequency band df. This is classically estimated using the Fast Fourier Transform (FFT) algorithm, requiring consecutive values, with no missing data. The power spectra are represented here in log–log plot. Deterministic forcing is visible by spikes in the spectrum, noise by horizontal parts in the spectrum, and power-law regimes of the form \(E(f)\sim f^{-\beta }\) appear as straight lines of slope \(\beta\).

3.2 Return times estimated using the exceedance function

Return times are dynamical quantities for a stochastic process, characterizing the recurrence of a given state. For natural phenomena, the concept of return times has a special importance when considering extreme events, which can have huge societal impacts, e. g., storms, earthquakes, floods. Return times correspond to the time between the end of an extreme event and the beginning of the next one. Successive return times can be built from a given time series. If the process considered is stochastic, return times are also stochastic. Of frequent interest is the mean return time associated with a given threshold. The larger the threshold, the larger the mean return time associated with this threshold. It is known for discrete and stationary stochastic processes that the mean return time \(\overline{T}\) for a cell A of probability P(A) can be written as (Kac 1947; Nicolis and Nicolis 2007):

$$\begin{aligned} \overline{T}=\frac{\tau }{P(A)}, \end{aligned}$$
(2)

where \(\tau\) is the time step of the process. This result is a theorem, using the hypothesis that the probability of large return time is very small. By considering the events larger than the threshold, one obtains the complementary cumulative distribution function, or exceedance function:

$$\begin{aligned} F(x)=\Pr (X>x) \end{aligned}$$
(3)

where X is the process studied and x the threshold. The mean return is then estimated using F(x). If there are N measurement values, and \(n_0\) occurrences of events larger than the threshold s, then \(p=\Pr (X>x)=F(s)=n_0/N\), and the mean return time is written as:

$$\begin{aligned} \overline{T}(s)=\frac{\tau }{p}=\frac{\tau N}{n_0}. \end{aligned}$$
(4)

For example, for \(N=10^6\) measurement values, and \(n_0=10\) occurrences, for hourly data, one has \(\overline{T}=100,000\) h \(=11.4\) years. This approach is useful, since the return times do not need to be directly estimated; one needs only to consider the exceedance function and to estimate from it a mean return time.

3.3 Empirical mode decomposition and Hilbert spectral analysis

The objective of the empirical mode decomposition (EMD) method is to decompose a time series into a sum of modes, each one being narrow banded in frequency. Then, Hilbert spectral analysis (HSA) is applied on each mode time series, in order to extract locally, for each time step, instantaneous amplitude and frequency information. This approach was introduced in the end of the 1990s (Huang et al. 1998) and has since received a large audience, with several thousand citations for the original article. It has been applied on turbulence time series and generalized to study intermittency (Huang et al. 2008, 2011). A recent reference describes in detail the use of EMD and HSA for the study of scaling intermittency (Schmitt and Huang 2016).

The first step is the EMD algorithm, applied on the original time series. The EMD is similar with Fourier or wavelet analysis, but has a main difference, due to the fact that no basis function is assumed a priori: The method is adaptative, meaning that each mode has a different shape, given by the data. However, the method is not mathematically based and relies on an iterative algorithm. We do not provide here all the details on this algorithm and refer to the original paper for this (Huang et al. 1998), or to the recent book detailing the calculation procedure (Schmitt and Huang 2016). The algorithm extracts p modes \(C_i, i\in [1,p]\), from high to low frequencies. At the end of the decomposition, the original time series X(t) is the sum of the p modes and a residual r(t):

$$\begin{aligned} X(t) = \sum _{i=1}^p C_i(t) +r(t) \end{aligned}$$
(5)

The number p is of the order of \(\log _2 N\), where N is the number of data points of the original series. This method can be used to filter a series (excluding the first modes corresponding to higher frequencies), or to consider the fluctuations at some given scales (by considering the modes corresponding to these scales) (Flandrin and Gonçalvès 2004). This decomposition is followed by HSA, applied to each mode, in order to locally extract an amplitude and a frequency. A Hilbert transform is applied to each mode:

$$\begin{aligned} \hat{X}(t)=\frac{1}{\pi }\int \frac{X(u){\mathrm{d}}u}{t-u}. \end{aligned}$$
(6)

This Hilbert transform is used to build a so-called analytical signal as \(Z(t)=X(t)+j\hat{X}(t)=A(t)e^{j\theta (t)}\), where \(j^2=-1\), A(t) is the local amplitude and \(\theta (t)\) the phase. The derivative of the phase gives the instantaneous frequency \(\omega (t)=\theta '(t)/(2\pi )\). This provides a time–frequency–amplitude methodology providing for each time t p values of the amplitude and the phase. A 2D probability density \(p(\omega ,A)\) is a amplitude–frequency space that can be estimated from the data. This can be used to extract the energy associated with a frequency \(h(\omega )\):

$$\begin{aligned} h(\omega )=\int _0^{\infty }p(\omega ,A)A^2{\mathrm{d}}A \end{aligned}$$
(7)

This provides an approach which is similar to the Fourier spectral energy estimation E(f). The comparison between this EMD–HSA method and the Fourier power spectrum has been done first for turbulence data (Huang et al. 2008). Such approach is attractive by the fact that it can be used even when some data are missing, contrary to the classical algorithm for the Fourier transform (the Fast Fourier Transform (FFT)), which requires equally spaced data. A generalization of this approach has been proposed to study the intermittent properties of time series in a new scaling framework (Huang et al. 2008, 2011; Schmitt and Huang 2016). The idea is to consider moments of order \(q>0\) of the amplitude, and not only the second moment. This writes:

$$\begin{aligned} L_q(\omega )=\int _0^{\infty }p(\omega ,A)A^q{\mathrm {d}}A \end{aligned}$$
(8)

In case of scale invariance of the fluctuations, these q order moments of the fluctuations characterize intermittency through a power law as follows:

$$\begin{aligned} L_q(\omega )\sim \omega ^{-\xi (q)}, \end{aligned}$$
(9)

where \(\xi (q)\) is an intermittency exponent, which characterizes the multiscale statistics of the process considered. Larger moments q correspond to larger fluctuations.

It is related to the classical structure functions scaling exponent \(\zeta (q)\) (Frisch 1995) by \(\xi (q)=\zeta (q)+1\) (Huang et al. 2008, 2011; Schmitt and Huang 2016). This can also be compared to some classical scaling stochastic models. For example for Brownian motion, \(\zeta (q)=q/2\), for fractional Brownian motion, a classical fractional generalization of Brownian motion, \(\zeta (q)=qH\), with \(0<H<1\), where H is the Hurst exponent, which characterizes the scaling of the mean fluctuations. If the moment function is linear, the process is monofractal; the intermittency is larger for multifractal processes, for which the moment function \(\zeta (q)\) or \(\xi (q)\) is nonlinear and concave. This method has been shown to have less influence from periodic signals superposed to scaling stochastic fluctuations, compared to the classical structure functions analysis (Huang et al. 2011; Schmitt and Huang 2016). It is hence well designed for the analysis of the surge time series, which has less influence from tidal deterministic harmonics than the water level signal itself, but where there are still deterministic peaks remaining.

4 Dynamical and intermittent properties of the tidal times series and of the measurements

4.1 Power spectra

We consider here a first dynamical property of the series X(t) and T(t): their Fourier power spectra. Since the FFT algorithm requires consecutive values, we have taken here for the Fourier spectral analysis, recent subsamples of each series, as long as possible, with no missing values. These are indicated in Table 2: There are about 37,000 points for Boulogne-sur-Mer, 44,000 for Calais, and 32,000 for Dunkirk, which is more than enough to perform a Fourier transform and display the spectral density. The corresponding spectra are shown in Fig. 3a–c. Each figure represents superposed spectra for the tidal signal and the measured signal. For the tidal signal, corresponding to theoretical predictions obtained by modeling, the spectra have been estimated on the same portions of time series as for the measured data.

In each case, the superposition of the two spectra is rather limited: Tidal spectra display the different harmonics used for the numerical calculation; these are mainly discrete spectra, with strong peaks corresponding to frequencies used for the reconstruction of the deterministic tidal signal. Measurement spectra are close to tidal spectra only for some peaks at high frequency. For the rest of the frequencies, and many for low frequencies, both spectra are markedly different; for lower frequencies, the measured time series have spectral energy density 3 to 5 orders of magnitude higher than the spectral energy density of the tidal component. Some scaling regimes can be seen for lower frequencies, with values of \(\beta\) given in Table 2. For the Calais series, the value obtained can be compared to the value of 5/3, which corresponds to the Kolmogorov exponent of fully developed turbulence (Frisch 1995; Pope 2000; Schmitt and Huang 2016). For Boulogne-sur-Mer and Dunkirk, the slopes could be compared to the value of \(\beta =1.27\pm 0.02\) found for velocity in the atmosphere (Katul and Chu 1998; Calif and Schmitt 2014). However, there is no theoretical explanation to be proposed here to better explain this precise value.

The spectra of sea level measured time series hence possess peaks corresponding to the deterministic tidal forcing, and a scale invariant part corresponding to a stochastic forcing associated with the meteorology. The pressure fluctuations have been analyzed (not shown here): They have a scaling power spectrum with a slope close to − 7/3, as expected in the field of turbulence, showing that the pressure information in itself cannot alone explain such scaling behavior in the water level dynamics. There could be an influence of turbulence, for scales going from hours to months, in this meteorological forcing, corresponding to unknown nonlinear interactions between wind, pressure, and temperature.

Ideally, for comparison purposes, it would have been better to study the same time periods for the three series, in order to consider the same events. However, the spectral analysis is estimated over periods of several years, which is assumed here to be enough to average out specific meteorological events. It can also be underlined that the effect of longer term variability in sea level data is not captured by such duration time series (Haigh et al. 2011; Mawdsley et al. 2015). However, shorter timescales information is well captured here.

Table 2 Information about measurement time series, subsamples without missing values, used for the spectral analysis
Fig. 3
figure 3

Spectral analysis of measured and tidal series a Boulogne-sur-Mer, b Calais, and c Dunkirk. Straight lines are also represented, corresponding to power-law fits of the low-frequency part of the measured data. The power-law slopes are given in Table 2

4.2 Return times

Here, another dynamical quantity is considered, the return times for both tidal and measured time series.

Figure 4a represents the exceedance probability estimated for all six time series: tidal model and measurement data in Boulogne, Calais, and Dunkirk. The absolute water level values are different for each port, since the tidal amplitude increases from Dunkirk to Boulogne-sur-Mer. For low threshold values, one can see that, for each series, the probability curves for tidal and measurement data are superimposed. The difference between tidal and measurement data is becoming important and visible only for the larger thresholds, as already found recently (Eelsalu et al. 2014). These curves are transformed into mean return times in Fig. 4b. For large thresholds (larger than 9 m for Boulogne, 7.5 m for Calais, and 6 m for Dunkirk), it is visible that for a given threshold, the mean return time is much larger for the model than for measurements. Inversely, for a given return time, for mean return times larger than 0.02 years = 1 week, the threshold provided by the model and by measurements is different. For example, for a mean return time of 10 years, the threshold is about 40 cm larger for the measurements than for the tidal model.

Finally, let us mention that the extrapolation of such analysis for very large return times is often information requested by local or national authorities. As an example, new dikes are sometimes foreseen, corresponding to a threshold associated with a mean return time of 1000 years. For such estimation, measurements are not long enough, and an extrapolation of the curves must be performed, with an adequate model. Let us emphasize here that the curve obtained for experimental measurements must be used, and not the curve obtained from tidal model data.

The difference for a mean return time of 10 years is 50 cm, and an extrapolation shows that for a mean return time of 100 years, the difference is larger than 1 m, whereas for 1000 years it may reach 4 m. In several aspects, this is a classical problem in coastal engineering and management. It may also be addressed using an ensemble approach (Eelsalu et al. 2014).

Fig. 4
figure 4

Left: exceedance probability for the six time series, with each time a comparison between model (tidal data) and measurements (B = Boulogne, C = Calais, D = Dunkirk). Right: transformation of the exceedance probability into a mean return time in years, versus the threshold

4.3 About the pdf of return times

Sometimes engineering consultants or even politics indicate that an event having a mean return time of 100 years has 1/100 probability to happen every year. The general idea is to say that an event having a mean return time of \(\overline{T}\) years has a \(\tau /\overline{T}\) probability to happen every year (where here \(\tau\) = 1 year). This is an ambiguous formulation. It is correct to state that the overall probability of such event is \(\tau /\overline{T}\), but to indicate that “every year” this is the case, corresponds to assume independence of successive events.

The underlying hypothesis (non stated) is that these events are independent Bernoulli process of parameter \(\rho =\tau /\overline{T}\). The probability of not having an event is \(1-\rho =(\overline{T}-\tau )/\overline{T}\), and since each consecutive event is supposed to be independent, the probability to have k successive “non-events” is \((1-\rho )^k=\left( \frac{\overline{T}-\tau }{\overline{T}}\right) ^k\). For \(\overline{T}/\tau =100\) this is \(0.99^k\). The probability density of return times in this case is an exponential of the form:

$$\begin{aligned} p(k)=e^{-\lambda k} \end{aligned}$$
(10)

with \(\lambda =\log \frac{\overline{T}}{\overline{T}-\tau }\). For such process, the mean return time has the correct value, since it is \(1/\rho =\overline{T}/\tau\). But the pdf of return times or real processes is in general not an exponential. For example, independent extremes follow the Generalized Extreme Value Distribution that is indeed very different from the exponential function: see, e.g., Coles (2004) and Wahl et al. (2017).

This has been verified on a long almost continuous portion, from Dunkirk time series: 112,499 data from September 30, 1997 to July 21, 2010. This portion contains only \(0.66 \%\) of missing data. The pdf of return times has been estimated for two thresholds: \(S=5\)m, for which there is a mean return time of \(\overline{T}=11.3\) h, and 7801 return events, and \(S=6.2\)m, for which there is a mean return time of 156.9 h, and 708 different return events. In both cases, a Bernoulli time series of length 5,000,000 data points with parameter \(\rho =\tau /\overline{T}\) is generated (here \(\tau=1\) h), and its pdf of return times is compared to the measured ones. Figure 5 shows the exceedance probability F(x) for measured return times, and for a Bernoulli process with the same mean return time. The exponential decrease of the Bernoulli process is clearly visible in each case, whereas the measured exceedance probability has a complex shape, very different from an exponential.

Hence, it is correct to state that the mean return time of an event having a probability of 1/P is \(\overline{T}=P\), but it is not correct to consider that every year its probability of occurring is 1 / P; furthermore, these successive probabilities are not independent.

Fig. 5
figure 5

Exceedance probability F(x) for measured return times, and for a Bernoulli process with the same mean return time. Left: for the threshold \(S=5\) m. Right: the same figure, for the threshold \(S=6.2\) m

5 Analysis of the non-tidal residual time series

We analyze here non-tidal residual (NTR) time series, taken as \(S(t)=X(t)-T(t)\), the measurements minus the tidal component. Figure 6 shows a portion (from Boulogne-sur-Mer site) of 500 consecutive days of the series S(t), illustrating its stochastic and multiscale aspect. We first analyze its statistics, then its dynamical properties, and finally the tide–surge interactions, in the form of the relation between the water level and the non-tidal residual.

5.1 Probability density function of non-tidal residual time series

Table 3 provides the mean, median, standard deviation, minimum, and maximum value of the NTR series for the three ports. The mean and median are very small for each series (of the order of cm); the standard deviation is very similar in each case (of the order of 20 cm). Maxima and minima are very close for Calais and Dunkirk, and for Boulogne-sur-Mer, they are slightly larger. Since there are several hundred thousand data points, the probability density function (PDF) for each series can be estimated with a rather good precision. This is shown in Fig. 7. Both PDFs are superposed, where \(\log\) of the probability is represented, in order to better visualize the rare events. For positive values, the PDF is very close, and for negative values, the Dunkirk case seems slightly different from the two others. A dotted line is also represented, corresponding to an exponential decrease. The exponential fit is provided by the equation:

$$\begin{aligned} p(x)=\exp (-K \vert x \vert ) \end{aligned}$$
(11)

with the value \(K=6.3\). Some PDF plots of the non-tidal residual can be found in Rabassa and Beck (2015), for the sea levels at various coastal locations in the UK. In this paper, for some plots, an approximate exponential decay can be found, however, with no estimation of the parameter K. An analysis of the variation of the exponent K for both negative and positive surges for time series of sea levels numerically reconstructed for the eastern Baltic sea coasts using an ocean model is provided in Soomere et al. (2015).

Fig. 6
figure 6

A portion (from Boulogne-sur-Mer site) of 500 consecutive days of non-tidal residual series S(t). The extract shown here is from November 13, 1975 to March 27, 1977

Fig. 7
figure 7

Probability density function of the non-tidal residual time series, superposed with an exponential fit of equation \(p(x)=\exp (-K \vert x \vert )\) with \(K=6.3\)

Table 3 Some statistical values of NTR data for each series

5.2 Intermittency study of NTR dynamics

We consider here the intermittent properties of the non-tidal residual time series, using the empirical mode decomposition (EMD) methodology, associated with the Hilbert spectral analysis (HSA).

The analysis has been performed on subsamples of the series, for which there are no missing data, as given in Table 2. Hilbert spectral analysis is applied, indicating scaling behavior of the NTR time series, for the range of frequencies \(0.03 <\omega <0.5\) day\(^{-1}\) (2 days–1 month) (Fig. 8a). Compensated spectra show plateaus for each spectrum, indicating that the scaling is rather well respected (Fig. 8b). The arbitrary order HSA is also applied on the same time series, \(L_q(\omega )\) versus \(\omega\), and on the same range of frequencies, the exponents \(\zeta (q)=\xi (q)-1\) are extracted with their error bars. The curves \(\zeta (q)\) versus the moment q (Fig. 9) are nonlinear in each case, indicating multiscaling behavior of the series. Such property provides a way to generate realistic synthetic time series of surge data. It also characterizes the intermittency of such series, for the scales considered here, from 2 days to 1 month.

Fig. 8
figure 8

Energy spectra using HSA analysis, \(L_q(\omega )\) versus \(\omega\), for all three ports, applied to the subsample of the data, for which there are no missing values, given in Table 2. The scaling range indicated in the figure is \(0.03\,<\omega <\,0.5\) day\(^{-1}\) (2 days–1 month). The compensated spectra are shown in the figure on the right, the plateaus showing that the scaling is rather well respected

Fig. 9
figure 9

Scaling exponents \(\zeta (q)\) versus the moment q, for all three ports, estimated on the range \(0.03<\,\omega\, <\,0.5\) day\(^{-1}\) (2 days–1 month). The error bars are also shown. The nonlinearity indicates multiscaling behavior

5.3 Tide–surge interactions

The relationship between surge and tide has been a topic of interest in many old and recent studies of water level time series (Prandle and Wolf 1978; Horsburgh and Wilson 2007; Bernier and Thompson 2007; Wolf 2009; Idier et al. 2012). The general idea is to consider that surge may have some nonlinear interactions with the tide and that both signals may not be independent. Such result has strong interest in relation to the interpretation and prediction of the surge signal. Some of the previous studies have shown that the surge is larger a few hours before the high tide: For example, it has been shown that in the North Sea on the East Coast of England, the peak surges occurrence is 3–5 h before the nearest high water, and sometimes 1–2 h before the nearest high tide (Horsburgh and Wilson 2007).

Here, we do not consider the surge value versus the tidal phase information, since the latter is variable, with many different tidal amplitudes. We prefer to consider, as done in other works, the relation between NTR and height of tide (Haigh et al. 2011; Mawdsley and Haigh 2016). In this framework, we consider something which has not, to our knowledge, been analyzed before and which is of more direct interest for hazard evaluation: the conditional average of the NTR, with respect to the value predicted by the tidal model). For this, the different tidal level values are partitioned into several classes (25 classes in our analysis, chosen to have enough statistics in each class), and for each class, a conditional mean value of the NTR is estimated. This shows the possible dependency of the NTR value with respect to the tidal level: If there are no dependency, the corresponding curve should be flat. For such analysis, all the available data are considered, as given in Table 1. Figure 10a–c shows the conditional mean NTR for, respectively, Boulogne-sur-Mer, Calais, and Dunkirk. There is clearly a tendency, with large mean NTR values for medium to large values of the tidal level. The same analysis is performed by considering separately positive and negative NTR events. The corresponding curves are shown in the same plots. In each case, the larger mean values of the NTR (negative and positive parts) are obtained for the medium level of the tidal value. The case of Boulogne-sur-Mer is clearer, with mean values of the NTR of 0.2 m for tidal values from 4 to 5 m. The situation in Calais is similar, and in Dunkirk qualitatively less marked. Let us also note that the mean conditional NTR is not the difference between the positive and negative mean conditional curves, since there could be relatively more positive or negative events.

Table 4 presents the total number of positive and negative NTR events, showing that for the Boulogne-sur-Mer case, there are clearly more positive NTR events, whereas for Calais, there are less positive NTR events, and in Dunkirk, there seems to be a relatively even number of positive and negative events. However, these numbers do not necessarily have the same repartition with respect to the tidal level. Indeed, Fig. 10d shows the ratio of positive to negative number of NTR, for each bin used to consider the conditional mean: This shows that there are much more positive NTR events for large tidal values. This is very clear for the Boulogne-sur-Mer case with a rather uniform increase. For the Calais and Dunkirk cases, there is a relative maximum, around 5 m for Calais and 6 m for Dunkirk. This non-uniform number of positive versus negative surge events explains that the difference between conditional positive and negative curves in Fig. 10a–c does not give directly the conditional mean NTR.

Table 4 Information about the number of positive and negative NTR events
Fig. 10
figure 10

Conditional average of the NTR, with respect to the absolute tidal level. Top left: Boulogne-sur-Mer, top-right: Calais, bottom-left: Dunkirk. In each case, the mean NTR conditioned on the tidal total water level is displayed, together with the conditional mean of the positive NTR and the conditional mean of the negative NTR. Bottom-right: the ratio of positive to negative number of NTR, for each bin used to consider the conditional mean. Here, the curves for all ports are superposed

6 Discussion and conclusion

Water level time series are complex, multiscale, with a deterministic tidal component and a stochastic part. The deterministic part is a sum of monochromatic trigonometric functions, which amplitudes result from numerical estimations taking into account the geometry of the oceanic basin. The stochastic part is not included in the model and corresponds to meteorological influences. In this paper, we have compared modeled (tidal) and measured time series, in the ports of Boulogne-sur-Mer, Calais, and Dunkirk, in the northern coast of France, in the eastern English Channel, and South of North Sea.

We have first considered the dynamics of both time series, by estimating their Fourier power spectra. For this, we have considered subsamples of the series, for which there are no missing values in the measurements. In each case, power spectra have shown that the model tidal signal is very far from measurements for the low-frequency part, for times larger than 1 day. Only high-frequency peaks are well captured by the model tidal signal. One can also detect a power-law behavior for the measured data, which is visible for two or more decades lower than 1 day\(^{-1}\), with slopes ranging from 1.27 to 1.71. These values could be linked in some way to a turbulence influence, which could be found through pressure, temperature, and wind fields. It is not due to pressure alone, which has a power law close to − 7/3 at these scales. We have also estimated mean return times, using the exceedance probability. We have shown that for large thresholds, mean return times are very different between the tidal model time series and the measurements.

We have also considered the residual, or the surge, which is the difference between the measurements and the tidal signal. We have estimated its probability density function, which is rather close to an exponential distribution, for negative and positive surge events, for all three ports. Its dynamics have been considered for the same subsample, for which there are no missing data, using the empirical mode decomposition and Hilbert spectral analysis. It was shown to be scaling on a range of scales from 2 days to 1 month, and over this range of scales, the scaling exponents \(\zeta (q)\) have been estimated for all three series.

A quantification of the relation between surge and tide has also been realized. Such relation has been explored in other works, mainly by considering the surge events versus the tidal phase. Here, we considered the conditional mean of surge, of positive surges and of negative surges, with respect to the tidal absolute level. We also explored the number of positive or negative surges, with respect to the tidal level.

These approaches are new methodologies. They have been tested here on three times series from ports in Northern France. A perspective of this work is to extend the analysis to other water level series in the English Channel and the North Sea. These methodologies provide a quantitative expression of the dependency between surge and tide, which can be used for predictions of future surges. Such perspectives of the current work can be explored in two directions. On the one hand, synthetic water level series can be generated, with the generation of stochastic NTR variables having correct statistics and dynamics, which are added to the tidal level prediction model. On the other hand, the shape of the mean return time estimated from such series can be compared to measurements and extended in order to provide information on the thresholds associated with much longer mean return times.

We have not considered the slow elevation of the mean sea level. Several studies have shown that the mean sea level elevation is approximately 1.8 mm per year in Dunkirk (Hequette 2010; Haigh et al. 2011). This is a rather weak signal, compared to the high values reached by the surge data (1 or 2 m). A few millimeter trend will not change the exceedance probability at a scale of few years. However, if sea level continues to rise at similar rates or higher over several decades, the mean return times of extreme water levels will necessarily decrease (see discussions in, e.g.,  Haigh et al. 2011; Mudersbach et al. 2013; Soomere and Pindsoo 2016). In such case, this can be easily incorporated in the present analysis and modeling.

Finally, let us note that the topic of return times is a complex question, especially for time series presenting a strong variability at many different scales. There is a rather strong potential societal impact, via the risk evaluation and political decisions that need to be taken in answer of hazard assessments. Researchers must be careful enough to use well-defined concepts and methods. For a stochastic process, return times are also stochastic. If the last big event was in 1950 and if the estimated mean return time of such event is 70 years, it does not mean that it has a large probability to occur in 2020. Successive return times for a same threshold may have very different magnitudes, and there may be large fluctuations around the mean value. The different results which have been obtained here may provide a methodology to assess the possible universal properties of the surge, and also approaches for stochastic modeling of this complex series for hazard assessment in response to societal needs.