Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

This chapter develops a compact tri-colored rooted-tree theory for the order conditions for general ERKN methods. The bottleneck of the original tri-colored rooted-tree theory is the existence of numerous redundant trees. This chapter first introduces the extended elementary differential mappings. Then, the new compact tri-colored rooted tree theory is established based on a subset of the original tri-colored rooted-tree set. This new theory makes all redundant trees no longer appear, and hence the order conditions of ERKN methods for general multi-frequency and multidimensional second-order oscillatory systems are greatly simplified.

8.1 Introduction

Runge–Kutta–Nyström (RKN) methods (see [12]) are very popular for solving second-order differential equations. This chapter develops the rooted-tree theory and B-series for extended Runge–Kutta–Nyström (ERKN) methods solving general multi-frequency and multi-dimensional oscillatory second-order initial value problems (IVPs) of the form

$$\begin{aligned} \left\{ \begin{aligned}&\varvec{y}''(t)+M \varvec{y}(t)=\varvec{f}\big (\varvec{y}(t),\varvec{y}'(t)\big ), \qquad t\in [t_0,T],\\&\varvec{y}(t_0)=\varvec{y}_0,\quad \varvec{y}'(t_0)=\varvec{y}'_0, \end{aligned}\right. \end{aligned}$$
(8.1)

where M is a \(d\times d\) constant matrix implicitly containing the dominant frequencies of the system, \(\varvec{y}\in \mathbb {R}^d,\) and \(\varvec{f}:\mathbb {R}^d\times \mathbb {R}^d \rightarrow \mathbb {R}^d\), with position \(\varvec{y}\) and velocity \(\varvec{y}'\) as arguments. In the special case where the right-hand side function of (8.1) does not depend on velocity \(\varvec{y}'\), (8.1) reduces to the following special second-order oscillatory system

$$\begin{aligned} \left\{ \begin{aligned}&\varvec{y}''(t)+M \varvec{y}(t)=\varvec{f}\big (\varvec{y}(t)\big ), \qquad t\in [t_0,T],\\&\varvec{y}(t_0)=\varvec{y}_0,\quad \varvec{y}'(t_0)=\varvec{y}'_0. \end{aligned}\right. \end{aligned}$$
(8.2)

Furthermore, if M is symmetric and positive semi-definite and \(\varvec{f}(\varvec{q})=-\nabla \varvec{U}(\varvec{q})\), then, with \(\varvec{q}=\varvec{y}\), \(\varvec{p}=\varvec{y}'\), (8.2) becomes identical to a multi-frequency and multidimensional oscillatory Hamiltonian system

$$\begin{aligned} \left\{ \begin{array}[c]{ll} \varvec{p}'(t)=-\nabla _{\varvec{q}} H(\varvec{p}(t),\varvec{q}(t)),&{}{\varvec{p}}(t_{0})=\varvec{p}_{0}, \\ \varvec{q}'(t)=~~~\nabla _{\varvec{p}} H(\varvec{p}(t),\varvec{q}(t)), \qquad \ &{} \varvec{q}(t_{0})=\varvec{q}_{0}, \end{array}\right. \end{aligned}$$
(8.3)

with the Hamiltonian

$$\begin{aligned} H(\varvec{p},\varvec{q})=\dfrac{1}{2}\varvec{p}^{\intercal }\varvec{p} +\dfrac{1}{2}\varvec{q}^{\intercal }M\varvec{q}+\varvec{U}(\varvec{q}), \end{aligned}$$

where \(\varvec{U}(\varvec{q})\) is a smooth potential function. For solving the multi-frequency, multi-dimensional, oscillatory system (8.3), a large number of studies have been made (see, e.g. [6, 25, 28]). The methods for problems (8.1) and (8.2) are especially important when M has large positive eigenvalues, as in the case where the wave equations is semi-discretised in space (see, e.g. [11, 14, 26, 27, 29]). Such problems arise in a wide range of fields such as astronomy, molecular dynamics, classical mechanics, quantum mechanics, chemistry, biology and engineering.

ERKN methods were proposed originally in the papers [23, 32] to solve the special oscillatory system (8.2). ERKN methods work well in practical numerical simulation, since they are specially designed to be adapted to the structure of the underlying oscillatory system and do not depend on the decomposition of the matrix M. ERKN methods have been widely investigated and used in numerous applications in the fields of science and engineering. For example, the idea of ERKN methods has been extended to two-step hybrid methods (see, e.g. [8, 9]), to Falkner-type methods (see, e.g. [7]), to Störmer–Verlet methods (see, e.g. [17]), to energy-preserving methods (see, e.g. [11, 15, 24]), and to symplectic and multi-symplectic methods (see, e.g. [14, 16, 19, 20]). Meanwhile, further research on ARKN methods, including the symplectic conditions and symmetry, has been carried out in the following papers [10, 13, 21, 22, 31].

In a recent paper [33], ERKN methods were extended to the general oscillatory system (8.1), and a tri-colored tree theory called extended Nyström tree theory (EN-T theory) was analysed for the order conditions. Unfortunately, however, the EN-T theory is not completely satisfactory due to the existence of redundant trees. For example, there are 7 redundant trees out of 16 trees for third order ERKN methods. In practice, in order to gain the order conditions for a specific ERKN method of order r, one needs to draw all the trees of order up to r first, and then from them select and delete about half of the redundant trees. This will lead to inefficiency in the use of the EN-T theory to achieve the order conditions for ERKN methods.

Hence, in this chapter, we will present an improved theory to eliminate all such redundant trees. In a similar approach to the case of the special oscillatory system (8.2) in [30], extended elementary differentials are required, and we will discuss this in detail in Sect. 8.4.

This chapter is organized as follows. We first summarise the ERKN method for the general oscillatory system (8.1) in Sect. 8.2, and then in Sect. 8.3 we illustrate drawbacks of the EN-T theory proposed in [33]. In Sect. 8.4, we introduce the set of improved extended-Nyström trees and show how this relates to other tree sets in the literature. Section  8.5 focuses on the B-series associated with the ERKN method for the general oscillatory system (8.1), and Sect. 8.6 analyses the corresponding order conditions for the ERKN methods, when applied to the general oscillatory system (8.1). In Sect. 8.7 we derive some ERKN methods of order up to four, exploiting the advantages of the new tree theory. The numerical experiments are made in Sect. 8.8. Conclusive remarks are included in Sect. 8.9.

8.2 General ERKN Methods

To begin with, we summarise the following general ERKN method based on the matrix-variation-of-constants formula (see [23]) and quadrature formulae.

Definition 8.1

(See [33]) An s-stage general extended Runge–Kutta–Nyström (ERKN) method for the numerical integration of the IVP (8.1) is defined by the following scheme

$$\begin{aligned} \left\{ \begin{aligned}&Y_i=\phi _0(c_i^2V)\varvec{y}_n+c_i\phi _1(c_i^2V)h\varvec{y}'_n +h^2\sum \limits _{j=1}^{s}\bar{a}_{ij}(V)\varvec{f}(Y_j,Y'_j), \qquad i=1,\ldots ,s, \\&hY'_{i}=-c_iV\phi _1(c_i^2V)\varvec{y}_n+\phi _0(c_i^2V)h\varvec{y}'_n +h^2\sum \limits _{j=1}^{s}a_{ij}(V)\varvec{f}(Y_j,Y'_j), \qquad i=1,\ldots ,s, \\&\varvec{y}_{n+1}=\phi _0(V)\varvec{y}_n+\phi _1(V)h\varvec{y}'_n +h^2\sum \limits _{i=1}^s\bar{b}_i(V)\varvec{f}(Y_i,Y'_i),\\&h\varvec{y}'_{n+1}=-V\phi _1(V)\varvec{y}_n+\phi _0(V)h\varvec{y}'_n +h^2\sum \limits _{i=1}^sb_i(V)\varvec{f}(Y_i,Y'_i), \end{aligned}\right. \end{aligned}$$
(8.4)

where \(\phi _0(V), \phi _1(V),\) \( \bar{a}_{ij}(V)\), \(a_{ij}(V)\), \(\bar{b}_{i}(V)\) and \(b_i(V)\) for \(i,j=1,\ldots ,s\) are matrix-valued functions of \(V=h^2M\), and are assumed to have the following series expansions

$$\begin{aligned}\begin{aligned}&\bar{a}_{ij}(V)=\sum \limits _{k=0}^{+\infty }\frac{\bar{a}^{(2k)}_{ij}}{(2k)!}V^k,\, a_{ij}(V)=\sum \limits _{k=0}^{+\infty }\frac{a^{(2k)}_{ij}}{(2k)!}V^k,\; \\&\bar{b}_{i}(V)=\sum \limits _{k=0}^{+\infty }\frac{\bar{b}^{(2k)}_{i}}{(2k)!}V^k,\, b_i(V)=\sum \limits _{k=0}^{+\infty }\frac{b^{(2k)}_i}{(2k)!}V^k,\, \phi _i(V)=\sum \limits _{k=0}^{+\infty }\frac{(-1)^k}{(2k+i)!}V^{k} \end{aligned} \end{aligned}$$

with real coefficients \(\bar{a}^{(2k)}_{ij}\), \(a^{(2k)}_{ij}\), \(\bar{b}^{(2k)}_{i}\), \(b^{(2k)}_i\) for \(k=0,1,2,\ldots \).

The ERKN method (8.4) in Definitions 8.1 can also be represented compactly in a Butcher tableau of the coefficients [4]:

$$\begin{aligned} \begin{array}{c|cccc|cccc} c_1 &{} \bar{a}_{11}(V) &{}\bar{a}_{12}(V) &{}\cdots &{} \bar{a}_{1s}(V) &{} a_{11}(V) &{}a_{12}(V) &{}\cdots &{} a_{1s}(V) \\ c_2 &{} \bar{a}_{21}(V) &{}\bar{a}_{22}(V) &{}\cdots &{} \bar{a}_{2s}(V) &{} a_{21}(V) &{}a_{22}(V) &{}\cdots &{} a_{2s}(V) \\ \vdots &{} \vdots &{}\vdots &{} \ddots &{}\vdots &{} \vdots &{} \vdots &{} \ddots &{} \vdots \\ c_s &{} \bar{a}_{s1}(V) &{} \bar{a}_{s2}(V) &{} \cdots &{} \bar{a}_{ss}(V) &{} a_{s1}(V) &{} a_{s2}(V) &{} \cdots &{} a_{ss}(V) \\ \hline &{} \bar{b}_1(V) &{} \bar{b}_2(V) &{} \cdots &{} \bar{b}_s(V) &{} b_1(V) &{} b_2(V) &{} \cdots &{} b_s(V) \end{array}. \end{aligned}$$
(8.5)

In essence, ERKN methods incorporate the particular structure of the oscillatory system (8.1) into both the internal stages and the updates. Throughout this chapter, we call methods for the general oscillatory system (8.1) general ERKN methods , and standard ERKN methods , for the special case (8.2).

8.3 The Failure and the Reduction of the EN-T Theory

The EN-T theory for general ERKN methods was presented in the recent paper [33] in which some tri-colored trees are supplemented to the classical Nyström trees (N-Ts). The idea of the EN-T theory comes from the fact that the numbers of the N-Ts and of the elementary differentials are completely different. The paper [33] tried to eliminate the difference and then to make one elementary differential correspond to one tree uniquely. Unfortunately, however, the paper [33] did not succeed on this point. For example, the two different trees shown in Table 8.1 have the same elementary differentials \(\mathscr {F}(\tau )(\varvec{y},\varvec{y}')\).

Table 8.1 Two EN-Ts which have the same elementary differentials \(\mathscr {F}(\tau )(\varvec{y},\varvec{y}')\)

Moreover, the great limitation of the EN-T theory is the existence of great numbers of redundant trees that cause trouble in applications. For example, in Table 8.2 (left), there are seven EN-Ts but five of them are redundant since their order \(\rho (\tau )\), density \(\gamma (\tau )\), weight \(\varPhi _i(\tau )\), and the consequent order conditions can be implied by others for the general ERKN methods (8.4).

Table 8.2 Some EN-Ts and the redundance

Here, it should be pointed out that it is not necessary that one tree corresponds to one elementary differential. In other words, one tree may correspond to a set of elementary differentials. For example, just as shown in Table 8.2, the sum of the products of the coefficient \(\alpha (\tau )\) and the elementary differentials \(\mathscr {F}(\tau )(\varvec{y},\varvec{y}')\) is meaningful. In fact, we have

$$\begin{aligned} \varvec{f}^{(1)}_{\varvec{y}}\varvec{y}'+\varvec{f}^{(1)}_{\varvec{y}'}(-My)=D^1_h \varvec{f}\Big (\phi _0(h^2M)\varvec{y}+\phi _1(h^2M)h\varvec{y}',{\phi _0(h^2M)\varvec{y}'-hM\phi _1(h^2M)\varvec{y}}\Big ), \end{aligned}$$

namely, \(\varvec{f}^{(1)}_{\varvec{y}}\varvec{y}'+\varvec{f}^{(1)}_{\varvec{y}'}(-My)\) is the first-order derivative of function \(\varvec{f}\) with respect to h, at \(h=0\), where the function \(\varvec{f}\) is evaluated at point \((\varvec{\hat{y}},\varvec{\hat{y}}')\) with

$$\begin{aligned} \varvec{\hat{y}}= & {} \phi _0(h^2M)\varvec{y}+\phi _1(h^2M)h\varvec{y}', \end{aligned}$$
(8.6)
$$\begin{aligned} \varvec{\hat{y}}'= & {} {\phi _0(h^2M)\varvec{y}'-hM\phi _1(h^2M)\varvec{y}}. \end{aligned}$$
(8.7)

Thus, in Table 8.2, we can choose these two bi-colored trees to respectively represent the sums, and omit all trees with meagre vertices. In this way, we can get rid of the redundance as shown in Table 8.2 (right).

On the other hand, although almost all tri-colored trees are redundant, there indeed exist tri-colored trees which are absolutely necessary in the research of order conditions for the general ERKN methods (8.4). For example, the fifth tree which is tri-colored in the fifth line in the Table 2 in [33] undoubtedly works for the order conditions. In a word, the theory for the general ERKN methods (8.4) is a tri-colored tree theory, but it is based on a subset of the EN-T set.

Hence, it is quite natural that this chapter starts from the Nth derivative of the function \(\left. \varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n}\right| _{(\varvec{\hat{y}},\varvec{\hat{y}}')}\)with respect to h, at \(h=0\). For details about multivariate Taylor series expansions and some related knowledge, readers are referred to [1, 30]. In what follows we will denote this derivative as \(D^N_h\varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n}\).

Remark 8.1

The dimension of the matrix \(D^N_h\varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n}\) is \( d\times d^{m+n}\) . If \(\varvec{z}\) is a \( d^{m+n}\times 1\) matrix, the dimension of \(D^N_h\varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n}\varvec{z}\) is \( d\times 1\).

Remark 8.2

If the matrix M is null,

$$\begin{aligned} D_h^{N}\varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n}\varvec{z}= \varvec{f}^{(m+n+N)}_{\varvec{y}^{m+N}\varvec{y}'^n} \Big (\underbrace{\varvec{y}',\cdots ,\varvec{y}'}_{N \,\text {fold}},\varvec{z}\Big ), \end{aligned}$$

where \(\varvec{f}^{(m+n+N)}_{\varvec{y}^{m+N}\varvec{y}'^n}\) is evaluated at the point \((\varvec{y}, \varvec{y}')\), and \((\cdot ,\cdots ,\cdot )\) is the Kronecker inner product (see [30]).

Remark 8.3

In the special case (8.2) where the function \(\varvec{f}\) is independent of \(\varvec{y}'\), \(D^N_h\varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n}\varvec{z}\) is exactly \(D^N_h\varvec{f}^{(m)}\varvec{z}\) in [30].

At the end of this section we give the following first three results of \(D_h^{N}\varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n}\varvec{z}\), which contribute significantly to our understanding of the extended elementary differentials (see Definition 8.3 in Sect. 8.4).

$$\begin{aligned} D_h^{1}\varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n}\varvec{z}= & {} \varvec{f}^{(m+n+1)}_{\varvec{y}^{m+1}\varvec{y}'^n}\Big (\varvec{y}',\varvec{z}\Big )+ \varvec{f}^{(m+n+1)}_{\varvec{y}^{m}\varvec{y}'^{n+1}}\Big (-M\varvec{y},\varvec{z}\Big ),\\ D_h^{2}\varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n}\varvec{z}= & {} \varvec{f}^{(m+n+2)}_{\varvec{y}^{m+2}\varvec{y}'^n}\Big (\varvec{y}',\varvec{y}',\varvec{z}\Big ) +\varvec{f}^{(m+n+1)}_{\varvec{y}^{m+1}\varvec{y}'^{n}}\Big (-M\varvec{y},\varvec{z}\Big ) +2\varvec{f}^{(m+n+2)}_{\varvec{y}^{m+1}\varvec{y}'^{n+1}}\Big (\varvec{y}',-M\varvec{y},\varvec{z}\Big )\\+ & {} \varvec{f}^{(m+n+2)}_{\varvec{y}^{m}\varvec{y}'^{n+2}}\Big (-M\varvec{y},-M\varvec{y},\varvec{z}\Big )+\varvec{f}^{(m+n+1)}_{\varvec{y}^{m}\varvec{y}'^{n+1}}\Big (-M\varvec{y}',\varvec{z}\Big ),\\ D_h^{3}\varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n}\varvec{z}= & {} \varvec{f}^{(m+n+3)}_{\varvec{y}^{m+3}\varvec{y}'^n}\Big (\varvec{y}',\varvec{y}',\varvec{y}',\varvec{z}\Big ) +3\varvec{f}^{(m+n+3)}_{\varvec{y}^{m+2}\varvec{y}'^{n+1}}\Big (\varvec{y}',\varvec{y}',-M\varvec{y},\varvec{z}\Big )\\+ & {} 3\varvec{f}^{(m+n+3)}_{\varvec{y}^{m+1}\varvec{y}'^{n+2}}\Big (\varvec{y}',-M\varvec{y},-M\varvec{y},\varvec{z}\Big )+\varvec{f}^{(m+n+3)}_{\varvec{y}^{m}\varvec{y}'^{n+3}}\Big (-M\varvec{y},-M\varvec{y},-M\varvec{y},\varvec{z}\Big )\\+ & {} 3\varvec{f}^{(m+n+2)}_{\varvec{y}^{m+2}\varvec{y}'^{n}}\Big (\varvec{y}',-M\varvec{y},\varvec{z}\Big )+3\varvec{f}^{(m+n+2)}_{\varvec{y}^{m+1}\varvec{y}'^{n+1}}\Big (-M\varvec{y},-M\varvec{y},\varvec{z}\Big )\\+ & {} 3\varvec{f}^{(m+n+2)}_{\varvec{y}^{m+1}\varvec{y}'^{n+1}}\Big (\varvec{y}',-M\varvec{y}',\varvec{z}\Big ) +3\varvec{f}^{(m+n+2)}_{\varvec{y}^{m}\varvec{y}'^{n+2}}\Big (-M\varvec{y},-M\varvec{y}',\varvec{z}\Big )\\+ & {} \varvec{f}^{(m+n+1)}_{\varvec{y}^{m+1}\varvec{y}'^{n}}\Big (-M\varvec{y}',\varvec{z}\Big ) +\varvec{f}^{(m+n+1)}_{\varvec{y}^{m}\varvec{y}'^{n+1}}\Big ((-M)^2\varvec{y},\varvec{z}\Big ). \end{aligned}$$

8.4 The Set of Improved Extended-Nyström Trees

In the study of order conditions for second-order differential equations, there are four theory systems listed in Table 8.3, where the abbreviation “SSEN-T” is for simplified special extended Nyström-tree [30], and here the word “compact” should be interpreted as meaning that any order condition derived from a tree belonging the underlining rooted tree set cannot be obtained by another from the same rooted tree set.

Table 8.3 Four theory systems for second order differential equations

The first two systems are very famous in the numerical analysis for ODEs, where the second is a special case of the first one. The rooted tree sets in these two systems are all bi-colored tree sets with the white vertex and the black vertex. The last two systems are constructed on tri-colored rooted tree sets by adding the meagre vertex to the graph of bi-colored trees. Similarly, the last system is the special case of the third.

Moreover, when the matrix M is null, the third is identical to the first, and the fourth to the second. In a word, the last two systems are the extensions of the first two systems respectively. However, the extension of the first system is not satisfied yet, since the last section in this chapter states that the third system is not compact. In order to make the extension better, a compact theory will be built to replace the third one, by introducing a completely new tri-colored rooted tree set and six mappings onto it. In this section, we will define the new tree set and study the relationships to the N-T set, the EN-T set, and the SSEN-T set.

8.4.1 The IEN-T Set and the Related Mappings

In what follows, we will recursively define a new set named the improved extended-Nyström tree set, and define six mappings on it.

Definition 8.2

The improved extended-Nyström tree (IEN-T) set, is recursively defined as follows:

  1. (a)

    , belong to the IEN-T set.

  2. (b)

    If \(\tau \) belongs to the IEN-T set, then the graph obtained by grafting the root of tree \(\tau \) to a new black fat node and then to a new meagre node, \(\cdots \) (p times), and then to a new black fat node and then last to a new white node, denoted by \(W_+B_+(b_+B_+)^{p}(\tau )\) (see Table 8.4), belongs to the IEN-T set for \(\forall p=0,1,2,{\ldots }\).

  3. (c)

    If \(\tau \) belongs to the IEN-T set, then the graph obtained by grafting the root of tree \(\tau \) to a new black fat node and then to a new meagre node, \(\cdots \) (p times), then last to a new white node, denoted by \(W_+(b_+B_+)^{p}(\tau )\) (see Table 8.4), belongs to the IEN-T set for \(\forall p=0,1,2,{\ldots }\).

  4. (d)

    If \(\tau _1,\,\ldots ,\, \tau _\mu \) belong to the IEN-T set, then \(\tau _1\times \cdots \times \tau _\mu \) belongs to the IEN-T set, where ‘\(\times \)’ is the merging product [4].

Table 8.4 Tree \(W_+B_+(b_+B_+)^{p}(\tau )\) (left), and tree \(W_+(b_+B_+)^{p}(\tau )\) (right) in Definition 8.2

Each tree \(\tau \) in the IEN-T set can be denoted by

$$\begin{aligned} \begin{array}{cc} \tau :=&{}\underbrace{\tau _*\times \cdots \times \tau _*}_{N-\text {fold}}\times \Big ( W_+B_+(b_+B_+)^{p_1}(\tau _1)\Big )\times \cdots \times \Big ( W_+B_+(b_+B_+)^{p_m}(\tau _m)\Big )\\ &{}\times \Big (W_+(b_+B_+)^{q_1}(\tau _{m+1})\Big )\times \cdots \times \Big ( W_+(b_+B_+)^{q_n}(\tau _{m+n})\Big ), \end{array}\end{aligned}$$
(8.8)

where \(\tau _*=\)

figure n

. Figure 8.1 gives the mode of the trees in the IEN-T set.

Fig. 8.1
figure 1

The mode of the trees in the IEN-T set

On the basis of Definition 8.2, the following rules for forming a tree \(\tau \) in the IEN-T set can be obtained straightforwardly:

  1. (i)

    The root of a tree is always a fat white vertex.

  2. (ii)

    A white vertex has fat black children, or white children, or meagre children.

  3. (iii)

    A fat black vertex has at most one child which can be white or meagre.

  4. (iv)

    A meagre vertex must has one fat black vertex as its child, and must have a white vertex as its descendant.

Definition 8.3

The order \(\rho (\tau )\), the extended elementary differential \(\mathscr {F}(\tau )(\varvec{y},\varvec{y}')\), the coefficient \(\alpha (\tau )\), the weight \(\varPhi _i(\tau )\), the density \(\gamma (\tau )\) and the sign \(S(\tau )\) on the IEN-T set are recursively defined as follows.

  1. 1.

    \(\rho (\)\()=1\), \(\mathscr {F}(\)\()=\varvec{f}\), \(\alpha (\)\()=1\), \(\varPhi _i(\)\()=1\), \(\gamma (\)\()=1\) and S( \()=1\).

  2. 2.

    For \(\tau \in \) IEN-T denoted by (8.8),

    • \(\rho (\tau )=1+N+\sum \limits _{i=1}^{m}\Big (1+2p_i+\rho (\tau _i)\Big ) +\sum \limits _{i=1}^{n}\Big (2q_i+\rho (\tau _{m+i})\Big )\),

    • \(\mathscr {F}(\tau )= D^N_h\varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n} \Big ((-M)^{p_1}\mathscr {F}(\tau _1),\cdots ,(-M)^{p_{m+n}}\mathscr {F}(\tau _{m+n}) \Big )\),

      where \(p_{m+i}=q_i\), \(i=1,\cdots ,n\), and \((\cdot ,\cdots ,\cdot )\) is the Kronecker inner product (see [30]),

    • \(\alpha (\tau )=(\rho (\tau )-1)!\cdot \frac{1}{N!}\cdot \prod \limits _{i=1}^{m}\Big (\frac{\alpha (\tau _i)}{(1+2p_i+\rho (\tau _i))!}\Big ) \cdot \prod \limits _{i=1}^{n}\Big (\frac{\alpha (\tau _{m+i})}{(2q_i+\rho (\tau _{m+i}))!}\Big ) \cdot \frac{1}{J_1!\ldots J_I!},\)

      where \(J_1,\cdots ,J_I\) count the same branches,

    • \(\varPhi _i(\tau )=c_i^N \cdot \prod \limits _{k=1}^{m}\Big (\sum \limits _{j=1}^{s}\bar{a}_{ij}^{(2p_k)}\varPhi _j(\tau _k)\Big ) \cdot \prod \limits _{k=1}^{n}\Big (\sum \limits _{j=1}^{s}a_{ij}^{(2q_k)}\varPhi _j(\tau _{m+k})\Big )\),

    • \(\gamma (\tau )=\rho (\tau )\cdot \prod \limits _{i=1}^{m} \Big (\frac{(1+2p_i+\rho (\tau _i))!\gamma (\tau _i)}{(2p_i)!\rho (\tau _i)!}\Big )\cdot \prod \limits _{i=1}^{n} \Big (\frac{(2q_i+\rho (\tau _{m+i}))!\gamma (\tau _{m+i})}{(2q_i)!\rho (\tau _{m+i})!}\Big )\),

    • \(S(\tau )=\prod \limits _{i=1}^{m}\Big ((-1)^{p_i}S(\tau _i)\Big )\cdot \prod \limits _{i=1}^{n}\Big ((-1)^{q_i}S(\tau _{m+i})\Big ),\)

where \(\sum \limits _{k=1}^{0}=0\) and \(\prod \limits _{k=1}^{0}=1\).

Definition 8.4

The set IEN-T\(_{m}\) is defined as

$$ \text {IEN-T}_{m}=\big \{\tau :\;\rho (\tau )=m,\,\tau \in \text {IEN-T}\big \}. $$

Remark 8.4

The order \(\rho (\tau )\) is the number of the tree \(\tau \)’s vertices.

Remark 8.5

The extended elementary differential \(\mathscr {F}(\tau )\) is a product of \((-M)^p\) (p is the number of meagre vertices between a white vertex and the next coming white vertex), and \(D^N_h\varvec{f}^{(n+m)}_{\varvec{y}^m\varvec{y}'^n}\) (N is the number of end vertices from the white vertex, m is the number of the non-ending black vertices from the white vertex, and n is the number of the meagre vertices from the white vertex). We will see that the extended elementary differential is not only one function but a weighted sum of the traditional elementary differential.

Remark 8.6

One IEN-T corresponds to one extended elementary differential \(\mathscr {F}(\tau )\).

Remark 8.7

The coefficient \(\alpha (\tau )\) is the number of possible different monotonic labelings of \(\tau \).

Remark 8.8

The weight \(\varPhi _i(\tau )\) is a sum over the indices of all white vertices and of all end vertices. The general term of the sum is a product of \(\bar{a}_{ij}^{(2p)}\) for \(W_+B_+(b_+B_+)^{p}(\tau )\), of \(a_{ij}^{(2p)}\) for \(W_+(b_+B_+)^{p}(\tau )\) (p is the number of the meagre vertices between the white vertices i and j), and of \(c_i^{m}\) (m is the number of end vertices from the white vertex i).

Remark 8.9

One IEN-T corresponds to one weight \(\varPhi _i(\tau )\) .

Remark 8.10

The density \(\gamma (\tau )\) is the product of the density of a tree by overlooking the differences between vertices, and \(\frac{1}{(2p)!}\), where p is the number of the meagre vertices between two white vertices.

Remark 8.11

The sign \(S(\tau )\) is 1 if the number of the meagre vertices is even, and \(-1\) if the number of the meagre vertices is odd.

Table 8.5 makes a list of the corresponding mappings: the order \(\rho \), the sign S, the density \(\gamma \), the weight \(\varPhi _i\), the symmetry \(\alpha \) and the extended elementary differential \(\mathscr {F}\) for each \(\tau \) in the IEN-T set of order up to 4.

Table 8.5 IEN-Ts and mappings of order up to 4 and the corresponding elementary differentials on the N-T set

8.4.2 The IEN-T Set and the N-T Set

In this subsection, we will see that with the disappearance of meagre vertices the IEN-T set is exactly the N-T set. In fact, in this case, each tree \(\tau \) in the IEN-T set has the form shown in Fig 8.2, and the rules to form the tree set are straightforwardly reduced to:

  1. (i)

    The root of a tree is always a fat white vertex.

  2. (ii)

    A white vertex has fat black children, or white children.

  3. (iii)

    A fat black vertex has at most one child which must be white.

In this case, from Remarks 8.48.10, the order \(\rho (\tau )\), the coefficient \(\alpha (\tau )\) and the density \(\gamma (\tau )\) are exactly the same as the ones on the N-T set respectively. If M is null, the weight \(\varPhi _i(\tau )\) and the extended elementary differential \(\mathscr {F}(\tau )(\varvec{y},\varvec{y}')\) on the IEN-T set are exactly the same as the ones on the N-T set respectively, too. In fact, from Definition 8.3, with the disappearance of meagre vertices, these two mappings are recursively defined respectively, for \(\tau \) denoted by Fig. 8.2, as follows:

$$\begin{aligned}\begin{aligned} \varPhi _i(\tau )&=c_i^N \cdot \prod \limits _{k=1}^{m}\Big (\sum \limits _{j=1}^{s}\bar{a}_{ij}\varPhi _j(\tau _k)\Big ) \cdot \prod \limits _{k=1}^{n}\Big (\sum \limits _{j=1}^{s}a_{ij}\varPhi _j(\tau _{m+k})\Big ), \\ \mathscr {F}(\tau )&=D^N_h\varvec{f}^{(m+n)}_{\varvec{y}^m\varvec{y}'^n}\Big (\mathscr {F}(\tau _1),\cdots ,\mathscr {F}(\tau _{m+n}) \Big ). \end{aligned} \end{aligned}$$
Fig. 8.2
figure 2

The form of the trees with meagre vertices disappearing

Clearly, the IEN-T set is really an extension of the N-T set (see, Table 14.3 on p. 292 in [4]). It can also be seen from Tables 8.5 and 8.6 that one 4th order tree, six 5th order trees are appended to the N-T set to form the IEN-T set. All these special and new appended trees have a meagre vertex (or some vertices) which correspond to nothing in the N-T set. In fact, the weights \(\varPhi _i\) in Table 8.6 are all the functions of \(\bar{a}_{ij}^{(2k)}\) and \(a_{ij}^{(2k)}\), high-order derivatives of \(\bar{a}_{ij}(V)\) and \(a_{ij}(V)\) with respect to h, at \(h=0\).

Table 8.6 Tri-colored Trees which are appended to the set N-T\(_{5}\) to form the set IEN-T\(_{5}\)

8.4.3 The IEN-T Set and the EN-T Set

First of all, we note that there are just five mappings defined on the EN-T set in the paper [33], while there are six mappings on the IEN-T set in this chapter. In the paper [33], the authors introduced the signed density \(\tilde{\gamma }(\tau )\), but in this chapter we replace \(\tilde{\gamma }(\tau )\) by the product of the two mappings, the density \(\gamma (\tau )\) and the sign \(S(\tau )\).

The IEN-T set is a subset of the EN-T set, once one overlooks the (extended) elementary differential \(\mathscr {F}(\tau )\) on them.

8.4.4 The IEN-T Set and the SSEN-T Set

From the rules of the IEN-T set and of the SSEN-T set (see [30]), if the function \(\varvec{f}\) in the system (8.1) does not containing \(\varvec{y}'\) explicitly, the IEN-T set is exactly the SSEN-T set.

8.5 B-Series for the General ERKN Method

In Sect. 8.4 we presented the IEN-T set, on which six mappings are defined. With these preliminaries, motivated by the concept of B-series, we will describe a totally different approach from the one described in [33] to deriving the theory of order conditions for the general ERKN method.

The main results of the theory of B-series have their origins in the profound paper [2] of Butcher in 1972, and then are introduced in detail by Hairer and Wanner [5] in 1974. In what follows, we present the following two elementary theorems.

Theorem 8.1

With Definition 8.3, \(\varvec{f}(\varvec{y}(t+h),\varvec{y}'(t+h))\) is a B-series

$$ \varvec{f}(\varvec{y}(t+h),\varvec{y}'(t+h))=\sum \limits _{\tau \in \text {IEN-T}} \frac{h^{\rho (\tau )-1}}{(\rho (\tau )-1)!}\alpha (\tau )\mathscr {F}(\tau )(\varvec{y},\varvec{y}'). $$

Proof

First, we expand \(\varvec{f}(\varvec{y}(t+h),\varvec{y}'(t+h))\) at point \((\varvec{\hat{y}},\varvec{\hat{y}}')\), with the definitions of (8.6) and (8.7).

$$\begin{aligned} \varvec{f}(\varvec{y}(t+h),\varvec{y}'(t+h)) =\sum \limits _{m\ge 0,n\ge 0}\frac{1}{(m+n)!}\left. \varvec{f}_{\varvec{y}^m\varvec{y}'^n}^{(m+n)}\right| _{(\varvec{\hat{y}},\varvec{\hat{y}'})} \Big (\varvec{y}(t+h)-\varvec{\hat{y}}\Big )^{\otimes m}\otimes \Big (\varvec{y}'(t+h)-\varvec{\hat{y}}'\Big )^{\otimes n}, \end{aligned}$$
(8.9)

where the second term \(\left. \varvec{f}_{\varvec{y}^m\varvec{y}'^n}^{(m+n)}\right| _{(\varvec{\hat{y}},\varvec{\hat{y}'})}\) in this series is the matrix-valued function of h.

Definition 8.3 ensures that \(\varvec{f}(\varvec{y}(t+h),\varvec{y}'(t+h))\) is a B-series. In fact, if \(\varvec{f}(\varvec{y}(t+h),\varvec{y}'(t+h))\) is a B-series, from the matrix-variation-of-constants formula with \(\mu =1\), (see [33]), and from the properties of the \(\phi \)-functions (see e.g. [25]), we have

$$\begin{aligned} \begin{array}{ll} \varvec{y}(t+h)-\varvec{\hat{y}} &{} =h^2\int _0^1(1-z)\phi _1((1-z)^2V)\varvec{f}(\varvec{y}(t+hz),\varvec{y}'(t+hz))\mathrm {d}z \\ &{}=\sum \limits _{\tau \in {{\mathrm{IEN-T}}}}\int _0^1(1-z)\phi _1((1-z)^2V)\frac{z^{\rho (\tau )-1}}{(\rho (\tau )-1)!} \mathrm {d}z\cdot \Big (h^{\rho (\tau )+1}\alpha (\tau )\mathscr {F}(\tau )(\varvec{y},\varvec{y}')\Big )\\ {} &{}=\sum \limits _{\tau \in {{\mathrm{IEN-T}}}}\phi _{\rho (\tau )+1}(V)\cdot h^{\rho (\tau )+1}\alpha (\tau )\mathscr {F}(\tau )(\varvec{y},\varvec{y}') \\ &{}=\sum \limits _{\tau \in {{\mathrm{IEN-T}}}} \sum \limits _{p\ge 0}\frac{(-1)^p V^{p}}{(\rho (\tau )+1+2p)!} h^{\rho (\tau )+1}\alpha (\tau )\mathscr {F}(\tau )(\varvec{y},\varvec{y}'), \end{array} \end{aligned}$$
(8.10)

and

$$\begin{aligned} \begin{array}{ll} \varvec{y}'(t+h)-\varvec{\hat{y}}'&=\sum \limits _{\tau \in {{\mathrm{IEN-T}}}} \sum \limits _{q\ge 0}\frac{(-1)^q V^{q}}{(\rho (\tau )+2q)!} h^{\rho (\tau )}\alpha (\tau )\mathscr {F}(\tau )(\varvec{y},\varvec{y}'). \end{array} \end{aligned}$$
(8.11)

Taking the Taylor series of \(\left. \varvec{f}_{\varvec{y}^m\varvec{y}'^n}^{(m+n)}\right| _{(\varvec{\hat{y}},\varvec{\hat{y}'})}\) at \(h=0\), and from (8.10) and (8.11), the Eq. (8.9) becomes

$$\begin{aligned} \begin{array}{rl} &{}\varvec{f}(\varvec{y}(t+h),\varvec{y}'(t+h))=\sum \limits _{N,n,m}\sum \limits _{\tau \in {\mathrm{IEN-T}}}\frac{h^s}{N!(m+n)!} D_h^{N}\varvec{f}_{\varvec{y}^m\varvec{y}'^n}^{(n+m)}\Big ( \frac{(-M)^{p_1}\alpha (\tau _1)\mathscr {F}(\tau _1)(\varvec{y})}{(\rho (\tau _1)+1+2p_1)!},\ldots ,\\ \\ &{} \frac{(-M)^{p_m}\alpha (\tau _m)\mathscr {F}(\tau _m)(\varvec{y})}{(\rho (\tau _m)+1+2p_m)!}, \frac{(-M)^{q_1}\alpha (\tau _{m+1})\mathscr {F}(\tau _{m+1})(\varvec{y})}{(\rho (\tau _{m+1})+2q_1)!},\ldots , \frac{(-M)^{q_n}\alpha (\tau _{m+n})\mathscr {F}(\tau _{m+n})(\varvec{y})}{(\rho (\tau _{m+n})+2q_n)!}\Big ), \end{array} \end{aligned}$$
(8.12)

where

$$\begin{aligned} s=N+\sum \limits _{k=1}^m(2p_k+\rho (\tau _k)+1)+\sum \limits _{k=1}^n(2q_k+\rho (\tau _{m+k})). \end{aligned}$$

By Definition 8.3, the proof is complete.

Theorem 8.2

Given a general ERKN method (8.4), by Definition 8.3, each \(\varvec{f}(Y_i,Y'_i)\) is a series of the form

$$ \varvec{f}(Y_i,Y'_i)=\sum \limits _{\tau \in \text {IEN-T}}\frac{h^{\rho (\tau )-1}}{\rho (\tau )!}\varvec{a}_i(\tau ), $$

where \(\varvec{a}_i(\tau )=\varPhi _i(\tau )\cdot \gamma (\tau )\cdot S(\tau )\cdot \alpha (\tau )\cdot \mathscr {F}(\tau )(\varvec{y}_n,\varvec{y}'_n).\)

Proof

In a similar way to the proof of Theorem 8.1, we expand \(\varvec{f}(Y_i,Y'_i)\) at \((\varvec{\tilde{y}},\varvec{\tilde{y}}')\) for the general ERKN method (8.4), where \(\varvec{\tilde{y}}=\phi _0(c_i^2V)\varvec{y}_n+\phi _1(c_i^2V)c_ih\varvec{y}'_n\) and \(\varvec{\tilde{y}'}=\phi _0(c_i^2V)\varvec{y}' _n-c_ihM\phi _1(c_i^2V)\varvec{y}_n\), and obtain the Taylor series expansion as follows:

$$\begin{aligned} \varvec{f}(Y_i,Y'_i)=\sum \limits _{m,n\ge 0}\frac{1}{(m+n)!}\left. \varvec{f}_{\varvec{y}^m\varvec{y}'^n}^{(m+n)} \right| _{\varvec{\tilde{y}},\varvec{\tilde{y}'}} \Big (h^2\sum _j\bar{a}_{ij}(V)\varvec{f}(Y_j,Y'_j)\Big )^{\otimes m}\otimes \Big (h\sum _ja_{ij}(V)\varvec{f}(Y_j,Y'_j)\Big )^{\otimes n}, \end{aligned}$$
(8.13)

where the second term \(\left. \varvec{f}_{\varvec{y}^m\varvec{y}'^n}^{(m+n)} \right| _{\varvec{\tilde{y}},\varvec{\tilde{y}'}}\) is a function of \(c_ih\). Then the Taylor series expansion of \(\left. \varvec{f}_{\varvec{y}^m\varvec{y}'^n}^{(m+n)}\right| _{(\varvec{\hat{y}},\varvec{\hat{y}'})}\) at \(h=0\) is given by

$$\begin{aligned} \left. \varvec{f}_{\varvec{y}^m\varvec{y}'^n}^{(m+n)}\right| _{(\varvec{\hat{y}},\varvec{\hat{y}'})} =\sum \limits _{N\ge 0}\frac{c_i^N}{m!}h^ND_h^{N}\varvec{f}_{\varvec{y}^m\varvec{y}'^n}^{(m+n)}. \end{aligned}$$
(8.14)

Definition 8.3 ensures that each \(\varvec{f}(Y_i,Y'_i)\) for \(i=1,\ldots ,s\) is a B-series. In fact, the third and fourth terms in the Eq. (8.13) are given by

$$\begin{aligned} \begin{array}{ll} h^2\sum _j\bar{a}_{ij}(V)\varvec{f}(Y_j,Y'_j)&=\sum \limits _{\tau \in {{\mathrm{IEN-T}}}}\sum \limits _{p\ge 0} \frac{\sum _j\bar{a}_{ij}^{(2p)}}{\rho (\tau )!}\frac{V^p}{(2p)!}h^{\rho (\tau )+1}\varvec{a}_j(\tau ), \end{array} \end{aligned}$$
(8.15)

and

$$\begin{aligned} \begin{array}{ll} h\sum _ja_{ij}(V)\varvec{f}(Y_j,Y'_j)&=\sum \limits _{\tau \in {{\mathrm{IEN-T}}}}\sum \limits _{q\ge 0} \frac{\sum _j a_{ij}^{(2q)}}{\rho (\tau )!}\frac{V^q}{(2q)!}h^{\rho (\tau )}\varvec{a}_j(\tau ). \end{array} \end{aligned}$$
(8.16)

We then obtain

$$\begin{aligned} \begin{array}{rl} \varvec{f}(Y_i,Y'_i) =\sum \limits _{N,n,m}\sum \limits _{\tau \in {\mathrm{IEN-T}}} &{} \frac{c_i^N h^s}{N!(n+m)!}D_h^{N}\varvec{f}_{\varvec{y}^m\varvec{y}'^n}^{(m+n)} \Big (\frac{\sum _j\bar{a}_{ij}^{(2p_1)}}{\rho (\tau _1)!}\frac{M^{p_1}}{(2p_1)!}\varvec{a}_j(\tau _1),\ldots , \frac{\sum _j\bar{a}_{ij}^{(2p_m)}}{\rho (\tau _m)!}\frac{M^{p_m}}{(2p_m)!}\varvec{a}_j(\tau _m),\\ {} &{} \frac{\sum _ja_{ij}^{(2q_1)}}{\rho (\tau _{m+1})!}\frac{M^{q_1}}{(2q_1)!}\varvec{a}_j(\tau _{m+1}),\ldots , \frac{\sum _ja_{ij}^{(2q_n)}}{\rho (\tau _{m+n})!}\frac{M^{q_n}}{(2q_n)!}\varvec{a}_j(\tau _{m+n})\Big ), \end{array} \end{aligned}$$
(8.17)

where \(s=N+\sum \limits _{k=1}^m(2p_k+\rho (\tau _k)+1)+\sum \limits _{k=1}^n(2q_k+\rho (\tau _{m+k}))\). Using Definition 8.3, we complete the proof.

8.6 The Order Conditions for the General ERKN Method

Theorem 8.3

The scheme (8.4) for the general multi-frequency and multidimensional oscillatory second-order initial value problems (8.1) has order r if and only if the following conditions

$$\begin{aligned} \sum \limits _{i=1}^{s}\bar{b}_i(V)S(\tau )\gamma (\tau )\varPhi _i(\tau )= & {} \rho (\tau )!\phi _{\rho (\tau )+1}+O(h^{r-\rho (\tau )}), \quad \forall \tau \in \text {IEN-T}_m,\;m\le r-1, \nonumber \end{aligned}$$
(8.18)
$$\begin{aligned} \\ \sum \limits _{i=1}^{s}b_i(V)S(\tau )\gamma (\tau )\varPhi _i(\tau )= & {} \rho (\tau )!\phi _{\rho (\tau )}+O(h^{r-\rho (\tau )+1}),\quad \forall \tau \in \text {IEN-T}_m ,\;m\le r, \nonumber \\ \end{aligned}$$
(8.19)

are satisfied.

Proof

It follows from the matrix-variation-of-constants formula, Theorems 8.1 and 8.2 that

$$\begin{aligned} \begin{aligned} \varvec{y}_{n+1}&=\phi _0(V)\varvec{y}_n+h\phi _1(V)\varvec{y}'_n \\&+\sum \limits _{\tau \in {{\mathrm{IEN-T}}}}\frac{h^{\rho (\tau )+1}}{\rho (\tau )!}\sum \limits _{i=1}^s \bar{b}_i(V)\varPhi _i(\tau )S(\tau )\gamma (\tau )\alpha (\tau )\mathscr {F}(\tau )(\varvec{y}_n,\varvec{y}'_n), \end{aligned} \end{aligned}$$
(8.20)
$$\begin{aligned} \begin{aligned} \varvec{y}(t+h)&=\phi _0(V)\varvec{y}+h\phi _1(V)\varvec{y}'\\ {}&+\sum \limits _{\tau \in \text {IEN-T}}h^{\rho (\tau )+1}\alpha (\tau ) \mathscr {F}(\tau )(\varvec{y},\varvec{y}) \int _0^1(1-z)\frac{z^{\rho (\tau )-1}}{(\rho (\tau )-1)!}\phi _1((1-z)V)\,\mathrm {d}z. \end{aligned} \end{aligned}$$
(8.21)

Comparing the Eqs. (8.20) with (8.21) and using the properties of the \(\phi \)-functions, we obtain the first result of Theorem 8.3. Likewise, we deduce the second part of the theorem.

Theorem 8.3 in this chapter and Theorem 4.1 in [33] share the same expression. However, it should be noted that there exist redundant order conditions in [33], while any order condition in this chapter cannot be replaced by others, provided the entries \(\bar{a}_{ij}(V)\), \(a_{ij}(V)\), \(b_i(V)\) and \(\bar{b}_i(V)\) in the general ERKN method (8.4) are independent. Obviously, the elimination of redundant order conditions makes the construction of high-order general ERKN methods (8.4) much clearer and simpler.

It is easy to see that Theorem 8.3 implies the order conditions for the standard ERKN methods in [23, 30] when the right-hand side function \(\varvec{f}\) does not depend on \(\varvec{y}'\). It is noted that, if the matrix M is null, Theorem 8.3 reduces to the classical general RKN method when applied to \(\varvec{y}''=\varvec{f}(\varvec{y},\varvec{y}')\), since the IEN-T set is exactly the N-T set in this special case.

8.7 The Construction of General ERKN Methods

In this section, using Theorem 8.3, we present some general ERKN methods (8.4) of order up to 4. The approach to constructing new methods in this section is different from that described in [33].

8.7.1 Second-Order General ERKN Methods

From Theorem 8.3 and the three IEN-Ts with order no more than 2 which are listed in Table 8.5, for an s-stage general ERKN method (8.4) expressed in the Butcher tableau (8.5), we have the following second order conditions:

$$\begin{aligned} \begin{array}{cc} \sum \limits _{i=1}^s \bar{b}_i(V)=\phi _2(V)+O(h), &{} \sum \limits _{i=1}^s b_i(V)=\phi _1(V)+O(h^2), \\ \sum \limits _{i=1}^s b_i(V)c_i=\phi _2(V)+O(h), &{} \sum \limits _{i=1}^s b_i(V)a_{ij}^{(0)}=\phi _2(V)+O(h). \end{array} \end{aligned}$$

Comparing the coefficients of \(h^0\) and h, we obtain 4 equations:

$$\begin{aligned} \sum \limits _{i=1}^s \bar{b}_i^{(0)}=\frac{1}{2}, \quad \sum \limits _{i=1}^s b_i^{(0)}=1,\quad \sum \limits _{i=1}^s b_i^{(0)}c_i=\frac{1}{2}, \quad \sum \limits _{i=1}^s b_i^{(0)}a_{ij}^{(0)}=\frac{1}{2}. \end{aligned}$$

It can be observed that these equations are exactly the second order conditions for the following traditional RKN method

$$\begin{aligned} \left\{ \begin{aligned}&Y_i=\varvec{y}_n+c_ih\varvec{y}'_n +h^2\sum \limits _{j=1}^{s}\bar{a}_{ij}^{(0)}\Big (\varvec{f}(Y_j,Y'_j)-M Y_j\Big ), \qquad i=1,\cdots ,s, \\&Y'_{i}=\varvec{y}'_n+h\sum \limits _{j=1}^{s}a_{ij}^{(0)}\Big (\varvec{f}(Y_j,Y'_j)-M Y_j\Big ), \qquad i=1,\cdots ,s, \\&\varvec{y}_{n+1}=\varvec{y}_n+h\varvec{y}'_n + h^2\sum \limits _{i=1}^s\bar{b}_i^{(0)}\Big (\varvec{f}(Y_i,Y'_i)-M Y_i\Big ),\\&\varvec{y}'_{n+1}=\varvec{y}'_n+h\sum \limits _{i=1}^s b_i^{(0)}\Big (\varvec{f}(Y_i,Y'_i)-M Y_i\Big ), \end{aligned}\right. \end{aligned}$$
(8.22)

applied to the initial value problems (8.1), with the tableau

$$\begin{aligned} \begin{array}{c|cccc|cccc} c_1 &{} \bar{a}_{11}^{(0)} &{}\bar{a}_{12}^{(0)} &{} \cdots &{}\bar{a}_{1s}^{(0)} &{} a_{11}^{(0)} &{}a_{12}^{(0)} &{} \cdots &{}a_{1s}^{(0)} \\ c_2 &{} \bar{a}_{21}^{(0)} &{}\bar{a}_{22}^{(0)} &{} \cdots &{}\bar{a}_{2s}^{(0)} &{} a_{21}^{(0)} &{} a_{22}^{(0)} &{} \cdots &{} a_{2s}^{(0)} \\ \vdots &{}\vdots &{}\vdots &{}\ddots &{}\vdots &{} \vdots &{} \vdots &{}\ddots &{}\vdots \\ c_s &{} \bar{a}_{s1}^{(0)} &{} \bar{a}_{s2}^{(0)} &{}\cdots &{} \bar{a}_{s,s}^{(0)} &{} a_{s1}^{(0)} &{} a_{s2}^{(0)} &{}\cdots &{} a_{s,s}^{(0)} \\ \hline &{} \bar{b}_1^{(0)} &{} \bar{b}_2^{(0)} &{}\cdots &{} \bar{b}_{s}^{(0)} &{} b_1^{(0)} &{} b_2^{(0)} &{}\cdots &{} b_{s}^{(0)} \end{array}. \end{aligned}$$
(8.23)

This means that we can easily solve \(\Big (c_i,\bar{a}_{ij}^{(0)},a_{ij}^{(0)},\bar{b}_i^{(0)},b_i^{(0)}\Big )\) in terms of a classical general RKN method. For example, from the explicit 2 stage second-order general RKN method with the Butcher tableau

$$\begin{aligned} \begin{array}{c|cc|cc} 0 &{} &{} &{} &{}\\ \frac{2}{3} &{}0&{} &{}{\frac{2}{3}}_{}&{} \\ \hline &{} {\frac{1}{4}}^{} &{} \frac{3}{4}&{} \frac{1}{4} &{} \frac{1}{4} \end{array}, \end{aligned}$$
(8.24)

we can obtain 2 stage second-order explicit general ERKN methods. Two examples are given below.

Example 1

The first 2 stage second-order explicit general ERKN method (8.4) has Butcher tableau

$$\begin{aligned} \begin{array}{c|cc|cc} 0 &{} &{} &{} &{}\\ \frac{2}{3} &{}0 &{} &{}{\frac{2}{3}}_{}I&{} \\ \hline &{} \frac{1}{4}I &{} \frac{3}{4}I&{} {\frac{1}{4}}^{}I &{} \frac{1}{4}I \end{array}. \end{aligned}$$
(8.25)

Example 2

The Butcher tableau of the second one is

$$\begin{aligned} \begin{array}{c|cc|cc} 0 &{} &{} &{} &{}\\ {\frac{2}{3}}_{} &{}0&{} &{}\frac{2}{3}\phi _0(\frac{4}{9}V)&{} \\ \hline &{} \frac{1}{4}\phi _1(V) &{} \frac{3}{4}\phi _1(\frac{1}{9}V)&{} {\frac{1}{4}}^{}\phi _0(V) &{} \frac{1}{4}\phi _0(\frac{1}{9}V) \end{array}. \end{aligned}$$
(8.26)

8.7.2 Third-Order General ERKN Methods

From Theorem 8.3 and 9 trees in the set of IEN-T\(_m, (m\le 3)\) in Table 8.5, for an s-stage general ERKN method (8.4) expressed in the Butcher tableau (8.5), we have the third order conditions as follows:

$$\begin{aligned} \begin{array}{ccc} \sum \limits _{i=1}^s \bar{b}_i(V)=\phi _2(V)+O(h^2), &{} \sum \limits _{i=1}^s \bar{b}_i(V)c_i=\phi _3(V)+O(h), &{} \sum \limits _{i=1}^s\sum \limits _{j=1}^s \bar{b}_i(V)a_{ij}^{(0)}=\phi _3(V)+O(h), \\ \sum \limits _{i=1}^s b_i(V)=\phi _1(V)+O(h^3), &{} \sum \limits _{i=1}^s b_i(V)c_i=\phi _2(V)+O(h^2), &{} \sum \limits _{i=1}^s\sum \limits _{j=1}^s b_i(V)a_{ij}^{(0)}=\phi _2(V)+O(h^2), \\ \sum \limits _{i=1}^s b_i(V)c_i^2=2\phi _3(V)+O(h), &{} \sum \limits _{i=1}^s \sum \limits _{j=1}^s b_i(V)c_ia_{ij}^{(0)}=2\phi _3(V)+O(h), &{} \sum \limits _{i=1}^s \sum \limits _{j=1}^s \sum \limits _{k=1}^s b_i(V)a_{ij}^{(0)}a_{ik}^{(0)}=2\phi _3(V)+O(h), \\ \sum \limits _{i=1}^s b_i(V)\bar{a}_{ij}^{(0)}=\phi _3(V)+O(h), &{} \sum \limits _{i=1}^s \sum \limits _{j=1}^s b_i(V)a_{ij}^{(0)}c_j=\phi _3(V)+O(h), &{} \sum \limits _{i=1}^s \sum \limits _{j=1}^s \sum \limits _{k=1}^s b_i(V)a_{ij}^{(0)}a_{jk}^{(0)}=\phi _3(V)+O(h). \end{array} \end{aligned}$$

Equating coefficients for each power of h, we obtain 13 equations, where 12 equations are exactly the third order conditions for the classical general RKN method (8.22) with the Butcher tableau (8.23)

$$\begin{aligned}&\sum \limits _{i=1}^{s}\bar{b}_i^{(0)}\gamma (\tau )\varPhi _i(\tau )=\frac{1}{\rho (\tau )+1},\quad \forall \tau \in \text {N-T}_m,\;m\le 2,\end{aligned}$$
(8.27)
$$\begin{aligned}&\sum \limits _{i=1}^{s}b_i^{(0)}\gamma (\tau )\varPhi _i(\tau )=1,\quad \forall \tau \in \text {N-T}_m ,\;m\le 3. \end{aligned}$$
(8.28)

The extra equation is \(\sum \limits _{i=1}^{s}\bar{b}_i^{(2)}=-\frac{1}{3}\). We can solve \(\Big (c_i,\bar{a}_{ij}^{(0)},a_{ij}^{(0)},\bar{b}_i^{(0)},b_i^{(0)}\Big )\) from the Eqs. (8.27) and (8.28) via a classical general RKN method. We can then find \(b_i^{(2)}\) from the extra equation. Using this approach, we can complete the construction of the general ERKN methods of order three. For example, from the explicit 3 stage third-order general RKN method with the Butcher tableau

$$\begin{aligned} \begin{array}{c|ccc|ccc} 0 &{} &{} &{} &{} &{}&{}\\ \frac{1}{2} &{}0&{} &{} &{}\frac{1}{2}&{} &{} \\ 1 &{}1&{}0 &{} &{}-1&{}2 &{} \\ \hline &{} \frac{1}{6} &{} {\frac{2}{6}}^{}&{}0&{} \frac{1}{6}&{} \frac{4}{6}&{}\frac{1}{6} \end{array} \end{aligned}$$
(8.29)

we can construct the 3 stage third-order explicit general ERKN methods straightforwardly. The three examples are listed below.

Example 3

The first 3 stage third-order explicit general ERKN method (8.4) is expressed in the Butcher tableau

$$\begin{aligned} \begin{array}{c|ccc|ccc} 0 &{} &{} &{} &{} &{}&{}\\ \frac{1}{2} &{}0&{} &{} &{}\frac{1}{2} I&{} &{} \\ 1 &{} I&{}0 &{} &{}- I&{}2 I &{} \\ \hline &{} \frac{1}{6} I &{} {\frac{2}{6}}^{} I&{}0&{} \frac{1}{6} (I-\frac{9}{20}V)&{} \frac{4}{6}( I-\frac{3}{20}V)&{}\frac{1}{6} (I+\frac{1}{20}V) \end{array}. \end{aligned}$$
(8.30)

Example 4

The Butcher tableau of the second 3 stage third-order explicit general ERKN method (8.4) is given by

$$\begin{aligned} \begin{array}{c|ccc|ccc} 0 &{} &{} &{} &{} &{}&{}\\ \frac{1}{2} &{}0&{} &{} &{}\frac{1}{2}I&{} &{} \\ 1 &{}I&{}0 &{} &{}-I&{}2I &{} \\ \hline &{} \frac{1}{6}(I-\frac{1}{6}V) &{} \frac{2}{6}(I-\frac{1}{24}V)&{}0&{} \frac{1}{6}(I-\frac{1}{2}V) &{} \frac{4}{6}(I-\frac{1}{8}V)&{}{\frac{1}{6}}^{}I \end{array}. \end{aligned}$$
(8.31)

Example 5

The third 3 stage third-order explicit general ERKN method (8.4) is denoted by the Butcher tableau

$$\begin{aligned} \begin{array}{c|ccc|ccc} 0 &{} &{} &{} &{} &{}&{}\\ \frac{1}{2} &{}0&{} &{} &{}\frac{1}{2}\phi _0(\frac{1}{4}V)&{} &{} \\ 1 &{}\phi _1(V)&{}0 &{} &{}-\phi _0(V)&{}2\phi _0(\frac{1}{4}V)_{} &{} \\ \hline &{} \frac{1}{6} \phi _1(V)&{} \frac{2}{6}\phi _1(\frac{1}{4}V)&{}0&{} \frac{1}{6}\phi _0(V)&{} \frac{4}{6}\phi _0(\frac{1}{4}V)&{}{\frac{1}{6}}^{}I \end{array}. \end{aligned}$$
(8.32)

8.7.3 Fourth-Order General ERKN Methods

From Theorem 8.3 and Table 8.5, comparing the coefficients of the power of h of (8.18) and (8.19), for an s-stage general ERKN method (8.4) with the coefficient \(\Big (\bar{a}_{ij}(V),a_{ij}(V),\bar{b}_{i}(V),b_{i}(V)\Big )\) displayed in the Butcher tableau (8.5), we can obtain 41 fourth order conditions, in which 36 conditions are as follows:

$$\begin{aligned}&\sum \limits _{i=1}^{s}\bar{b}_i^{(0)}\gamma (\tau )\varPhi _i(\tau )=\frac{1}{\rho (\tau )+1},\quad \forall \tau \in \text {N-T}_m,\;m\le 3,\end{aligned}$$
(8.33)
$$\begin{aligned}&\sum \limits _{i=1}^{s}b_i^{(0)}\gamma (\tau )\varPhi _i(\tau )=1,\quad \forall \tau \in \text {N-T}_m ,\;m\le 4. \end{aligned}$$
(8.34)

The remaining 5 conditions are

$$\begin{aligned} \begin{aligned}&\sum \limits _{i=1}^{s}\sum \limits _{j=1}^{s}b_i^{(0)}a_{ij}^{(2)}=-\frac{1}{12}, \; \sum \limits _{i=1}^{s}b_i^{(2)}=-\frac{1}{3},\; \sum \limits _{i=1}^{s}b_i^{(2)}c_i=-\frac{1}{12}, \\&\sum \limits _{i=1}^{s}\sum \limits _{j=1}^{s}b_i^{(2)}a_{ij}^{(0)}=-\frac{1}{12}, \; \sum \limits _{i=1}^{s}\bar{b}_i^{(2)}=-\frac{1}{12}. \end{aligned} \end{aligned}$$
(8.35)

For each specific classical general RKN method of order four, we can solve for \(\Big (c_i,\bar{a}_{ij}^{(0)},a_{ij}^{(0)},\bar{b}_i^{(0)},b_i^{(0)}\Big )\) from (8.33) and (8.34), since these 36 conditions are exactly the order conditions for the classical general RKN method (8.22) with the Butcher tableau (8.23). Then we can find \(\Big (a_{ij}^{(2)}, \bar{b}_i^{(2)},b_i^{(2)}\Big )\) from conditions (8.35). In this way, we construct the general ERKN methods (8.4) of order four.

In what follows, we will construct explicit 4 stage fourth order general ERKN methods from the following explicit 4 stage fourth-order classical general RKN method (8.22) with the Butcher tableau

$$\begin{aligned} \begin{array}{c|cccc|cccc} 0 &{} &{} &{} &{} &{} &{} &{} &{} \\ \frac{1}{2} &{} \frac{1}{8} &{} &{} &{} &{} \frac{1}{2} &{} &{} &{} \\ \frac{1}{2} &{} \frac{1}{8} &{} 0 &{} &{} &{} 0 &{} \frac{1}{2} &{} &{} \\ 1 &{} 0 &{} 0 &{} {\frac{1}{2}}_{} &{} &{} 0 &{} 0 &{} 1 &{} \\ \hline &{} \frac{1}{6} &{} \frac{1}{6} &{} \frac{1}{6} &{} 0 &{} \frac{1}{6} &{}\frac{2}{6} &{} \frac{2}{6} &{}{\frac{1}{6}}^{} \end{array}. \end{aligned}$$
(8.36)

Some general ERKN methods of order four constructed in this approach are shown below.

Example 6

The Butcher tableau of the first explicit 4 stage fourth-order general ERKN method (8.4) is given by

$$\begin{aligned} \begin{array}{c|cccc|cccc} 0 &{} &{} &{} &{} &{} &{} &{} &{} \\ \frac{1}{2} &{} {\frac{1}{8}}_{} I &{} &{} &{} &{} \frac{1}{2} I &{} &{} &{} \\ \frac{1}{2} &{} \frac{1}{8} I &{} 0 &{} &{} &{} 0 &{} \frac{1}{2} I &{} &{} \\ 1 &{} 0 &{} 0 &{} \frac{1}{2} I &{} &{} 0 &{} 0 &{} I-{\frac{1}{4}}_{}V &{} \\ \hline &{} \frac{1}{6}(I-\frac{1}{12}V)&{} \frac{1}{6}(I-\frac{1}{12}V) &{} \frac{1}{6}(I-\frac{1}{12}V) &{} 0 &{} \frac{1}{6}(I-\frac{1}{2}V) &{}\frac{2}{6}(I-\frac{1}{8}V) &{} \frac{2}{6}(I-\frac{1}{8}V) &{}{\frac{1}{6}}^{}I \end{array}.\nonumber \\ \end{aligned}$$
(8.37)

Example 7

The second explicit 4 stage fourth-order general ERKN method is expressed in the Butcher tableau

$$\begin{aligned} \begin{array}{c|cccc|cccc} 0 &{} &{} &{} &{} &{} &{} &{} &{} \\ \frac{1}{2} &{} {\frac{1}{8}}_{} I &{} &{} &{} &{} \frac{1}{2}( I-\frac{1}{8}V) &{} &{} &{} \\ \frac{1}{2} &{} \frac{1}{8} I &{} 0 &{} &{} &{} 0 &{} \frac{1}{2} I &{} &{} \\ 1 &{} 0 &{} 0 &{} \frac{1}{2} I &{} &{} 0 &{} 0 &{} I-{\frac{1}{8}}_{}V &{} \\ \hline &{} \frac{1}{6}( I-\frac{1}{6}V)&{} \frac{1}{6}( I-\frac{1}{24}V) &{} \frac{1}{6}( I-\frac{1}{24}V) &{} 0 &{} \frac{1}{6}( I-\frac{1}{2}V) &{}\frac{2}{6}( I-\frac{1}{8}V) &{} \frac{2}{6}( I-\frac{1}{8}V) &{}{\frac{1}{6}}^{} I \end{array}.\nonumber \\ \end{aligned}$$
(8.38)

Example 8

The third explicit 4 stage fourth-order general ERKN method (8.4) has the Butcher tableau as follows:

$$\begin{aligned} \begin{array}{c|cccc|cccc} 0 &{} &{} &{} &{} &{} &{} &{} &{} \\ \frac{1}{2} &{} \frac{1}{8}\phi _1{(\frac{1}{4}V)}_{} &{} &{} &{} &{}\frac{1}{2}\phi _0(\frac{1}{4}V) &{} &{} &{} \\ \frac{1}{2} &{} \frac{1}{8}\phi _1(\frac{1}{4}V) &{} 0 &{} &{} &{} 0 &{}\frac{1}{2} I &{} &{} \\ 1&{} 0 &{} 0 &{} \frac{1}{2}\phi _1(\frac{1}{4}V) &{} &{} 0 &{} 0 &{} \phi _0(\frac{1}{4}V)_{} &{} \\ \hline &{} \frac{1}{6}\phi _1(V) &{} \frac{1}{6}\phi _1(\frac{1}{4}V) &{} \frac{1}{6}\phi _1(\frac{1}{4}V) &{} 0 &{} \frac{1}{6}\phi _0(V) &{}\frac{2}{6}\phi _0(\frac{1}{4}V) &{} \frac{2}{6}\phi _0(\frac{1}{4}V)&{}{\frac{1}{6}}^{} I \end{array}. \end{aligned}$$
(8.39)

8.7.4 An Effective Approach to Constructing the General ERKN Methods

In the paper [33], in order to construct 4th order general ERKN methods for the systems (8.1), the authors first considered all 62 graphs of the EN-Ts (see Tables 1 and 2 in [33]), and then selected and deleted 34 redundant trees. Finally, they obtained 28 non-redundant EN-Ts (see Tables 3 and 4 in [33]). With these 28 EN-Ts, the authors in [33] achieved special 4th-order conditions, and then the authors derived a 4th-order ERKN method under two auxiliary simplifying assumptions.

Obviously, as shown in the paper [33] more than half of the construction effort was spent on drawing the redundant trees. In a word, the process described in the paper [33] is difficult to follow since the number of the redundant trees in the EN-T set is large.

However, in this chapter, these 28 trees can be directly obtained since 27 of them are exactly the classical N-Ts as shown in Sect. 8.4.2. In this way, it becomes quite easy to get the 4th-order conditions for the general ERKN method (8.4). Then using expansions of these order conditions, and equating each power of h, we can see that most are exactly the order conditions for the classical general RKN method (8.22). This approach to constructing the general ERKN integrators is very effective and efficient in practice, as shown in the previous sections where 2nd, 3rd and 4th order general ERKN methods are constructed as examples.

8.8 Numerical Experiments

In this section, some numerical experiments are implemented to illustrate the potential of the general ERKN methods (8.4) in comparison with the others in the literature. The criterion used in the numerical comparisons is the base-10 logarithm of the maximum global error (\(\log _{10}\Vert \text {MGE}\Vert \)) versus the base-2 logarithm of the stepsizes (\(\log _2(h)\)). The following 11 methods are used to solve the general system (8.1) for the comparison:

  • RKN2: The 2 stage second-order general RKN method (8.24).

  • ERKN2a: The first 2 stage second-order general ERKN method (8.25) given in Sect. 8.7 of this chapter.

  • ERKN2b: The second 2 stage second-order general ERKN method (8.26) given in Sect. 8.7 of this chapter.

  • RKN3: The 3 stage third-order general RKN method (8.29).

  • ERKN3a: The first 3 stage third-order general ERKN method (8.30) given in Sect. 8.7 of this chapter.

  • ERKN3b: The second 3 stage third-order general ERKN method (8.31) given in Sect. 8.7 of this chapter.

  • ERKN3c: The third 3 stage third-order general ERKN method (8.32) given in Sect. 8.7 of this chapter.

  • RKN4: The 4 stage fourth-order general RKN method (8.36).

  • ERKN4a: The first 4 stage fourth-order general ERKN method (8.37) given in Sect. 8.7 of this chapter.

  • ERKN4b: The second 4 stage fourth-order general ERKN method (8.38) given in Sect. 8.7 of this chapter.

  • ERKN4c: The third 4 stage fourth-order general ERKN method (8.39) given in Sect. 8.7 of this chapter.

Problem 1

We consider the damped equation

$$\begin{aligned} my''+by'+ky=0, \end{aligned}$$

as one of the test problems. When the damping constant b is small we would expect the system to still oscillate, but with decreasing amplitude as its energy is converted to heat. In this numerical test, the problem is integrated on the interval [0, 300] with \(m=1\), \(b=0.01\), \(k=3\) and the initial conditions \(\big (y(0),y'(0)\big )= \big (1,0\big )\). The analytic solution to the problem is given by

$$\begin{aligned} y(t)=e^{-\frac{0.01}{2}t}\Big (\cos (\frac{\sqrt{12-0.01^2}}{2}t)+\frac{0.01}{\sqrt{12-0.01^2}}\sin (\frac{\sqrt{12-0.01^2}}{2}t)\Big ). \end{aligned}$$

The numerical results are displayed in Fig. 8.3, where the small stepsizes for the methods are \(h=\frac{1}{2^j}\) for \( j=3,\ldots ,8\) and the big stepsizes are \(h=\frac{j}{8}\) for \( j=2,\ldots ,6\).

Fig. 8.3
figure 3

Problem 1 integrated on [0, 300]

Problem 2

We consider the initial value problem

$$\begin{aligned} y''(t)+\left( \begin{array}{cc}13&{}-12\\ -12&{}13\end{array}\right) y(t) = \frac{12\varepsilon }{5}\left( \begin{array}{cc}3&{}2\\ -2&{}-3\end{array}\right) y'(t) +\varepsilon ^2\left( \begin{array}{c}\frac{36}{5}\sin (t)+24\sin (5t)\\ -\frac{24}{5}\sin (t)-36\sin (5t)\end{array}\right) , \end{aligned}$$

with the initial values \(y(0)=(\varepsilon ,\varepsilon )^\intercal \) and \(y'(0)=(-4,6)^\intercal \). The analytic solution is given by

$$\begin{aligned} y(t)=\left( \begin{array}{c}\sin (t)-\sin (5t)+\varepsilon \cos (t)\\ \sin (t)+\sin (5t)+\varepsilon \cos (5t)\end{array}\right) . \end{aligned}$$

In the numerical experiment, we choose the parameter value \(\varepsilon =10^{-3}\) and integrate this problem on the interval [0, 300]. The numerical results are displayed in Fig. 8.4. The small stepsizes are \(h=\frac{1}{2^j}\) for \( j=3,\ldots ,8\) and the big stepsizes are \(h=\frac{j}{8}\) for \( j=2,\ldots ,6\). In this numerical test with the big stepsizes, the classical general RKN methods (RKN2, RKN3 and RKN4) give disappointing numerical results. Thus we do not depict the corresponding points in Fig. 8.4.

Fig. 8.4
figure 4

Problem 2 integrated on [0, 300]

Problem 3

Consider the damped wave equation with periodic conditions (wave propagation in a medium, see e.g. Weinberger [18])

$$\begin{aligned} \left\{ \begin{array}{l} \frac{\partial ^2u}{\partial t^2}+\delta \frac{\partial u}{\partial t}=\frac{\partial ^2 u}{\partial x^2}-f(u), \quad -1<x<1,\; t>0,\\ {} u(-1,t)=u(1,t), \end{array}\right. \end{aligned}$$

where \(f(u)=-\sin {u},\) (i.e., the damped sine Gordon equation) and \(\delta =1\). A semi-discretization in the spatial variable by second-order symmetric differences leads to the following system of second-order ODEs in time

$$\begin{aligned} \ddot{U}+MU=F(U,\dot{U}),\quad 0<t\le t_{end}, \end{aligned}$$

where \(U(t)=\big (u_1(t),\cdots ,u_N(t)\big )^\intercal \) with \(u_i(t)\approx u(x_i,t)\) for \(i=1,\ldots ,N,\)

$$\begin{aligned} M=\frac{1}{\triangle x^2} \left( \begin{array}{ccccc} 2&{}-1&{} &{} &{} -1\\ -1&{}2 &{} -1 &{} &{} \\ &{} \ddots &{} \ddots &{} \ddots &{} \\ &{} &{} -1&{}2&{}-1\\ -1&{} &{} &{} -1 &{}2 \end{array}\right) , \end{aligned}$$

\(\triangle x=2/N\), \(x_i=-1+i\triangle x\) and \(F(U,\dot{U})=\big (f(u_1)-\delta \dot{u}_1,\cdots ,f(u_N)-\delta \dot{u}_N\big )^\intercal \). Following the paper [3], we take the initial conditions as

$$\begin{aligned} U(0)=\Big (\pi ,\cdots ,\pi \Big )^\intercal ,\quad U_t(0)=\sqrt{N}\Big (0.01+\sin \big (\frac{2\pi }{N}\big ),\cdots , 0.01+\sin \big (\frac{2\pi N}{N}\big )\Big )^\intercal , \end{aligned}$$

with \(N=64\) and integrate the problem on the interval [0, 300] with small stepsizes \(h=\frac{1}{2^j}\) for \( j=5,\ldots ,8\) and with big stepsizes \(h=\frac{j}{128}\) for \( j=5,6,8,10\). The numerical results are displayed in Fig. 8.5. In this numerical test for the big stepsizes, the classical general RKN methods (RKN2, RKN3 and RKN4) all behave badly, yielding large errors.

Fig. 8.5
figure 5

Problem 3 integrated on [0, 300]

It can be observed from Figs. 8.3, 8.4 and 8.5 that

  • The general ERKN methods perform more efficiently than the classical general RKN methods.

  • The higher order general ERKN methods are more efficient than the lower ones.

  • As the stepsize decreases, the difference among the general ERKN methods of the same order becomes negligible.

  • The general ERKN methods behave perfectly for the large stepsizes.

8.9 Conclusions and Discussions

In this chapter, we have established an improved theory for the order conditions for the general ERKN methods designed specially for solving multi-frequency oscillatory system (8.1). The original tri-colored tree theory and the order conditions for the general ERKN methods presented in the paper [33] are not satisfied yet due to the existence of large numbers of redundant trees. This chapter has succeeded in making a simplification, by defining the IEN-T set on which some special mappings (especially the extended elementary differential mapping) are introduced.

This simplification of the order conditions for the general ERKN methods when applied to the oscillatory system (8.1) is of great importance. The new tri-colored tree theory and the B-series theory for the general ERKN methods when solving the general system (8.1) reduce to those for standard ERKN methods when solving special system (8.2), where the right-hand side vector-valued function \(\varvec{f}\) does not depend on \(\varvec{y}'\) (see [23, 30]).

This successful simplification makes the construction of the general ERKN methods much simpler and more efficient for the system (8.1). In light of the reduced tree theory analysed in this chapter, almost one half of algebraic conditions in the paper [33] can be eliminated. Furthermore, in this chapter, from the relation between the theories of order conditions for the general RKN method and for general ERKN method, we propose a simple approach to constructing new integrators. The numerical results show that the general ERKN methods are more suitable for long-term integration with a large stepsize, in comparison with the RKN methods in the literature.

The previous eight chapters concentrated on numerical integrators of oscillatory ordinary differential equations, although their applications to partial differential equations were implemented as well. However, in the next four chapters we will turn to structure-preserving schemes for partial differential equations.

The material of this chapter is based on the work by Zeng et al. [34].