Abstract
We provide an approximate zero \(\widetilde{S}(g,L)\) for the hyperbolic Kepler’s equation \(S-g\,{{\mathrm{arcsinh}}}(S)-L=0\) for \(g\in (0,1)\) and \(L\in [0,\infty )\). We prove, by using Smale’s \(\alpha \)-theory, that Newton’s method starting at our approximate zero produces a sequence that converges to the actual solution S(g, L) at quadratic speed, i.e. if \(S_n\) is the value obtained after n iterations, then \(|S_n-S|\le 0.5^{2^n-1}|\widetilde{S}-S|\). The approximate zero \(\widetilde{S}(g,L)\) is a piecewise-defined function involving several linear expressions and one with cubic and square roots. In bounded regions of \((0,1) \times [0,\infty )\) that exclude a small neighborhood of \(g=1, L=0\), we also provide a method to construct simpler starters involving only constants.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Keplerian orbits can be described by two parameters, the semilatus rectum \(p>0\) and the eccentricity \(e\ge 0\), from which all other orbital constants can be derived. According to Kepler’s laws, the coordinates (x, y) at time t of a satellite in the standard reference frame (central body at the origin and x-axis pointing to the periapsis) are given by
Elliptic (\(e<1\)) | Parabolic (\(e=1\)) | Hyperbolic (\(e>1\)) |
---|---|---|
\(x=\frac{p}{1-e^2}(\cos (E)-e)\) | \(x=\frac{p}{2}(1-D^2)\) | \(x=\frac{p}{1-e^2}(\cosh (H)-e)\) |
\(y=\frac{p}{\sqrt{1-e^2}}\sin (E)\) | \(y=pD\) | \(y=\frac{p}{\sqrt{e^2-1}}\sinh (H)\) |
\(E-e\sin (E)=M\) | \(D+\frac{D^3}{3}=M\) | \(e\sinh (H)-H=M\) |
\(M=\sqrt{\frac{\mu (1-e^2)^3}{p^3}}(t-t_0)\) | \(M=\sqrt{\frac{4\mu }{p^3}}(t-t_0)\) | \(M=\sqrt{\frac{\mu (e^2-1)^3}{p^3}}(t-t_0)\) |
where \(\mu \) is the gravitational parameter of the central body and \(t_0\) is the time of passage through periapsis (Montenbruck and Pfleger 1994). The value M, called mean anomaly, is a linear function of the time. E, D and H are the elliptic, parabolic and hyperbolic eccentric anomalies respectively, which are functions of the time which are related to M through Kepler’s equation (third line of formulas in the table above). The eccentric anomalies are needed to compute the coordinates x and y, hence the need for a method to solve Kepler’s equation.
In the case of circular orbits (\(e=0\)), the equation for the eccentric anomaly reduces to \(E=M\), and in the case of parabolic trajectories (\(e=1\)), the cubic equation for D can be solved exactly:
Therefore, the real problem resides on the cases \(0<e<1\) and \(e>1\), where the equation involves a combination of algebraic and trascendental functions.
For simplicity, the hyperbolic Kepler’s equation is usually written as \(\sinh (H)-gH=L\), where \(g=\frac{1}{e}\in (0,1)\) and \(L=\frac{M}{e}\in (-\infty ,\infty )\). Since the left side of the equation is an odd function of H, it is enough to consider \(L\in [0,\infty )\). Moreover, assuming that H can be found, the formula for the y coordinate requires \(\sinh (H)\), so it makes more sense to directly find \(S=\sinh (H)\) solving
Once that is done, we can compute y as \(y=\frac{p}{\sqrt{e^2-1}}S\) and, since \(\cosh (H)=\sqrt{1+\sinh ^2(H)}\), x as \(x=\frac{p}{1-e^2}(\sqrt{1+S^2}-e)\).
While there are many articles discussing solutions for the elliptic case (see Avendano et al. 2014; Colwell 1993; Danby and Burkardt 1983; Montenbruck and Pfleger 1994; Ng 1979; Odell and Gooding 1986; Taff and Brennan 1989, among others), the hyperbolic case has received less attention. Prussing (1977) and Serafin (1986) gave upper bounds for the actual solution of the hyperbolic Kepler’s equation, which can be used as starters for Newton’s method since \(f_{g,L}''\ge 0\). Gooding and Odell (1988) solved the hyperbolic Kepler’s equation by using Newton’s method starting from a well-tuned formula depending on the parameters g and L. Their approach gives a relative accuracy of \(10^{-20}\) with only two iterations. Although their starter is a single formula that works on the entire region \((0,1)\times [0,\infty )\), it is too complicated to provide an efficient implementation.
In contrast, Fukushima (1997) focused on the efficiency and simplicity of the starter rather than trying to find a universal formula, and produced a starter that is defined by different formulas, each valid in a stripe-like region. He showed that his starter converges under Newton’s method, but not necessarily at quadratic speed.
More recently, Farnocchia et al. (2013), gave a method to avoid numerical issues that sometimes arise when applying Newton’s method in near parabolic orbits. They used instead a non-singular iterative technique that avoids rounding off problems.
Our approach to solve Eq. (1.1) is to use Newton’s method starting from a value \(\widetilde{S}(g,L)\) that is close enough to the actual solution S(g, L) to guarantee quadratic convergence speed, i.e. if \(S_n\) denotes the value obtained after n iterations, then \(|S_n-S|\le 0.5^{2^n-1}|\widetilde{S}-S|\). We use a simple criterion, Smale’s \(\alpha \)-test (Smale 1986) with the constant \(\alpha _0\) improved by Wang and Han (1989), to decide whether a starter gives the claimed convergence rate. Values that satisfy the test are called approximate zeros.
Definition 1.1
(Smale’s \(\alpha \)-test) Let \(f:(a,b)\subseteq \mathbb {R}\rightarrow \mathbb {R}\) be an infinitely differentiable function. The value \(z\in (a,b)\) is an approximate zero of f if it satisfies the following condition
where
and \(\alpha _0=3-2\sqrt{2}\approx 0.1715728\).
In an earlier paper (Avendano et al. 2014), we obtained the following simple approximate zero \(\widetilde{E}(e,M)\) for the elliptic case.
Theorem 1.2
The starter
is an approximate zero of \(E-e\sin (E)-M\) for all \(e\in [0,1)\) and \(M\in [0,\pi ]\).
Our main result in this paper, proven in Sect. 2, is a starter for the hyperbolic Kepler’s equation that is a piecewise-defined function in eight stripe-like regions. In seven of them, we use only linear expressions in g and L, while in the last one we need a cubic and a square root. Trying to construct a starter with only linear expressions is impractical because, as we approach the corner, the stripes become thinner and therefore a large number of them will be needed. Besides, numerical evidence suggests that it is impossible to cover a neighborhood of the corner \(g=1\), \(L=0\) with only linear expressions.
Theorem 1.3
The starter
is an approximate zero of \(f_{g,L}\) for all \(g\in (0,1)\) and \(L\in [0,\infty )\) (Fig. 1).
Remark 1.4
The exact solution \(\widetilde{S}\) of the cubic equation \((1-g)\widetilde{S}+g\frac{\widetilde{S}^3}{6} =L\) is given by
which can also be written as
where \(A=\root 3 \of {\frac{3L}{g}+\sqrt{\frac{9L^2}{g^2}+\frac{8(1-g)^3}{g^3}}} \).
For any bounded region of \((0,1) \times [0,\infty )\) that excludes a neighborhood of \(g=1, L=0\), we give in Sect. 3 a construction of another piecewise-defined starter using only constants.
Theorem 1.5
For any \(0<\varepsilon <\frac{1}{4}\) and \(L_{\max }>\varepsilon '= \frac{\sqrt{3}\alpha _0\varepsilon ^\frac{3}{2}}{(1-\varepsilon )^\frac{1}{2}}\), there is a piecewise constant function \(\widetilde{S}\) defined in \(R=(0,1)\times [0,L_{\max }] \setminus (1-\varepsilon ,1) \times [0,\varepsilon ']\) that is an approximate zero of \(f_{g,L}\).
The previous result is important for building look-up table solutions of the hyperbolic Kepler’s equation. On the one hand, the starter obtained in this way is valid only on a bounded region of space, since L must be bounded above by a predetermined value \(L_{\max }\). On the other hand, once the look-up table has been built, the starter can be retrieved immediately for any point within the region. The starter of Theorem 1.3 is better if the region is unknown in advance or if we are interested in solving the equation only a few times.
The method of Theorem 1.5 cannot be extended indefinitely to include a neighborhood of the corner, as shown in the following result.
Theorem 1.6
Let \(U \subseteq (0,1) \times [0,\infty )\) be an open set such that \(\bar{U} \supseteq \{ 1 \} \times [0,\varepsilon ]\) for some \(\varepsilon >0\). For any constant starter \(S_0\ge 0\), there exists a point \((g_0,L_0) \in U\) such that \(S_0\) is not an approximate zero of \(f_{g_0,L_0}\).
Finally, in Sect. 4, we include a comparison of the CPU timings and number of Newton’s method iterations of the starter of Theorem 1.3 with the ones given by Gooding and Odell (1988) and Fukushima (1997).
2 Starters for the hyperbolic Kepler’s equation
The aim of this section is to prove Theorem 1.3. To do that, we first prove some necessary technical results. Then we show in Theorem 2.4 that the solution of the cubic equation \((1-g)\widetilde{S}+g\frac{\widetilde{S}^3}{6} =L\) is an approximate zero when \(0\le L\le 1-\frac{5}{6} g\). Finally, we study in Theorem 2.5 the family of starters \(\widetilde{S}=L+ag\), and in Corollary 2.6 those with \(a \in \{0.91, 1.02, 1.16, 1.33, 1.56, 1.90, 2.30 \}\).
Lemma 2.1
For all \(k\ge 1\), we have
where \(P_k (x)\) are polynomials of degree \(k-1\). The monomials of \(P_k(x)\) have all even or odd degree:
The leading coefficient is \((-1)^{k-1} (k-1)!\), the coefficients alternate signs, and the sum of the absolute value of all the coefficients is \(||P_k (x) ||_1 =1\cdot 3 \cdot 5\cdots (2k-3)=(2k-3)!!\) The independent term satisfies
Proof
We proceed by induction. Note that \({{\mathrm{arcsinh}}}'(x)=(1+x^2)^{-\frac{1}{2}}\), so the Lemma holds for \(k=1\) by setting \(P_1(x)=1\). Assume now that the Lemma is true for some \(k\ge 1\). Differentiating the k-th derivative, we get
so we define \(P_{k+1}(x)=(1-2k)xP_k(x)+(1+x^2)P'_k(x)\). This shows that \(P_{k+1}(x)\) is a polynomial of degree at most k. Moreover, if \(P_k(x)\) has only terms of odd (or even) degree, then \(P_{k+1}(x)\) has only terms of even (or odd) degree, respectively. A more careful study of the leading coefficient of \(P_{k+1}(x)\) from the recurrence gives
showing that the degree of \(P_{k+1}(x)\) is k and that the leading coefficient of \(P_{k+1}(x)\) is \((-1)^kk!\), as we needed for the inductive step. The other coefficients of \(P_{k+1}(x)\) can be also found by using the recurrence:
for \(r=1,\ldots ,\left[ \frac{k}{2}\right] \). Since \(-k-2r<0\), \(k-2r+1>0\), and the signs of \({c_{k-2r-1}^{(k)}}\) and \({c_{k-2r+1}^{(k)}}\) are different, we have \({\mathrm{sgn}(c_{k-2r}^{(k+1)})=\mathrm{sgn}(c_{k-2r+1}^{(k)})}\), which alternate by the inductive hypothesis. Finally, for a polynomial with coefficients of alternating signs and all even (or odd) exponents, we have \(||P_{k+1}(x)||_1=|P_{k+1}(i)|=|(1-2k)iP_k(i)|=(2k-1)||P_k(x)||_1=(2k-1)!!\), where \( i=\sqrt{-1}\) is the imaginary unit.
According to Jeffrey and Dai (2008), the power series expansion of \({{\mathrm{arcsinh}}}(x)\) at \(x=0\) is
so \(P_{2n}(0)=0\), and \(P_{2n+1}(0)=\frac{(-1)^n(2n-1)!!(2n+1)!}{(2n+1)(2n)!!}=(-1)^n(2n-1)!!^2\). \(\square \)
Lemma 2.2
Given a real number \(x\ne 0\) and an integer \(n\ge 2\), then
Proof
Assume first that \(|x|\le n\). Then \(\max \{ 1,\left| \frac{x}{n} \right| ^{\frac{1}{n-1}} \}=1\). Moreover,
and
Therefore,
Now consider the case \(|x| >n\). Then \(\max \left\{ 1, \left| \frac{x}{n} \right| ^{\frac{1}{n-1}} \right\} =\left| \frac{x}{n} \right| ^{\frac{1}{n-1}}\). Moreover,
since the supremum is greater than or equal to the first term, and
since the sequence \({\left| \frac{x}{n} \right| ^{\frac{1}{k-1}}}\) is decreasing with respect to k. Therefore,
which concludes the proof. \(\square \)
The previous two lemmas give us the following upper bound for \(\gamma (f_{g,L},\widetilde{S})\).
Lemma 2.3
For all \(n\ge 2\), and \((g,L) \in (0,1) \times [0,\infty )\),
where
Proof
By Lemma 2.1,
Again by Lemma 2.1,
By Lemma 2.2,
\(\square \)
Theorem 2.4
The exact solution \(\widetilde{S}\) of the cubic equation \((1-g)\widetilde{S}+g\frac{\widetilde{S}^3}{6} =L\) is an approximate zero of \(f_{g,L}\) in the region \(R_1\), where
Proof
For any point \((g,L) \in R_1\), the solution \(\widetilde{S}(g,L)\) of the cubic equation satisfies \(\widetilde{S} \in [0,1]\). Therefore,
Using the previous inequality in the definition of \(\beta (f_{g,L},\widetilde{S})\) and the fact that \(g<1\), we obtain
On the other hand, it follows from Lemma 2.3 for \(n=3\), \(g<1\) and \(|\widetilde{S}| \le 1\) that
since \(\frac{\widetilde{S}}{2 \left( \sqrt{1+\widetilde{S}^2}-1 \right) } \le \frac{2}{\sqrt{3} \left( \sqrt{1+\widetilde{S}^2}-1\right) ^{\frac{1}{2}} }\).
Multiplying the inequalities for \(\beta (f_{g,L},\widetilde{S})\) and \(\gamma (f_{g,L},\widetilde{S})\) together, we have that the \(\alpha \)-test is satisfied if
A standard analytic study of the previous one-variable function shows that it is bounded above by \(\alpha _0\) in the interval \(\widetilde{S} \in [0,1]\). \(\square \)
Theorem 2.5
The starter \(\widetilde{S}(g,L)=L+a \,g\) is an approximate zero of \(f_{g,L}\) in the stripe
if \(\widetilde{S}_{\min }(a),\widetilde{S}_{\max }(a)\) are chosen such that
for all \(x\in [\widetilde{S}_{\min }(a),\widetilde{S}_{\max }(a)]\).
Proof
By definition, we have
By Lemma 2.3 with \(n=2\),
Therefore, we obtain that
which is less than \(\alpha _0\) in \(R_2(a)\) when \(\widetilde{S} \in [\widetilde{S}_{\min }(a),\widetilde{S}_{\max }(a)]\). \(\square \)
Corollary 2.6
The starter \(\widetilde{S}(g,L)=L+a \,g\) is an approximate zero of \(f_{g,L}\) in the stripe
for the values of a, \(\widetilde{S}_{\min }(a)\), \(\widetilde{S}_{\max }(a)\) given in the following table:
a | \(\widetilde{S}_{\min }(a)\) | \(\widetilde{S}_{\max }(a)\) |
---|---|---|
0.91 | 0.99 | 1.12 |
1.02 | 1.12 | 1.32 |
1.16 | 1.32 | 1.60 |
1.33 | 1.59 | 2.01 |
1.56 | 2.00 | 2.74 |
1.90 | 2.73 | 4.00 |
2.30 | 4.00 | \(\infty \) |
Proof (of Theorem 1.3)
By Theorem 2.4, \(\widetilde{S}(g,L)\) is an approximate zero of \(f_{g,L}\) when \(0\le L \le 1-\frac{5}{6} g\). For the remaining cases, we use Corollary 2.6. It is easy to verify that for each \(a \in \{ 0.91, 1.02, 1.16, 1.33, 1.56, 1.90, 2.30\}\), the region where \(\widetilde{S}\) is defined as \(L+a \, g\) is contained in \(R_2(a)\). \(\square \)
3 An analytical study of piecewise constant starters
Theorem 3.1
The constant starter \(\widetilde{S}(g,L)=S_0 >0\) is an approximate zero of \(f_{g,L}\) in the stripe
where
The constant starter \(\widetilde{S}(g,L)=0\) is an approximate zero of \(f_{g,L}\) in the region
Proof
Consider first the case \(S_0>0\). By Definition 1.1 and the fact that \(g<1\),
On the other hand, it follows from Lemma 2.3 with \(n=3\) that
We will now distinguish between \(S_0\ge 1\) and \(0 < S_0 < 1\). In the first case, \(\gamma (f_{g,L},\widetilde{S}) \le \frac{2S_0}{1+S_0^2}\), so the \(\alpha \)-test follows from
which is equivalent to
When \(0 < S_0 < 1\), \(\gamma (f_{g,L},\widetilde{S}) \le \frac{2}{1+S_0^2} \max \left\{ 1,\left( \frac{1}{3(\sqrt{1+S_0^2}-1)}\right) ^{\frac{1}{2}} \right\} \) so the \(\alpha \)-test is satisfied if
which is equivalent to
Therefore, the \(\alpha \)-test is satisfied in the whole region \(R_3 (S_0)\).
When \(\widetilde{S}(g,L)=S_0=0\), we have
Therefore, the \(\alpha \)-test is satisfied if
which is equivalent to \((g,L) \in R_3(0)\). \(\square \)
Remark 3.2
The region \(R_3(0)\) of Theorem 3.1 can be written as
Lemma 3.3
The function \(\varDelta (S_0)\) of Theorem 3.1 is strictly increasing for \(S_0>0\) (Fig. 2).
Proof
When \(0<S_0 \le \frac{\sqrt{7}}{3}\), then
When \(S_0 > \frac{\sqrt{7}}{3}\), then
Since both expressions are the product of constant and strictly increasing functions and \(\varDelta (S_0)\) is continuous at \(S_0=\frac{\sqrt{7}}{3}\), then \(\varDelta (S_0)\) is strictly increasing.
Proof (of Theorem 1.5)
We decompose the region of the theorem as
From Remark 3.2, we have that \(A \subseteq R_3 (0)\).
Define the sequence \(S_0=\varepsilon '\) and \(S_{i+1}=S_i +2\varDelta (S_i)\), for \(i\ge 0\), where \(\varDelta (S_i)\) is the function of Theorem 3.1. Since \(\varDelta \) is a positive and strictly increasing function by Lemma 3.3, the sequence \(S_i\) is strictly increasing and satisfies \(S_i \ge S_0+2 i \varDelta (S_0) \longrightarrow \infty \) as \(i \rightarrow \infty \).
Since \(\displaystyle {\lim _{x \rightarrow \infty }} (x-{{\mathrm{arcsinh}}}(x))=\infty \), there exists a large enough N for which \(S_N-{{\mathrm{arcsinh}}}(S_N) >L_{\max }\). Once we prove that \(\displaystyle {B \subseteq C=\bigcup _{i=0}^N R_3 (S_i)}\), with \(R_3(S_i)\) as in Theorem 3.1, we will have that \(A \cup B\) is covered by the regions where the starters \(0,S_0,\ldots ,S_N\) are approximate zeros of \(f_{g,L}\). In particular, this will provide the piecewise constant starter \(\widetilde{S}\) that we looked for.
It only remains to prove that \(B \subseteq C\). Note first that \(R_3(S_i)\) are stripes whose union is
since the lower boundary of \(R_3(S_{i+1})\) is a segment that is always below the upper boundary of \(R_3(S_i)\). Indeed, it is enough to see that the endpoints of the first segment, \((0,S_{i+1}-\varDelta (S_{i+1}))\) and \((1,S_{i+1}-\varDelta (S_{i+1})-{{\mathrm{arcsinh}}}(S_{i+1}))\), are below the endpoints of the second segment, \((0,S_i+\varDelta (S_i))\) and \((1,S_i+\varDelta (S_i)-{{\mathrm{arcsinh}}}(S_i))\), respectively:
which is true because \(\varDelta \) and \({{\mathrm{arcsinh}}}\) are both strictly increasing.
On the other hand, the vertices of the rectangle B are in \(\bar{C}\), which means that \(B\subseteq C\). Indeed, it follows from \({{\mathrm{arcsinh}}}(\varepsilon '),{{\mathrm{arcsinh}}}(S_N),\varDelta (\varepsilon '),\varDelta (S_N)>0\); \(\varepsilon '<L_{\max }\) by hypothesis and \(S_N-{{\mathrm{arcsinh}}}(S_N)>L_{\max }\) by construction, that
This implies
concluding the proof. \(\square \)
Proof (of Theorem 1.6)
We proceed by contradiction, i.e. we assume that \(S_0\) is an approximate zero of \(f_{g,L}\) for all \((g,L)\in U\). By definition, this means that
for all \((g,L)\in U\). We will derive a contradiction from (3.1) for all \(S_0\ge 0\).
Case \(S_0=0\). Inequality (3.1) is equivalent to
Using Lemma 2.1, we obtain
This implies that \(L\le \frac{\sqrt{6}\alpha _0(1-g)^\frac{3}{2}}{g^\frac{1}{2}}\) in \(\bar{U}\). Therefore \(L\le 0\) in \(\bar{U}\cap \{g=1\}=\{1\}\times [0,\varepsilon ]\), which is impossible.
For the rest of the cases, we take limit as \(g \rightarrow 1^-\) and then limit as \(L \rightarrow 0^+\), obtaining
Case \(0<S_0<1.598\). Inequality (3.2) implies that
because \((t+1)^2/t \ge 4\) for all \(t>0\). Therefore, \(h(S_0)={{\mathrm{arcsinh}}}(S_0) -S_0+\frac{\alpha _0S_0^3}{2} \ge 0\), which is false in the interval (0, 1.598) because h has only one critical point there, which is a minimum, and \(h(0),h(1.598)\le 0\).
Case \(1.598\le S_0<3.1\). Inequality (3.2) is equivalent to
Since \(\frac{\left( \sqrt{1+S_0^2}+1\right) ^\frac{3}{2} \sqrt{2S_0^2-1}}{S_0 \sqrt{1+S_0^2}}\) is increasing in this interval, then
Therefore,
and thus \(S_0-{{\mathrm{arcsinh}}}(S_0)-0.13S_0^2<0\). This is false since the function \(h(S_0)=S_0-{{\mathrm{arcsinh}}}(S_0)-0.13S_0^2\) has an absolute minimum at \(S_0=3.1\) and \(h(3.1)>0\).
Case \(3.1\le S_0<9.62\). Inequality (3.2) is equivalent to
On the one hand,
since the expression on the left is decreasing. On the other hand, \(\frac{S_0-{{\mathrm{arcsinh}}}(S_0)}{S_0^\frac{4}{3}} > 0.24\) because \(h(S_0)=S_0-{{\mathrm{arcsinh}}}(S_0)-0.24S_0^\frac{4}{3}\) is increasing in the interval, so \(h(S_0)\ge h(3.1)>0\).
Substituting these inequalities in Eq. (3.3), we obtain
which is a contradiction.
Case \(S_0\ge 9.62\). Using that \(\frac{S_0-{{\mathrm{arcsinh}}}(S_0)}{\sqrt{1+S_0^2}}\ge \frac{1}{2}\) and that \(\sqrt{1+S_0^2}-1\le S_0\) in the interval, we obtain from Eq. (3.2) that
This implies that, for all \(k\ge 2\),
By Lemma 2.1, the leading term of \(P_k(S_0)\) is \(\pm (k-1)!S_0^{k-1}\), its coefficients add up to a maximum of \((2k-3)!!\) and it has no term of degree \(k-2\), so
which is equivalent to
This implies that
The degree 3 polynomial \(h_k\) has two local extrema: a local maximum at \(S_0=0\) where \(h_k(0)=2^{k-1}>0\) and a local minimum at \(S_0=r_{\min }(k)=\frac{2}{3k(2\alpha _0)^{k-1}}\). Since \(h_k(r_{\min }(k))<0\) if and only if \(k\ge 5\), \(h_k\) has two positive roots when \(k\ge 5\), namely \(r_{\text {left}}(k)\) and \(r_{\text {right}}(k)\). Therefore, \(h_k(S_0)\ge 0\) is equivalent to \(S_0\in [0,r_{\text {left}}(k)]\cup [r_{\text {right}}(k),\infty )\) when \(k\ge 5\).
We will obtain a contradiction by showing that \(S_0\in (r_{\text {left}}(k),r_{\text {right}}(k))\) for some \(k\ge 5\). This is necessarily true since \(\cup _{k\ge 5}(r_{\text {left}}(k),r_{\text {right}}(k))= (r_{\text {left}}(5),\infty ) \supseteq (r_{\min }(5),\infty ) \supseteq [9.62,\infty )\ni S_0\).
It only remains to be proven that \(\cup _{k\ge 5}(r_{\text {left}}(k),r_{\text {right}}(k))=(r_{\text {left}}(5),\infty )\). This follows from \(r_{\text {right}}(k)>r_{\min }(k)=\frac{2}{3k(2\alpha _0)^{k-1}}\longrightarrow \infty \) as \(k\rightarrow \infty \), and the fact that \(r_{\text {left}}(k+1)<r_{\text {right}}(k)\) for all \(k\ge 5\). Indeed, since \(h_{k+1}(r_{\min }(k))<0\) and \(r_{\min }(k)<r_{\text {right}}(k)\), then \(r_{\text {left}}(k+1)<r_{\min }(k)<r_{\text {right}}(k)\).
4 Comparison with existing methods
In this section, we compare the accuracy and computational complexity (CPU time) of our starter given in Theorem 1.3 with the starters proposed by Gooding and Odell (1988) and Fukushima (1997). The more recent method presented in Farnocchia et al. (2013) cannot be directly compared with ours, since their approach is significantly different.
For the comparison, we follow the criterion of Table II of Fukushima (1997). We evaluate the starters in \(10^4\times 10^4\) equally-spaced grid points \((g,\log _{10}L)\) with \(0<g<1\) and \(-3\le \log _{10}L\le 8\). At each point, we record the CPU time and the number of Newton’s iterations needed to reach a solution that satisfies \(|f_{g,L}(S)|<5 \times 10^{-15}(1+L)\). The factor \(1+L\) was introduced by Fukushima to account for both relative and absolute errors. Table 1 shows the average CPU time (in \(10^{-6}\) s) and number of steps needed.
Table 1 shows that our proposed starter is faster than the other two. Moreover, the fact that the average number of Newton iterations is about 1.77, shows that our method requires less than 2 iterations in most cases. Although Gooding and Odell’s starter requires fewer steps on average, it is slower that ours due to its complexity.
5 Conclusions
We provided in Theorem 1.3 a very simple starter \(\widetilde{S}(g,L)\) for the hyperbolic Kepler’s equation which uses a single addition and multiplication when \(L>1-\frac{5}{6}g\) and a few arithmetic operations, a cubic and square root otherwise. Our starter can be implemented efficiently on modern computers and has guaranteed quadratic convergence speed from the first iteration, so the relative error becomes negligible after just a few iterations for any value of \(g\in (0,1)\) and \(L\in [0,\infty )\).
Indeed, on a modern computer it takes \(0.33\,\mu \)s on average to solve an instance of the equation (evaluated over a big grid of different possible inputs) and in most cases it requires at most two iterations of the Newton method to get an error of order \(5\times 10^{-15}\), as defined in Sect. 4. We have done this test on an Intel Core i5-2410M CPU @ 2.30 GHz\(\times \) 4 machine running Ubuntu \(14.04LTS\;64\)- bits.
If an even simpler starter is necessary, we showed in Theorem 1.5 that it is possible to produce a piecewise constant starter in any bounded region of \((0,1)\times [0,\infty )\), excluding a neighborhood of the corner \(g=1,L=0\). This starter also has quadratic convergence rate and its evaluation requires no operations, so it can be efficiently implemented as a look-up table.
References
Avendano, M., Martín-Molina, V., Ortigas-Galindo, J.: Solving Kepler’s equation via Smale’s \(\alpha \)-theory. Celest. Mech. Dyn. Astron. 119, 27–44 (2014)
Colwell, P.: Solving Kepler’s Equation Over Three Centuries. Willmann-Bell, Richmond (1993)
Danby, J.M.A., Burkardt, T.M.: The solution of Kepler’s equation I. Celest. Mech. 31, 95–107 (1983)
Farnocchia, D., Bracali Cioci, D., Milani, A.: Robust resolution of Kepler’s equation in all eccentricity regimes. Celest. Mech. Dyn. Astron. 116, 21–34 (2013)
Fukushima, T.: A method solving Kepler’s equation for hyperbolic case. Celest. Mech. Dyn. Astron. 68, 121–137 (1997)
Gooding, R.H., Odell, A.W.: The hyperbolic Kepler equation (and the elliptic equation revisited). Celest. Mech. 44, 267–282 (1988)
Jeffrey, A., Dai, H.H.: Handbook of mathematical formulas and integrals, 4th edn. Elsevier, Amsterdam (2008)
Montenbruck, O., Pfleger, T.: Astronomy on the Personal Computer, 2nd edn. Springer, Berlin (1994)
Ng, E.W.: A general algorithm for the solution of Kepler’s equation for elliptic orbits. Celest. Mech. 10, 243–249 (1979)
Odell, A.W., Gooding, R.H.: Procedures for solving Kepler’s equation. Celest. Mech. 38, 307–334 (1986)
Prussing, J.E.: Bounds on the solution to Kepler’s problem. J. Astronaut. Sci. 25, 123–128 (1977)
Serafin, R.A.: Bounds on the solution to Kepler’s equation. Celest. Mech. 38, 111–121 (1986)
Smale, S.: Newton’s method estimates from data at one point. In: Ewing, R., Gross, K., Martin, C. (eds.) The Merging of Disciplines: New Directions in Pure, Applied, and Computational Mathematics. Springer, Berlin (1986)
Taff, L.G., Brennan, T.A.: On solving Kepler’s equation. Celest. Mech. Dyn. Astron. 46, 163–176 (1989)
Wang, X., Han, D.: On dominating sequence method in the point estimate and Smale’s Theorem. Scientia Sinica Ser. A, 905–913 (1989)
Acknowledgments
The first author is partially supported by the MINECO grant ESP2013-44217-R, the second author by the MINECO Grant MTM2011-22621 and the FQM-327 group (Junta de Andalucía, Spain).
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Avendano, M., Martín-Molina, V. & Ortigas-Galindo, J. Approximate solutions of the hyperbolic Kepler equation. Celest Mech Dyn Astr 123, 435–451 (2015). https://doi.org/10.1007/s10569-015-9645-0
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10569-015-9645-0