Abstract
The Kantorovich theory plays an important role in the study of nonlinear equations. It is used to establish the existence of a solution for an equation defined in an abstract space. The solution is usually determined by using an iterative process such as Newton’s or its variants. A plethora of convergence results are available based mainly on Lipschitz-like conditions on the derivatives, and the celebrated Kantorovich convergence criterion. But there are even simple real equations for which this criterion is not satisfied. Consequently, the applicability of the theory is limited. The question there arises: is it possible to extend this theory without adding convergence conditions? The answer is, Yes! This is the novelty and motivation for this paper. Other extensions include the determination of better information about the solution, i.e. its uniqueness ball; the ratio of quadratic convergence as well as more precise error analysis. The numerical section contains a Hammerstein-type nonlinear equation and other examples as applications.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Let \({\mathcal {U}}, {\mathcal {V}},\) be Banach spaces and \(L({\mathcal {U}},{\mathcal {V}}) \) stand for the space of all continuous linear operators mapping \({\mathcal {U}}\) into \({\mathcal {V}}.\) Consider, a differentiable mapping as per Fréchet \({\mathcal {L}}:\Omega \subseteq {\mathcal {U}}\longrightarrow {\mathcal {V}},\) and its corresponding nonlinear equation
with \(\Omega \) denoting a non-empty open set. The task of determining a solution \(x_*\in \Omega \) is very challenging but important since applications from numerous computational disciplines are brought in the form (1.1) (Argyros and Magréñan 2018; Argyros 2004a, b; Ezquerro et al. 2010; Ortega and Rheinboldt 1970; Verma 2019). The analytic form of \(x_*\) is rarely attainable. That is why mainly iterative processes are used to generate approximations to the solution \(x_*.\)
Among these processes, the most widely used is Newton’s and its variants. In particular, Newton’s process (NP) is defined by
There exists a plethora of results related to the study of NP (Argyros and Magréñan 2018; Argyros 2004a; Argyros and Hilout 2010). These studies are based on the theory inaugurated by Kantorovich and its variants (Argyros 2021, 2022, 2004a, b; Argyros and Magréñan 2018; Argyros and Hilout 2010; Dennis 1968; Ezquerro et al. 2010; Gragg and Tapia 1974; Hernandez 2001; Ezquerro and Hernandez 2018; Kantorovich and Akilov 1982; Ortega and Rheinboldt 1970; Potra and Pták 1980, 1984; Proinov 2010; Rheinboldt 1968; Tapia 1971).
The following conditions (A) are used in non-affine or affine invariant form.
Suppose:
(A1) \(\exists \) point \(x_0\in \Omega \) and parameter \( \lambda \ge 0:\) \({\mathcal {L}}'(x_0)^{-1}\in L({\mathcal {V}}, {\mathcal {U}})\) and
(A2) \(\exists \) parameter \(M_1 > 0:\) Lipschitz condition
holds \(\forall w_1\in \Omega \) and \(w_2\in \Omega .\)
(A3)
(A4) \(B[x_0, \rho ]\subset \Omega ,\) where parameter \(\rho > 0\) is given later.
Let us denote \(B[x_0, r]:=\{x\in \Omega :\Vert x-x_0\Vert \le r\}\) for \(r > 0.\) Set \(\rho =r_1=\frac{1-\sqrt{1-2M_1\lambda }}{M_1}.\)
There are many variants of Kantorovich’s convergence result for NP. One of those follows (Chen and Yamamoto 1989; Deuflhard 2004; Kantorovich and Akilov 1982).
Theorem 1.1
Under the conditions A for \(\rho =r_1;\) the NP is contained in \(B(x_0,r_1),\) convergent to a solution \(x_*\in B[x_0, r_1]\) of Eq. (1.1) and
where the scalar sequence \(\{t_n\}\) is given by
Moreover, the convergence is linear if \(\lambda =\frac{1}{2M_1}\) and quadratic if \(\lambda < \frac{1}{2M_1}.\) Furthermore, the solution is unique \(B[x_0, r_1]\) in the first case and in \(B(x_0, r_2)\) in the second, where \(r_2=\frac{1+\sqrt{1-2M_1\lambda }}{M_1}.\)
A plethora of studies has used conditions A (Argyros and Magréñan 2018; Argyros 2004a; Argyros and Hilout 2010).
Example 1.2
Consider cubic polynomial
Let \(\Omega =B(x_0, 1-\mu )\) for some parameter \(\mu \in (0, \frac{1}{2}).\) Choose \(x_0=1.\) Then, the conditions A are verified for \(\lambda =\frac{1-\mu }{3}\) and \(M_1=2(2-\mu ).\) It follows that the estimate
holds \(\forall \mu \in (0, \frac{1}{2}).\) That is condition (A3) is not satisfied. Therefore, the convergence is not assured by this theorem also used in Chen and Yamamoto (1989), Dennis (1968), Deuflhard (2004), Ezquerro et al. (2010), Gragg and Tapia (1974), Hernandez (2001), Ezquerro and Hernandez (2018), Kantorovich and Akilov (1982), Ortega and Rheinboldt (1970), Potra and Pták (1980, 1984), Proinov (2010), Rheinboldt (1968), Tapia (1971), Yamamoto (1987a, 1987b, 2000) and Zabrejko and Bnuen (1987). But the NP converges. Hence, clearly, there is a need to improve the results based on condition A, which is only sufficient but not necessary.
In this paper, several avenues are presented for achieving this goal. The idea is to replace the Lipschitz parameter \(M_1\) with smaller ones.
Consider the center Lipschitz condition
the set \(\Omega _1=B[x_0, \frac{1}{M_0}]\cap \Omega \) and the Lipschitz -2 condition
Notice that by the definition of the set \(\Omega _1\)
Then, the Lipschitz parameters are related by
and
Notice also since parameters \(M_0\) and M are specializations of parameter \(M_1,\, M_1=M_1(\Omega ),\, M_0=M_0(\Omega ), \) but \(M=M(\Omega _1),\) where by \(M_1(\Omega )\) we mean that the parameter \(M_1\) depends on the set \(\Omega .\) Therefore, no additional work is required to find \(M_0\) and M (see also the numerical examples).
Moreover, the ratio \(\frac{M_0}{M}\) can be very small (arbitrarily).
Example 1.3
Define scalar function
for \(x_0=0,\) where \(b_j,\, j=0,1,2,3\) are real parameters. It follows by this definition that for \(b_3\) sufficiently large and \(b_2\) sufficiently small, \(\frac{M_0}{M_1}\) can be small (arbitrarily), i.e., \(\frac{M_0}{M_1}\longrightarrow 0.\)
Other extensions involve tighter majorizing sequences for NP (see Sect. 2) and improved uniqueness results for solution \(x_*\) (Sect. 3). The applications appear in Sect. 4 followed by and the conclusions in Sect. 5.
2 Real sequences
Let \(K_0, M_0, K, M\) and \(\lambda \) be positive parameters. An important role in the study of NM is played by the majorizing sequence \(\{s_n\}\) defined for \(s_0=0,\, s_1=\lambda ,\) as
That is why some convergence results are listed for it in what follows in this section.
Lemma 2.1
Suppose conditions \(K_0\lambda <1\) and \(s_{n+1} < \frac{1}{M_0}\) hold for all \(n=1,2,\ldots \) Then, the following assertions hold
and there exists \(s_*\in [\lambda ,\frac{1}{M_0}]\) such that \(\lim _{n\longrightarrow \infty }s_n=s_*.\)
Proof
The definition of sequence \(\{s_n\}\) and the conditions of the Lemma imply the assertion and \(\lim _{n\longrightarrow \infty }s_n=s_*\in [\lambda , \frac{1}{M_0}].\) Notice that \(s_*\) is the least upper bound (unique) of the sequence \(\{s_n\}.\) \(\square \)
Next, criteria stronger than those in Lemma 2.1 are developed for the convergence of the sequence (2.1). However, these criteria are easier to verify than those of the Lemma 2.1.
Define parameter \(\gamma \) by
This parameter plays a role in the study of NP.
Suppose from now on that \(K_0\le M_0.\) Define the real quadratic polynomials \(q, q_1, q_2\) by
and
The discriminants \(D, D_1, D_2\) of these polynomials can be given as
and
respectively. It follows by the definition of \(\gamma , q_1\) and \(q_2\) that
since after multiplying the polynomial \(q_2\) by \(\frac{2M_0\gamma }{1-\gamma },\) we obtain the polynomial \(q_1\) i.e.,
That is polynomials \(q_1\) and \(q_2\) have the same roots. Denote by \(\frac{1}{2r_1}\) the unique positive root of polynomial q. This root is given explicitly by the quadratic formula and can be written as
Moreover, denote by \(\frac{1}{2r_2}\) the common positive root of polynomials \(q_1\) and \(q_2.\) This root can also be written as
Define parameter N by
Suppose
By the choice of parameters \(r_1, r_2\) polynomials \(q, q_1, q_2\) and condition (2.2), it follows that \(M_0 s_2 < 1,\) since \(q(\lambda ) < 0,\) \(K_0 \lambda <1,\) \(q_1(\lambda ) \le 0\) and \(q_2(\lambda )\le 0.\) Furthermore, the following estimate holds
where the parameter \(\gamma _0=\frac{M(s_2-s_1)}{2(1-M_0s_2)}.\) Indeed, the left hand side inequality reduces to \(q_1(\lambda ) \le 0\) and the right hand side to \(q_2(\lambda )\le 0.\) These assertions are true by the choice of \(\lambda .\)
Lemma 2.2
Under condition (2.2), sequence \(\{s_n\}\) satisfies
and is convergent to its least upper bound \(s_*\in (\lambda , \bar{s}_{**}]\) so that
Proof
Induction is used to show the estimate
It follows by the definition of roots \(\frac{1}{2r_1}, \frac{1}{2r_2}\) and polynomial \(g_1\) that estimate (2.7) holds for \(k=1.\) Using the Definition (2.1) of sequence \(\{s_n\}\) and parameter \(\gamma _0\)
Suppose that estimate (2.7) holds for \(k=1,2,\ldots , n-1.\) Then, similarly by (2.7) and the induction hypotheses, we obtain in turn
It follows by the definition (2.1) of sequence \(\{s_n\},\) estimate (2.3) and induction hypothesis (2.7) that
Then, the estimate (2.7) for \(k+1\) replacing k holds, if
or
or
or
where, the polynomial \(p_k:[0,1)\longrightarrow \mathbb {R}\) is defined by
There is a connection between consecutive polynomials:
It follows that
where
Notice that \(q_3(\gamma )=0\) by the definition of \(\gamma .\) Then, in particular
Define function \(p_\infty :[0,1)\longrightarrow \mathbb {R}\) by
By this definition and polynomials \(p_k\)
Consequently, assertion (2.8) holds if
which is true by the choice of parameter \(\frac{1}{2r_2}\) and polynomial \(p_2.\) The induction for assertion (2.7) is terminated leading to the conclusions. \(\square \)
Remark 2.3
The linear convergence of sequence \(\{s_n\}\) is shown under the condition (2.2). This condition provides the smallness of \(\lambda \) to force convergence. The quadratic convergence of sequence \(\{s_n\}\) can be shown if \(\lambda \) is chosen to be bounded above from by a smaller parameter than \(\frac{1}{2N}.\) Moreover, under condition (2.2) an upper bound on iterate \(s_k\) is obtained which can then be used in the proof to show quadratic convergence.
Lemma 2.4
Under condition (2.10) further suppose that for some \(\epsilon > 0,\, \beta =\frac{\epsilon }{\epsilon +1}\)
and
Then, the conclusions of Lemma 2.2 hold for sequence \(\{s_n\},\)
and
where \(b=\frac{M}{2}(1+\epsilon )\) and \(b\lambda <1.\)
Proof
Assertion (2.14) certainly holds if the following estimate is shown
The estimate (2.15) holds for \(k=1,\) since it is equivalent to \(M_0 \lambda \le \beta .\) But this is true by \(M_0\le 2N,\) condition (2.2) and inequality \(\frac{\epsilon M_0}{2(1+\epsilon )N}\le \beta .\)
Define polynomials \(g_n:[0, 1)\longrightarrow \mathbb {R}\) by
It follows from this definition that
Evidently, estimate (2.15) holds. Define function \(g_\infty :(0,1)\longrightarrow \mathbb {R}\) by
Hence, we get \(g_\infty (t)=\frac{(1+\epsilon )M_0\gamma _0(s_2-s_1)}{1-t}+(1+\epsilon )M_0(\lambda +s_2-s_1)-\epsilon .\) Evidently, the estimate
holds if instead
But this is identical to condition (2.11). The induction for the assertion (2.15). Then, it follows by estimate (2.15) and the definition of sequence \(\{s_n\}\) that assertion (2.13) holds. Using the definition of parameter b and estimate (2.13)
thus,
Notice that \(0< b\lambda <1\) by the condition (2.12). Hence, sequence \(\{s_k\}\) converges quadratically to \(t_*.\) \(\square \)
Remark 2.5
Condition (2.11) is left uncluttered. It can be expressed as a function of \(\lambda \) by
Suppose
It follows by the definition of polynomial \(\varphi \) and condition (2.17) that \(\varphi (0)=-\beta < 0.\) Moreover, \(\varphi (t)\longrightarrow \infty \) as \(t\longrightarrow r_3=\min \{\frac{1}{K_0}, \frac{1}{2r_2}\}.\) Hence, function \(\varphi _1\) has zeros in interval \((0, r_3)\) as a consequence of the intermediate value theorem. Let \(\lambda _0\) stand for the smallest such zero. Then, conditions (2.2), (2.11), and (2.12) are condensed as
If \(\frac{1}{2N_0}=\frac{\epsilon }{(1+\epsilon )M_0}\) or \(\frac{1}{2N_0}=\frac{2}{(1+\epsilon )M},\) then condition (2.18) should hold as a strict inequality.
3 Convergence of NP
The Lipschitz parameters are associated with operator \({\mathcal {L}}\) and its derivatives.
Suppose there exist parameters \(K_0> 0, K > 0\) such that
for \(x_1=x_0-{\mathcal {L}}'(x_0)^{-1}{\mathcal {L}}(x_0)\) and each \(\xi \in [0, 1]\) and
Conditions (A1), (1.4), (1.5), (3.1)–(3.3) and those of Lemma 2.1 or Lemma 2.2 are summarized by (H).
Next, under conditions H, we show the main convergence result for NP.
Theorem 3.1
Under conditions H sequence NP is convergent to a solution \(x_*\in B[x_0, t_*]\) of equation \({\mathcal {L}}(x)=0.\) Moreover, upper bounds
hold \(\forall \,\,i=0,1,2,\ldots \)
Proof
The assertions
and
are proven by induction \(\forall \,\,j=0,1,2,\ldots \) Using (A1)
Let \(u\in B[x_1, s_*-s_1].\) It follows by condition (A1)
so \(u\in B[x_1, s_*-s_1].\) That is assertions (3.5) and (3.6) hold if \(j=0.\) Assume these assertions hold if \(j=0,1,2,\ldots ,n.\) It follows for each \(\xi \in [0,1)\)
and
It follows by induction hypotheses, the Lemmas and conditions (3.1) and (1.1)
where \(\tilde{K}=\left\{ \begin{array}{cc} K_0,&{} j=0\\ M_0,&{} j=1,2,\ldots \end{array}\right. \) Hence, the inverse of linear operator \({\mathcal {L}}'(x_{j+1})\) exists. Notice that if \(j=0, K_0\) can be used, whereas if \(j=1,2,\ldots ,\) then \(M_0\) is utilized.
as a consequence of a lemma on linear operators that are invertible due to Banach’s perturbation lemma (Argyros and Magréñan 2018; Argyros 2004a, b; Argyros and Hilout 2010).
The identity can be given by NP
since
Then, using induction hypotheses, identity (3.8) and condition (1.5)
where \(\tilde{M}=\left\{ \begin{array}{cc} K,&{}j=0\\ M,&{}j=1,2,\ldots \end{array}\right. \)
It follows by NP, estimates (3.7), (3.9) and the definition (2.1) of sequence \(\{s_n\}\)
where \(\bar{K}=\left\{ \begin{array}{cc} K,&{}j=0\\ M,&{}j=1,2,\ldots \end{array}\right. \) and \(\bar{M}=\left\{ \begin{array}{cc} K_0,&{}j=0\\ M_0,&{}j=1,2,\ldots \end{array}\right. \)
Moreover, if \(v\in B[x_{j+2},s_*-s_{j+2}],\) it follows
Thus, the element \(v\in B[x_{j+1}, s_*-s_{j+1}]\) completing the induction for assertions (3.5) and (3.6). Notice that scalar majorizing sequence \(\{s_j\}\) is fundamental as convergent. Hence, the sequence \(\{x_j\}\) is also convergent to some \(x_*\in B[x_0, s_*].\) Furthermore, let \(j\longrightarrow \infty \) in estimate (3.9) to conclude \({\mathcal {L}}(x_*)=0.\) Finally, the proof of assertion (3.4) using estimate (3.5) as standard is omitted (Yamamoto 1987b). \(\square \)
Next, the uniqueness ball for the solution \(x_*\) is presented. Notice that not all conditions S are used.
Proposition 3.2
Under center-Lipschitz condition (1.4) further assume the existence of a solution \(p\in B(x_0, R)\subset \Omega \) of equation \({\mathcal {L}}(x)=0\) such that linear operator \({\mathcal {L}}'(p)\) is invertible for some \(R > 0;\) a parameter \(R_1 > R\) given by
Then, the element p solves uniquely equation \({\mathcal {L}}(x)=0\) in the set \(T=B(x_0, R_1)\cap \Omega .\)
Proof
Define linear operator \(Q=\int _0^1 {\mathcal {L}}'(\bar{p}+\xi (p-\bar{p}))\mathrm{{d}}\xi \) for some element \(\bar{p}\in T\) satisfying \(F(\bar{p})=0.\) By using the definition of parameter \(R_1\), set T and condition (1.4)
Then, the estimate (4.5) and the Banach lemma on linear operators (Argyros and Magréñan 2018; Argyros 2004a, b; Argyros and Hilout 2010) with inverses, imply the invertablility of linear operator Q. Moreover, by the approximation \(0= {\mathcal {L}}(p)-{\mathcal {L}}(\bar{p})=Q(p-\bar{p}),\) we deduce \(\bar{p}=p.\) \(\square \)
Remark 3.3
Notice that not all conditions of Theorem 3.1 are used in Proposition 3.2. But if they were, then, we can set \(p=x^*\) and \(R=s^*.\)
4 Special cases and Examples
It turns out that conditions of Theorem 3.1 reduce to the ones given by the earlier studies. But first we have the observations
Remark 4.1
Let us compare conditions H to conditions A:
It follows by these conditions that \(K_0\le K\le M_0.\) Consequently, replacing \(M_0\) or \(M_1\) by these tighter parameters gives previously mentioned benefits. Moreover, notice that parameters \(K_0, K, M_0, M\) are specializations of the originally used \(M_1.\) Hence, no additional cost is required in their computation.
-
(1)
The condition (A1) is common.
-
(2)
The condition (A2) always implies the conditions (3.1) and (3.2). However, the converse implication does not hold necessarily, unless \(K_0=K=M_1.\)
-
(3)
The new majorizing sequence \(\{s_n\}\) is tighter than the sequence \(\{t_n\}\) used by Kantorovich. In particular, under the conditions of the Theorem 1.1 and the Theorem 3.1, a simple inductive argument gives
$$\begin{aligned}{} & {} 0\le s_n\le t_n,\\{} & {} \quad 0\le s_{n+1}-s_n\le t_{n+1}-t_n \end{aligned}$$and
$$\begin{aligned} s_*\le \rho . \end{aligned}$$Notice also:
-
(4)
The conditions of Lemma 2.2 are stronger than those of the Lemma 2.1.
-
(5)
The conditions of the Kantorovich Theorem 1.1 imply
$$\begin{aligned} t_n < \frac{1}{M_1}. \end{aligned}$$This inequality implies the one by our Lemma 2.1 but not vice versa unless if \(M_0= M_1.\)
-
(6)
Next, a comparison between Lemma 2.2 and the corresponding one in Theorem 1.1 follows.
Case \(K_0=M_0=K=M.\)
(i) It follows by the definition of N that \(N=M\) and condition (2.2) reduces to
Furthermore, if \(M=M_1\) it reduces to the Kantorovich condition (A3) in the conditions A. But by estimate (1.8) it follows that if \(M < M_1\) then condition (A3) implies (4.1) but not vice versa. Hence, the new convergence criterion (4.1) weakens the Kantorovich criterion (A3) (see also the examples where \(M < M_1\)). (ii) The majorizing sequence reads
This sequence is more precise than \(\{t_n\}\) but not necessarily \(\{s_n\},\) unless if \(K=M\) and \(K_0=M_0.\)
(iii) The uniqueness ball is extended, since \(M_0\) is used for \(M_1\) in the formula. (see also Proposition 3.2).
Other specializations of the Lipschitz conditions give similar benefits (Table 1).
Example 4.2
The parameters using the example of the introduction are \(K_0=\frac{\mu +5}{3}, K=M_0=\frac{\mu +11}{6}.\) Moreover,\(\Omega _0=B(1, 1-\mu )\cap B(1, \frac{1}{M_0})=B(1, \frac{1}{M_0}).\) Set \(M=2(1+\frac{1}{3-\mu })\) \(M_0 < M_1\) and \( M < M_1\) for all \(\mu \in (0,0.5).\) Criterion (2.2) is then satisfied if \(\mu \in S_0:=[0.42, 0.5).\) Hence, the range of values for \(\mu \) for which NP converges is extended. Interval \(S_0\) can be enlarged if the condition of Lemma 2.1 is verified. Then, for \(\mu =0.4,\) we have the following \(\frac{1}{M_0}=0.5263,\)
Hence, the conditions of Lemma 2.1 hold. For \(\epsilon =0.63, \) we have
and
Thus, the conditions (2.11) and (2.12) hold, and the interval \(S_0\) is further enlarged.
Example 4.3
Let \({\mathcal {U}}={\mathcal {V}}=C[0,1]\) be the space of continuous real functions on the interval [0, 1]. The max-norm is used. Set \(\Omega =B[x_0,3].\) Define Hammerstein nonlinear integral operator \({\mathcal {L}}\) on \(\Omega \) as
where function \(y\in C[0,1],\) and G is a kernel related by Green’s function
It follows by this definition that \({\mathcal {L}}'\) is defined by
\(z\in C[0,1], w\in [0,1].\) Pick \(x_0(w)=y(w)=1.\) It then follows from (4.2)–(4.4) that \({\mathcal {L}}'(x_0)^{-1}\in L({\mathcal {V}}, {\mathcal {U}}),\)
and \(\Omega _0=B(x_0, 3)\cap B(x_0, 0.4167)=B(x_0, 0.4167),\) so \(M=1.5.\) Notice that \(M_0 < M_1\) and \(M < M_1.\) Set \(K_0=K=M_0.\) The Kantorovich convergence criterion (A3) is not satisfied, since \(2M_1\lambda =1.44 >1.\) Therefore convergence of NP is not assured. But our condition is satisfied, since \(2M\lambda =0.6 < 1.\)
Remark 4.4
-
(i)
Under conditions H, set \(p=x_*\) and \(R=s_*.\)
-
(ii)
Lipschitz condition (1.5) can be replaced by
$$\begin{aligned} \Vert {\mathcal {L}}'(x_0)^{-1}({\mathcal {L}}'(w_1)-{\mathcal {L}}'(w_2))\Vert \le d\Vert w_1-w_2\Vert , \end{aligned}$$(4.5)\(\forall \,\,w_1\in \Omega _0\) and \(w_2=w_1-{\mathcal {L}}'(w_1)^{-1}{\mathcal {L}}(w_1)\in \Omega _0.\) This, even smaller parameter d can replace M in the aforementioned results. The existence of iterate \(w_2\) is assured by (1.4).
-
(iii)
Another way to reduce Lipschitz constant M is given as follows: Suppose Lipschitz condition (1.5) is replaced by
$$\begin{aligned} \Vert {\mathcal {L}}'(x_0)^{-1}({\mathcal {L}}'(w_1)-{\mathcal {L}}'(w_2))\Vert \le d_0\Vert w_1-w_2\Vert , \end{aligned}$$(4.6)\(\forall w_1\in T_1\) and \(w_2=w_1-{\mathcal {L}}'(w_1)^{-1}{\mathcal {L}}(w_1)\in T_1,\) where \(T_1=B(x_1, \frac{1}{M_0}-\lambda )\) provided that \(\lambda \mu <1.\) Notice that \(d_0\le d\le M,\) since \(T_1\subset \Omega _0\subset \Omega _1.\) In the case of example 4.1, the parameters are \(d_0=\frac{5(4-\mu )^3+\mu (3-\mu )^3}{3(3-\mu )(4-\mu )^2}< d=\frac{6+2(3-\mu )(1+2\mu )}{3(3-\mu )} < M\,\, \forall \mu \in (0,0.5).\)
5 Conclusion
A new methodology extends the applicability of NP. The new results are finer than the earlier ones. Therefore, they can replace them. No additional conditions are used. The methodology is very general. Consequently, it can be applied to extend other procedures (Argyros 2004b; Chen and Yamamoto 1989; Deuflhard 2004; Yamamoto 1987b).
References
Argyros IK (2004a) A unifying local-semilocal convergence analysis and applications for two-point Newton-like methods in Banach space. J Math Anal Appl 228:374–397
Argyros IK (2004b) On the Newton–Kantorovich hypothesis for solving equations. J Comput Appl Math 169:315–332
Argyros IK (2021) Unified convergence criteria for iterative Banach space valued methods with applications. Mathematics 9(16):1942. https://doi.org/10.3390/math9161942
Argyros IK (2022) The theory and applications of iteration methods. Engineering series, 2nd edn. CRC Press, Taylor and Francis Group, Boca Raton
Argyros IK, Hilout S (2010) Extending the Newton–Kantorovich hypothesis for solving equations. J Comput Appl Math 234:2993–3006
Argyros IK, Magréñan AA (2018) A contemporary study of iterative schemes. Elsevier (Academic Press), New York
Chen X, Yamamoto T (1989) Convergence domain of certain iterative methods for solving nonlinear equations. Numer Funct Anal Optim 10:37–48
Dennis JE Jr (1968) On Newton-like methods. Numer Math 11:324–330
Deuflhard P (2004) Newton methods for nonlinear problems. Affine invariance and adaptive algorithms. Springer Series in Computational Mathematics, vol 35. Springer, Berlin
Ezquerro JA, Hernandez MA (2018) Newton’s scheme: an updated approach of Kantorovich’s theory. Cham, Geneva
Ezquerro JA, Gutiérrez JM, Hernandez MA, Romero N, Rubio MJ (2010) The Newton method: from Newton to Kantorovich (Spanish). Gac R Soc Mat Esp 13:53–76
Gragg WB, Tapia RA (1974) Optimal error bounds for the Newton–Kantorovich theorem. SIAM J Numer Anal 11:10–13
Hernandez MA (2001) A modification of the classical Kantorovich conditions for Newton’s method. J Comput Appl Math 137:201–205
Kantorovich LV, Akilov GP (1982) Functional analysis. Pergamon Press, Oxford
Ortega LM, Rheinboldt WC (1970) Iterative solution of nonlinear equations in several variables. Academic Press, New York
Potra FA, Pták V (1980) Sharp error bounds for Newton’s process. Numer Math 34:63–72
Potra FA, Pták V (1984) Nondiscrete induction and iterative processes. Research notes in mathematics, vol 103. Pitman (Advanced Publishing Program), Boston
Proinov PD (2010) New general convergence theory for iterative processes and its applications to Newton–Kantorovich type theorems. J Complex 26:3–42
Rheinboldt WC (1968) A unified convergence theory for a class of iterative processes. SIAM J Numer Anal 5:42–63
Tapia RA (1971) Classroom notes: the Kantorovich theorem for Newton’s method. Am Math Mon 78:389–392
Verma R (2019) New trends in fractional programming. Nova Science Publisher, New York
Yamamoto T (1987a) A method for finding sharp error bounds for Newton’s method under the Kantorovich assumptions. Numer Math 49:203–320
Yamamoto T (1987b) A convergence theorem for Newton-like methods in Banach spaces. Numer Math 51:545–557
Yamamoto T (2000) Historical developments in convergence analysis for Newton’s and Newton-like methods. J Comput Appl Math 124:1–23
Zabrejko PP, Bnuen DF (1987) The majorant method in the theory of Newton–Kantorovich approximations and the Pták error estimates. Numer Funct Anal Optim 9:671–684
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by Andreas Fischer.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Regmi, S., Argyros, I.K., George, S. et al. Extended Kantorovich theory for solving nonlinear equations with applications. Comp. Appl. Math. 42, 76 (2023). https://doi.org/10.1007/s40314-023-02203-2
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s40314-023-02203-2