1 Introduction
Let $E_1$ and $E_2$ denote complete normed spaces and $D\subset E_1$ be a nonempty, open, and convex set, and $G:D\subseteq E_1\longrightarrow E_2$ be a Fréchet differentiable operator. The determination of a locally unique solution $x^*\in D$ for the nonlinear equation of the form
is very challenging and of extreme importance. This is indeed the case, since many applications from diverse disciplines such as Mathematical: Biology; Chemistry; Ecology; Economics; Physics; Scientific Computing; and Engineering can be written in a form like (1.1) using Mathematical Modeling [Reference Argyros4–Reference Bonnans6, Reference Dontchev and Rockafellar8, Reference Nesterov and Nemirovskii16]. However, the analytical version of the solution $x^*$ is hard or expensive and it can be found only in special cases. That explains the reason why most practitioners and researchers develop iterative method generating a sequence approximating $x^*$ under certain conditions imposed on the initial data.
Newton’s method is defined for each $n=0,1,2,...$ by
There is a plethora of studies about Newton’s method [Reference Argyros3, Reference Argyros and George5, Reference Ortega and Rheinboldt18, Reference Proinov20]. Practical applications in convex programming can be found in [Reference Nesterov and Nemirovskii16]. Other applications can be found in [Reference Argyros4, Reference Bonnans6, Reference Magreńan and Argyros15, Reference Nocedal and Wright17, Reference Polyak19, Reference Traub and Wozniakowski24]. A usual hypothesis in such studies is some type of Lipchitz, Hölder, or majorant condition on $G'$ . Such hypotheses are important, since they allow some control on the derivative, which is important in both the local as well as the semi-local analyses of convergence of iterative methods. The local convergence analysis results are important, since they provide the degree of difficulty in choosing the starting points $x_0$ so that the sequence $\{x_n\}$ is convergent to $x^*$ .
In particular, we are motivated by the elegant work on the local analysis of convergence for Newton’s method in [Reference Ferreira9] (see also [Reference Ferreira and Svaiter10–Reference Ferreira and Svaiter13, Reference Wang and Ouyang25, Reference Zabrejko and Nguen27]) using the majorant principle and optimization considerations.
The new local analysis of convergence uses weaker majorant conditions resulting to the following advantages:
Novelty
-
(a 1) A larger radius of convergence. That allows a wider selection of starting points $x_0$ ensuring the convergence of the sequence $\{x_n\}$ to $x^*$ .
-
(a 2) Tighter upper error bounds on the distances $\|x_n-x^*\|$ . This way we use fewer iteration to achieve a predetermined error tolerance denoted by $\epsilon>0$ . That is, there exists N such that for each $n\geq N, \|x_n-x^*\|\leq \epsilon .$
-
(a 3) A larger than before domain is found containing $x^*$ as the only solution of the equation $G(x)=0$ .
It is worth noting that such advantages extend the number of classes of equations that can be solved using Newton’s method, the advantages $(a_1)-(a_3)$ are obtained under the same computational effort, since in practice the new and tighter majorant functions are special cases of the one used in [Reference Ferreira9]. The same advantages can be obtained in other studies, e.g., for solving generalized equations using Newton’s method [Reference Adly, Van Ngai and Nguyen1, Reference Araǵon Artacho, Belyakov, Dontchev and López2, Reference Dontchev and Rockafellar8, Reference Ferreira, Jean-Alexis, Pietrus and Silva11, Reference Ferreira and Svaiter13, Reference Robinson22] as long as they use the majorant conditions.
The rest of the article contains the following: The preliminaries in Section 2 followed by the properties of the majorant functions in Section 3. The local analysis of convergence for Newton’s method is developed in Section 4. Finally, the special cases and examples can be found in the concluding Section 5.
2 Preliminaries
Let $E_1$ and $E_2$ stand for complete normed spaces and $L(E_1,E_2)$ denote the space of continuous linear operators sending $E_1$ into $E_2$ . Moreover, define the open ball $U(z,\mu )$ and its closure $U[z,\mu ]$ , respectively, for $\mu>0$ as
and
A standard auxiliary result on the inverses of operator that are linear is useful.
Lemma 2.1 (Banach’s perturbation lemma) Assume that $T\in L(E_1, E_2)$ and satisfies $\|T-I\| <1$ , where $I:E_1 \longrightarrow E_1$ denotes the identity operator. Then, the following items are valid: the linear operator T is invertible and
Proof Set $A=I$ and $c=\|T-I\|$ in Lemma 1 of [Reference Smale, Ewing, Gross and Martin23, p. 189].
Some basic properties of convex functions are needed. More information about such functions can be found in [Reference Adly, Van Ngai and Nguyen1–Reference Argyros3, Reference Cibulka, Dontchev and Geoffroy7, Reference Hiriart-Verryty and Lemarechal14].
Lemma 2.2 Let $R>0$ be a given constant. Assume that $\psi : [0,R) \longrightarrow (-\infty , +\infty )$ is convex and differentiable. Then, the following items are valid:
-
(i)
$$ \begin{align*} \frac{\psi(s)-\psi(\theta s)}{s} \leq \psi '(s)(1-\theta) \nonumber \end{align*} $$for each $s \in (0,R)$ and $\theta \in [0,1];$ -
(ii)
$$ \begin{align*}\frac{\psi(u_1)-\psi(\theta u_1)}{u_1}\leq \frac{\psi(u_2)-\psi(\theta u_2)}{u_2} \nonumber \end{align*} $$for each $u_1,u_2 \in [0,R), u_1 <u_2$ and $\theta \in [0,1]$ .
Proof See Theorem 4.1.1 and Remark 4.1.2 in [Reference Hiriart-Verryty and Lemarechal14, p. 21].
We need a relationship between different types of majorant conditions. Let
Definition 2.1 A function $h_0:[0,R)\longrightarrow (-\infty , +\infty )$ which is twice continuously differentiable is said to be a center-majorant function for G on if for each
for some operator $M\in L(E_1, E_2)$ which is invertible, independent of y, and may or may not depend on $x^*$ (see also Remark 2.3(iv)).
-
(A1) The function $h_0'$ is convex and strictly increasing.
-
(A2) $h_0(0)=0$ and $h_0'(0)=-1.$
M is chosen to be any linear invertible operator satisfying this condition (see also Remark 2.3 and Section 5).
Let be such that
Definition 2.2 A function $h:[0,\rho )\longrightarrow (-\infty , +\infty )$ which is twice continuously differentiable is said to be a restricted-majorant function for G on $U(x^*,{\rho }),$ if for each $\theta \in [0,1],\,y\in U(x^*,\rho )$
Notice that h depends on the function $h_0.$
-
(A3) The function $h'$ is convex and strictly increasing.
-
(A4) $h(0)=0$ and $h'(0)=-1.$
Notice that the function $h_0$ depends on $x^*$ and whereas the function h on $x^*,\rho $ , and $h_0.$
Definition 2.3 A function $h_1:[0,R)\longrightarrow (-\infty , +\infty )$ which is twice continuously differentiable is said to be a majorant function for G on $U(x^*,{\rho }),$ if for each
-
(A3)’ The function $h_1'$ is convex and strictly increasing.
-
(A4)’ $h_1(0)=0$ and $h_1'(0)=-1.$
Remark 2.3 (i) It follows by these definitions that
(ii) Thus, the results in the literature using only $h_1$ (for $M=G'(x^*)$ ) (see [Reference Ferreira9]) can be replaced by the pair $(h_0,h)$ resulting to finer error distances, a larger convergence radius, and a more precise and larger uniqueness radius for the solution $x^*.$ These advantages are obtained under the same computational cost, since in practice the computation of the function $h_1$ requires that of $h_0$ and h as special cases.
(iii) A popular (but not the most flexible) choice for $M=G'(x^*).$ In this case, the solution is simple. However, assumptions allow the determination of a solution that is not necessarily simple.
(iv) The choice of the initial point can be chosen without the actual knowledge of $x^*.$ Suppose that the operator G satisfies the autonomous differential equation [Reference Argyros3, Reference Ortega and Rheinboldt18]:
where P is a continuous operator. By this definition, we obtain $G'(x^*)=P(G(x^*))=P(0),$ for any solution $x^*$ of the equation $G(x)=0.$ As an example, define ${G(x)=e^x-1}$ and choose $P(x)=x+1.$ Then, for the operator P satisfies $G'(x)=P(G(x)).$ Therefore, we can choose $M=G'(x^*)$ and $M=P(G(x^*))$ or $M=P(0)$ , which is known although $x^*$ is not known.
Definition 2.4 We define the parameters
and
Moreover, define the Newton iteration for solving the equation $h(s)=0$ given by
Define the radius
and the parameter
Next, the main local analysis convergence result for Newton’s method is stated.
Theorem 2.4 Assume that the conditions (A1)–(A5) are valid with the radius $\rho ^*$ as defined in (2.4). Then, if the starter $x_0\in U(x^*,\rho ^*)-\{x^*\},$ the following assertions hold:
(i) The scalar majorant sequence $\{s_n\}$ given by the formula (2.2) converges to zero, belongs in the interval $(0, \rho ^*),$ and the sequence $\{\frac {s_{n+1}}{s_n^2}\}$ is strictly decreasing.
(ii) The sequence $\{x_n\}\subset U(x^*,\rho ^*)$ converges Q-quadratically, to $x^*$ so that
for each $n=0,1,2,\ldots .$
(iii) Additionally, if for $\alpha \in (0,\rho ), \frac {h(\alpha )}{(\alpha h_0'(\alpha ))-1}=1,$ then $\rho ^*=\alpha $ is the largest convergence radius. Moreover, $x^* $ is the unique solution of equation (1.1) in the ball $U(x^*, c_4).$
3 The properties of the majorant functions
The parameters related to the majorant functions $h_0,h$ and the domain D are shown to exist and be positive in the auxiliary results that follow in this section. Moreover, the properties of the sequence $\{s_n\}$ which appears in Theorem 2.4 are investigated.
Lemma 3.1 The following items are valid:
and
for each $s\in (0,c_1).$
Proof By hypothesis $x^*\in D$ which is an open set. Thus, it follows that By the second condition in $(A_2), h_0'(0)=-1,$ there exists a parameter $\gamma>0$ so that ${h_0'(s)<0}$ for each $s\in (0,\gamma )$ leading us to deduce that $c_1>0.$ Moreover, by the condition $(A_2), h_0(0)=0$ and $h_0'(0)=-1.$ Then, there exists a parameter $\gamma>0$ so that $h_0(s)<0$ for each $s\in (0,\gamma ),$ so $c_3>0.$ By the conditions $(A_3),(A_4)$ and Lemma 2.2,
for each But if $s\in (0,c_1),$ then $h_0'(s)<0,$ showing (3.1) by (3.2.)
It follows by the condition $(A_1)$ and the definition of the parameter $c_1$ that the real function
exists in the interval $[0,c_1).$
Lemma 3.2 The following items are valid.
The function $\frac {|\psi _{h_0,h}(s)|}{s^2}$ is strictly increasing
and
for each $s\in (0,c_1).$
Proof The definition of the function $\psi _{h_0,h}, h_0(0)=0$ and $h_0'(s)<0$ for each $s\in [0,c_1)$ imply that
Thus, (3.4) holds, where we also used Lemma 2.2 to deduce that the function
for each $s\in (0,c_1),\theta \in (0,1)$ is strictly increasing, and positive as well as the function $s\rightarrow \frac {1}{|h_0'(s)|}$ is strictly increasing. This makes the right-hand side of (3.5) positive.
Lemma 3.3 The following items are valid:
and
for each $s\in (0,c_2).$
Proof By the last two lemmas,
and the function $\frac {|\psi _{h_0,h}(s)|}{s^2}$ is bounded close enough to zero. Thus, we can write
Hence, by the estimates (3.8) and (3.9), there exists a parameter $\gamma>0$ so that
Consequently, by the definition of the parameter $c_2,$ and the preceding double inequality, we get $c_2>0.$ Moreover, Lemma 3.2 gives that the function $s\rightarrow \frac {|\psi _{h_0,h}(s)|}{s^2}$ is strictly increasing for each $s\in (0,c_1).$ Furthermore, the definition of the parameter $c_2$ implies (3.7) for each $s\in (0,c_2).$
It follows by (3.4) that the scalar sequence $\{s_n\}$ can also be given as
for each $n=0,1,2,\ldots $ .
Lemma 3.4 The following items are valid: The scalar sequence $\{s_n\}$ given in (3.10) exists in the interval $(0,c_2)$ for each $n=0,1,2,\ldots $ ;
i.e., the sequence $\{s_n\}$ is strictly decreasing; the sequence $\{\frac {s_{n+1}}{s^2_n}\}$ is strictly decreasing,
and
Proof Lemma 3.3 and the definition of the sequence $\{s_n\}$ given in (3.10) imply by a simple inductive argument that $s_{m+1}=|\psi _{h_0,h}(s_m)|<s_m,$ where we also used the restriction the first formula in (3.10) to deduce $s_0<\rho ^*\leq c_2.$ Thus, the sequence $\{s_m\}$ exists in the interval $(0,c_2),$ remains in $(0,c_2)$ , and satisfies (3.11). Moreover, by the definition (3.10),
since by Lemma 3.2, the sequence $\{\frac {s_{m+1}}{s_m^2}\}$ is strictly decreasing. Then, the estimate (3.14) and $s_m<s_0$ imply
where $c=\frac {|\psi _{h_0,h}(s_0)|}{s_0}\in [0,1)$ by Lemma 2.3.
Consequently, by (3.15), we deduce the validity of the item (3.12) the second item in Lemma 2.2.
Furthermore, we obtain
Finally, the item (3.13) follows by the estimates (3.14) and (3.16).
Remark 3.5 Let us show the implications of the new conditions $(A_1)-(A_2),$ when compared to the old ones $(A_3)'$ and $(A_4)',$ given in [Reference Ferreira9] for the specializations
and
for some parameters $l_0>0, l>0$ and $l_1>0.$ Notice that
and
Case 1: $l_0=l=l_1$ (Lipschitz). Then, the definitions of the parameters $c_2$ and $\rho ^*$ imply that
where in order to obtain the second element in the preceding set, we solved for the variable s the inequality
Case 2: By the condition $(A_5)$ ,
Then, the corresponding radius is
where we again solved for s the inequality
Clearly, it follows by (3.17)–(3.20) that
Notice that by (3.19) and (3.21) as $\frac {\ell }{\ell _1}\longrightarrow 0,$ the new radius is at least as three times larger. The value $\rho _1^*$ is due to Rheinboldt [Reference Rheinboldt, Tikhonov, Kuhnert, Kuznecov, Moszynski and Wakulicz21] and Traub [Reference Traub and Wozniakowski24].
In the numerical section, examples are developed where (3.17), (3.18), (3.20), and (3.22) are valid as strict inequalities. Moreover, under Case 1, the corresponding sequence $\{u_n\}$ to $\{s_n\}$ is
for $n=0,1,2,\ldots $ ,
whereas
for $n=0,1,2,\ldots $ . Therefore, by (3.17), (3.18), and these definitions,
for each $n=0,1,2,\ldots $ .
Hence, we conclude that the new error bounds $\{s_n\}$ are tighter than the old ones $\{u_n\}$ given in [Reference Ferreira9]. Notice also that these advantages hold even if $h_0, h$ , and $h_1$ are chosen in ways other than the Lipschitz, since $h_0$ and h are always tighter than $h_1.$ Moreover, the aforementioned advantages require no additional computational effort, since in practice the calculation to determine the function $h_1$ also requires that of the functions $h_0$ and h as special cases. Moreover, there are advantages concerning the uniqueness of the solution $x^*$ (see Remark 3.5). Furthermore, notice that the conditions on $c_1, c_2, c_3,\rho ^*$ are always weaker than the corresponding ones given in [Reference Ferreira9] for $h_0=h=h_1.$ That is, if the old convergence conditions hold, so do the new ones but not necessarily vice versa. Finally, notice that all other results involving the sequence $\{u_n\}$ and the function can be replaced by $\{s_n\}$ and the functions $h_0$ and h (see [Reference Ferreira9] and the references therein).
4 Local Analysis
In this section, we present the association of the majorant functions $h_0$ and h to the nonlinear operator $F.$ Next, we first present a Banach perturbation lemma of inverses for linear operators.
Lemma 4.1 Under the hypotheses of Lemmas 3.1–3.4, further assume that
Then, the following items are valid: $G'(x)^{-1}\in L(E_2,E_1)$ and
for each Thus, in particular, $G'$ is invertible on
Proof It follows by the condition (4.1) that $h_0'(\|x-x^*\|)<0.$ Thus, the condition $(A_2)$ and Definition 2.1 give
Hence, by Lemma 2.1, $M^{-1}G'(x)\in L(E_1,E_2),$ so $G'(x)^{-1}\in L(E_2,E_1).$
Moreover,
Finally, the last item is valid due to the inequality
It is well known that the Newton method at a single point is the solution of the linearization at the point at hand. Thus, it is important to study the linearization error at that point in D:
for each $u_2,u_1\in D.$
This error is controlled by the error in the linearization of the majorant functions $h_0, h$ , and $h_1$ defined as follows for each $v_1,v_2\in [0,R)$ :
and
Notice that the pair $(\Lambda _G,\lambda _{h_1})$ are used in the local analysis of convergence in [Reference Ferreira9] (see also [Reference Ferreira9–Reference Ferreira and Svaiter13]).
Lemma 4.2 Assume that Then, the following error estimate is valid:
for each
Proof Simply replace $\lambda _{h_1}$ by $\lambda _{h}$ and Definition 2.4 by Definition 2.2 in the corresponding proof of Lemma 2.10 in [Reference Ferreira9], provided that $M=G'(x^*).$
The invertibility of $G'$ is guaranteed by Lemma 4.1 in the ball $U(x^*,\rho ^*).$
Thus, the Newton iteration operator
is well defined on $U(x^*,\rho ^*).$
The next auxiliary lemma develops conditions to guarantee that the Newtonian iterates can be repeated indefinitely.
Lemma 4.3 Let $s\in (0,\rho ^*).$ Assume $x\in U(x^*,s).$ Then, the following error estimate is valid:
for each $x\in U(x^*,s).$
Proof Inequality (4.6) is trivially valid for $x=x^*,$ since $G'(x^*)=0.$ Thus, we can assume $\|x-x^*\|\in (0,s).$ Lemma 4.1 assures the invertibility of the linear operator $G'(x).$ Then, we have the identity
Consequently, by (4.7) and Lemmas 4.1 and 4.2, we get
since $h_0(0)=h(0)=0.$ Then, the application of Lemma 3.2 for $x\in U(x^*,s)$ implies
Next, by (4.8) and (4.9), we obtain
leading to the validation of the item (4.8).
Corollary 4.4 Let $s\in (0,\rho ^*).$ Then, the following are valid:
and
Proof By the application of Lemma 4.3 for $x\in U[x^*,s]$ and since $\|x-x^*\|\leq s,$ we obtain $\|N_G(x)-x^*\|\leq |\psi _{h_0,h}(s)|.$
Hence, the item (4.10) is valid. Moreover, Lemma 3.2 and $\rho ^* \leq c_2$ imply ${|\psi _{h_0,h}(s)| \leq s}$ . Therefore, the item (4.11) is also valid.
Next, a domain is determined that contains only one solution which is $x^*$ .
Proposition 4.5 Let . Assume that zero is the only solution of the equation $h_0(s)=0$ in the closed interval $[0,s]$ , i.e., $h_0(s)<0$ . Then, the limit point $x^*$ is the only solution of the equation $G(x)=0$ in the closed ball $U[x^*,s]$ . Consequently, the limit point $x^*$ is the only solution of the equation $G(x)=0$ in the open ball $U(x^*,c_3)$ .
Proof Simply replace the tighter function $h_0$ than $h_1$ that is actually needed in the proof of the corresponding Lemma 2.13 in [Reference Ferreira9].
Remark 4.6 If $h_0=h=h_1$ , Lemma 4.1 and Proposition 4.5 reduce to the corresponding Lemmas 2.9 and 2.13, respectively. Otherwise, since
and
The new results provide tighter upper error bounds on the norms $\|G'(x)^{-1}M\|$ and a larger radius of uniqueness for the solution $x^*$ (see also Remark 3.5). In particular, under the Lipchitz case, we have
In order to specify $c_3$ , we must solve the inequality $h_0(s)<0$ , leading to
The corresponding $c_3$ given in Lemma 2.13 in [Reference Ferreira9] is obtained if we solve for s the inequality $h_1(s)<0$ , leading to
Consequently, we get
Next, the proof of Theorem 2.4 is developed based on the abovementioned auxiliary results.
Proof of Theorem 2.4
It follows by (2.2) and (4.5) that the Newton iteration $\{x_n\}$ satisfies the identity
Next, notice that all the items concerning the scalar sequence $\{s_n\}$ are shown in Lemma 3.4. Moreover, the Newton iteration $\{x_n\}$ is well defined and belongs in $U(x^*,\rho ^*)$ . Indeed, for $x_0 \in U(x^*,\rho ^*)$ and, since $\rho ^* \leq c_1$ , the formula (4.12), the item $N_G (U(x^*,\rho ^*)) \subset U(x^*, \rho ^*)$ in Corollary 4.4 (see (4.11)) and Lemma 4.1 validate the claim. Next, we are going to determine that $\lim _{m \rightarrow + \infty } x_m =x^*$ . Let us show that the scalar sequence $\{s_m\}$ majorizes the Newton sequence $\{x_m\}$ , i.e.,
By the choice $s_0=\|x^*-x_0\|$ , inequation (4.13) holds if $m=0$ . We employ mathematical induction and assume that $\|x^*-x_m\| \leq s_m$ . Then, by Lemma 3.4, $\{s_m\} \subset (0,\rho ^*)$ . Furthermore, the application of the definition of the sequence $\{s_n\}$ given by the formula (3.10) and the identity (4.12) gives
The proof of the induction for inequation (4.13) is complete. Furthermore, by (4.13) and (3.12), $\lim _{m \rightarrow + \infty }x_m=x^*$ , since $\lim _{m \rightarrow +\infty } s_m =0$ . The claim about the optimality of the convergence radius $\rho ^*$ is given in Lemma 2.15 in [Reference Ferreira9]. The first inequality in (2.4) follows by Lemma 3.4 and (4.12) since
for each $m=0,1,2,...$ . Thus, the last inequation and the definition of the majorant sequence $\{s_m\}$ show the validity of the first inequation in (2.4). Finally, the uniqueness of the solution $x^*$ is shown in Proposition 4.5.
5 Special cases and examples
It is worth noticing that, if we simply specialize the functions $h_0$ and h, the results extend the applicability of the classical cases [Lipchitz] (see also Remark 3.5), under the Smale [Reference Smale, Ewing, Gross and Martin23] or Wang [Reference Wang and Ouyang25, Reference Wang26] conditions and under the Nesterov et al. conditions [Reference Nesterov and Nemirovskii16]. We leave the details to the motivational reader.
Next, we present two examples to test the convergence conditions and further validate the theoretical results. We have chosen $M=G'(x^*).$
Example 5.1 Consider, the choice $B_1=B_2=\mathbb {R}^3$ and $D=P[0,1].$ Define the operator on D as
and $t_1, t_2,t_3\in \mathbb {R}.$ Clearly, $t^*=(0,0,0)^T$ solves the equations. Then, by the definition (5.1), it follows that the derivative according to Fréchet $G'$ of the operator F is defined as
We have $h_0(s)=(e-1)s,\ h(s)= e^{\frac {l}{e-1}}s$ , and $h_1(s)=es$ . Then, $\rho _1^*=0.2453$ [Reference Ferreira9] and $\rho ^*=0.3827.$ Thus, the previous radius of convergence $\rho _1^*$ is smaller than the new one $\rho ^*.$ This allows for a wider choice of initial guesses $x_0,$ and other benefits as already mentioned under novelty in the introduction of this article.
Example 5.2 Let $K[0,1]$ stand the space of continuous functions mapping the interval $[0,1]$ into the real number system. Let $B_1=B_2=K[0,1]$ and $D=P[x^*,1]$ with $x^*(\mu )=0.$ The operator G is defined on $K[0,1]$ as
Then, the definition of the derivative according to Fréchet [Reference Argyros4–Reference Bonnans6, Reference Robinson22] gives for the function G
for each $w\in K[0,1].$ Then, the conditions (A1)–(A5) of Theorem 2.4 are validated, since $G'(x^*(\mu ))=I$ provided that $h_0(s)=h(s)=12s$ and $h_1(s)=24s.$ Then, ${\rho _1^*=0.0278}$ and $\rho ^*=0.0556.$ Notice that as in Example 5.1, $\rho _1^* < \rho ^*.$
6 Conclusion
A very general theory for studying the convergence of Newton-type methods is developed for generating sequences approximating a solution of a generalized equation involving set-valued operators. Both the local as well as the semi-local analyses of convergence rely on the weaker conditions and the concept of generalized continuity. Moreover, we provide upper error bounds on the norms $\|x_{n+1}-x_n\|$ and $\|x^*-x_n\|.$ In particular, the semi-local analysis of convergence is based on majorizing sequences for $\{x_n\}$ generated by the method (1.2). It was shown that even specializations of the operators involved lead to better results when compared to existing ones. The future direction of our research involves the application of the developed theory on other methods [Reference Adly, Van Ngai and Nguyen1, Reference Araǵon Artacho, Belyakov, Dontchev and López2, Reference Bonnans6–Reference Dontchev and Rockafellar8, Reference Nocedal and Wright17].
Acknowledgements
We would like to express our gratitude to the anonymous reviewers for the constructive criticism of this paper.