1. Introduction
Let
X and
Y be Banach spaces, and let
K represent a closed convex subset of
Y. Two inclusion problems have been extensively studied [
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31]. The first problem is defined as
where
is a differentiable operator in the Fréchet sense. The second problem is a convex composite optimization problem:
where
h is a real-valued convex function on
Y and
F is as defined in (
1). If
, the distance function related to
K, and (
2) is solvable, then (
2) reduces to (
1). Many optimization and mathematical programming problems can be reduced to solving (
1) or (
2) [
1,
2,
3,
4,
5,
6,
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30,
31]. Regarding problem (
1), Robinson [
25,
26] developed the extended Newton method under the conditions that
K is a closed convex cone and
X is a reflexive space. Let
.
Algorithm :
Given
,
, compute
as follows: First, define the set
If
, pick
to satisfy
, and set
In the case when
for some
, the algorithm is not well-defined. It is said to be well-defined if at least one sequence is generated by this algorithm. Let
be a convex process given by
Robinson in [
25,
26] provided two conditions:
and
is Lipschitz continuous with modulus
H,
is normed with
, and
Under these conditions, Algorithm
generates a sequence
converging to some
solving (
1). Numerous authors have presented relevant convergence results lacking the affine invariant property for the operator
F [
14]. Later, in the elegant work by Li and Ng in [
20], convergence conditions addressing this problem under weaker conditions were introduced. Their study also relied on the new concept of the weak-Robinson condition for convex processes introduced in
Section 2, Subsection 2.1. The semi-local convergence of the two algorithms is presented in
Section 2, Subsection 2.2. The hybrid methods are studied in
Section 2, Subsection 2.3. The concluding remarks appear in
Section 3 and the paper ends with discussion in
Section 4.
2. Materials and Methods
2.1. Weak-Robinson Condition and Properties of Convex Processes
Let
denote the open ball in
X with center
and radius
. Also, let
denote the distance from
x to
S. We assume familiarity with the concept of convex process
as defined by Rockafellar [
27,
28], as well as its properties. We list only some in order to make the study as self-contained as possible. The following sets are needed:
and
Proposition 1
([
28]).
Let be convex processes with and . Suppose that and that is closed for all . Then, the following assertions hold:
Next, we present Lipschitz-type conditions connected to convex processes for comparison.
Definition 1.
Let , where denotes the space of linear operators from X to Y. Let be a convex process, where Z is also a Banach space. The pair is said to be center-Lipschitz continuous on if
for all and some . Set .
Definition 2.
Let E and T be as in Definition 1. Then, the pair is said to be restricted-Lipschitz continuous on if
for all and some .
Definition 3.
Let E and T be as in Definition 1. Then, the pair is said to be Lipschitz continuous on if
for all and some .
Remark 1.
It follows from these Definitions, since , that
The convergence analysis in [
4] used
. However, the proofs can be repeated with
L replacing
. Note that
and
, while
. We assume that
Otherwise, i.e., if , the results hold with replacing L.
Definition 4.
The problem (1) is said to satisfy the Weak-Robinson condition at on if
It follows that for the problem (
1), the Robinson condition at
implies the Weak-Robinson condition at
on
X.
Proposition 2.
Let and . Suppose that the problem (1) satisfies the Weak-Robinson condition at and is -Lipschitz continuous on . Then, for all , the following assertions hold:
Additionally, if X is reflexive, then
Proof. Simply replace
by the smallest
that is actually needed in the proof of Proposition 2.3 in [
4]. □
2.2. The Gauss-Newton Method for the Problem
We assume in this section that
X is reflexive,
h is a continuous convex function, and the set
K is the set of minimum points defined as
For each
and
, let
be the set of all
such that
and
Then,
if and only if
d solves the convex minimization problem:
Let
,
, and
. Then, the Gauss-Newton method for solving the problem (
2) is given by the Algorithm:
Algorithm :
Given
, compute
as follows: If
then stop. Otherwise, select
so that
, and let
.
Define the parameters
and
by
The majorizing functions for all
are given by
It follows from (
19), (
7), and these definitions that
and
Define the majorizing sequences
by
and
The sequence
is used in [
4]. Here, we use
although
can also be used according to the proof of Theorem 3.1. Next, a Kantorovich condition is provided for the convergence of the sequence
also for the sequence
.
Lemma 1.
Then, the sequence is increasingly convergent to which is the smallest of the two roots of the function g guaranteed to exist by (22). Moreover for
Proof. Simply replace
by
in the proof of Lemma 3.1 in [
4] . □
Remark 2.
The Kantorovich condition in [4] is
but not necessarily vice versa unless . Moreover,
where is given by (23) with L replaced by . Thus, the new results are at least as good as those in [4].
It is also worth noting that the improved results are obtained under the same computational effort, since in practice the computation of requires that of and L as special cases. Next, we present the semi-local convergence result for the algorithm .
Theorem 1.
Suppose that the problem (1) satisfies the Weak-Robinson condition at on and that is -center-Lipschitz continuous on and L-restricted-Lipschitz on . Suppose that
Then, any sequence generated by the algorithm is convergent to some solving (1). Moreover, the following assertions hold:
Proof. Simply replace
by
, respectively, in the proof of Theorem 3.1 in [
4]. □
Theorem 2.
Suppose that the problem (1) satisfies the Weak-Robinson condition at on , , , , and is -center-Lipschitz on , L-Lipschitz continuous on , and . Then, any sequence generated by the Algorithm converges to a solution of the problem (1). Moreover, the following error assertion holds:
Proof. Replace
by
in the proof of Theorem 3.2 in [
4]. □
Remark 3.
- (i)
Theorem 2 can be reduced to a weaker version of Robinson’s [25,26] for , , and .
- (ii)
The Kantorovich conditions for the convergence of the sequence can be weakened further.
Indeed, let us demonstrate this in the case of Algorithm
and Theorem 3.4. According to the proof of Theorem 3.5:
The sequence majorizing
in [
4], i.e.,
, can be written as:
Moreover, the convergence conditions are given in [
2] and [
4], respectively, as:
and
where
. Notice that:
holds and
as
approaches 0. Thus, the new convergence condition is infinitely weaker than the original Kantorovich condition (
33). Moreover, the following hold:
and
Hence, the error bounds on the distances
are at least as tight and the information on the location of the solution is at least as precise. The results can also be rewritten using majorant functions replacing the Lipschitz constants
, and
(see [
2,
3,
4,
5]).
2.3. The Implementation of the Algorithms
The iterates in both algorithms are computed by finding the expensive
. In the papers [
3,
4], we developed a hybrid method to address this difficulty. Specifically, we define the method
where
,
k is a natural number,
is an invertible operator,
, and
. Here,
.
Similarly, we introduce hybrid methods to replace both Algorithms. We demonstrate this with Algorithm
. The method for
follows along the same lines. We take
for simplicity. As in [
3], define the parameters:
Define the sequence
with
,
,
The sequence shall be shown to majorizing for the sequence generated by the Algorithm defined next.
Algorithm :
Given
compute
as follows: First define the set
If , pick to satisfy and set .
Let us provide a convergence criterion for the sequence
. Suppose that there exists
such that for each
,
It follows by (
39) and (
38) that
, and there exists
such that
. The limit point
is the unique least upper bound of the sequence
. We can show the analog of Theorem 2 for the sequence
generated by Algorithm
.
Theorem 3.
Suppose that the problem (1) satisfies the weak-Robinson condition at on , and is -center Lipschitz on and L-Lipschitz on ,
and (39) holds. Then, any sequence generated by Algorithm converges to a solution of the problem (1). Moreover, the following error estimate holds:
Proof. Notice that for each
,
So,
and
. Then, we can write
We also need the estimate
By replacing
by
in the proof of Theorem 2 and using (
43), we get
leading to the existence of
solving
. Then, from (
44), we have for each
Finally, by letting
in (
45), we obtain (
41). □
Remark 4.
- (i)
It is worth noticing that the condition (39) is weaker than the Kantorovich condition used in the Theorem 2 or (33) since both of these conditions imply (39) but not necessarily vice versa. Thus, (39) can replace these stronger conditions in the Theorem 2 provided that the sequence (for ) replaces the sequence .
- (ii)
The results of the Theorem 3 can be presented in a more general setting if M is not necessarily chosen to be as follows: Simply replace by M in the Definition 1 and by M in (12) where is an invertible operator. Then, the conclusions of the Theorem 3 hold this more general setting.
- (iii)
The results can be further extended if the Lipschitz continuity is replaced by the generalized continuity of the operator along the lines of our work in [2,3,4,5].
3. Concluding Remarks
Remark 5.
The results in this work can also be expressed using majorant functions instead of the Lipschitz constants and . This approach allows for even tighter error estimates (see [1,2,3,4,5]).
For instance, the constant in Definition 1 can be replaced with a function that is continuous and non-decreasing. Consequently, for all , we obtain:
If , (46) reduces to the one stated in Definition 1. Similarly, and L in Definition 2 and Definition 3 can be replaced accordingly.
Let us introduce some parameters and real functions that play a role in the semi-local convergence of the Gauss-Newton algorithm under generalized continuity conditions used to control the derivative of the operator f. Set .
Suppose: There exists a function : , which is continuous and nondecreasing and a parameter such that the equation has at least one positive solution.Denote the smallest such solution by . Set
There exists a function
which is continuous and nondecreasing. Let
be a parameter. Define the scalar sequence
for
,
,
,
and each
by
The sequence shall be shown to be majorizing in Theorem 6.1. But, first a convergence condition is provided for the sequence .
There exists
such that for all
It follows by this condition and the definition of the sequence that for each and there exists such that It is known from calculus that is the least upper bound of the sequence which is unique.
The scalar functions
and
connect to the operators on the Gauss-Newton Algorithm. Let
and
be such that
. Define
by
Let be a convex process, where Z is also a Banach space. The pair () is center-generalized continuous on for , i.e., for all
Let be as in the condition . The pair is restricted generalized continuous on , i.e., for all .
.
.
Next, the main semi-local convergence of the Gauss-Newton algorithm is shown under the conditions .
Theorem 4.
Suppose that the inclusion (1) satisfies the weak-Robinson condition at on and ()-() hold. Then, the Gauss-Newton Algorithm is well defined and any sequence converges to some such that Moreover, the following items hold for all
Proof. It follows by the weak-Robinson condition that
So, by the definition of
, (
47), (
48) and the condition (
, we get
By the reflexivity of
X, (
54) and since
,there exists
such that
. Hence,
and
. Thus, the iterate
is well defined and
so, (
51) holds for
.
Moreover, by (
54) and the Gauss-Newton Algorithm
, i.e.,
. So,(
50) holds sot
. Suppose that (
50) and (
51) hold for
. Then, we get
Similarly, we have .
Then, by the condition
for all
so
. Hence, the iterate
exists and (
51) holds for
. Furthermore, for
. By applying the condition (
)
Then, by the weak-Robinson condition, the Proposition (2) is applicable to
replacing
x leading to
and
Thus, we get
and by the induction hypothesis, (
47) and the condition
Notice that this estimate holds for
replaced by
if
. Next, we must show
It follows from (
55) and (
56) that the set in the middle of (
58) is non-empty. To show the rest in (
58), set
and
, so
for some
. We must show
. But we have
and
So,
, since
K is a cone. By the definition of
w and the induction hypothesis, it follows that
Thus,
, showing the right-hand side of the assertion (
58). Hence,we can get from (
56) and (
58) in turn,
Notice that the sequence
is decreasing, so
. Thus, by (
59)
, which together with
imply that there exists
with
such that
. Hence, we have
and
. Thus, it follows by (
59) that
. Thus, the induction for (
50) and (
51) is completed. It follows that the sequence
is fundamental in
X (since
is also fundamental as convergent). Therefore,there exists
such that
and
. Finally,from (
50) for
and the triangle inequality, we obtain
leading to (
52) if
□
Remark 6.
- label=()
-
Let us specialize the functions and ψ to be and . Then, the condition becomes and , since by the condition . Notice that the sequence becomes by the definition of the functions and ψ
According to the proof of Theorem 6.1, the sequence given by (60) majorizes provided that the conditions and hold. These conditions are weaker than (1). Indeed, this is the case, since
Hence, the sequence can be replaced by the less tight given by the formula (18). Notice that the sequence converges provided that the (see (27)). Consequently, the sequence given by (60) is tighter than used in [18] and the sufficient convergence conditions wood are weaker than used in [18]. We conclude that under these choices of the functions and ψ, the results of Theorem 6.1 specialize to the ones in Theorem 1. Clearly, the same is true for Theorem 2 if we take in Theorem 6.1.
- lbbel=()
It is well known that generalized continuity provides even sharper bounds on the derivative than Lipschitz or Hölder or other conditions.
4. Discussion
The applicability of two Gauss-Newton methods for solving inclusion and convex-composition optimization problems for Banach space-valued operators is extended using more precise majorizing sequences and under the same computational cost, providing weaker yet sufficient semi-local convergence criteria and more precise error estimates on the distances and . Moreover, the implementation of these algorithms is addressed by developing the corresponding hybrid Gauss-Newton methods. The Lipschitz constants can be replaced by the generalized continuity of .
In future research, the ideas presented in this study shall be used to extend the applicability of other iterative methods used to solve these problems.
Author Contributions
All authors contributed equally.
Funding
This research received no external funding.
Institutional Review Board Statement
Not applicable
Informed Consent Statement
Not applicable
Data Availability Statement
Not applicable
Acknowledgments
IWe would like to express our gratitude to the anonymous reviewers for the constructive criticism of this paper.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Aragon Artacho, F.J.; Dontchev, A.L.; Gaydu, M.; Geoffroy, M.H.; Veliov, V.M. 2011. "Metric Regularity of Newton’s Iteration." SIAM Journal on Control and Optimization 49:339-362.
- Argyros, I.K. 2022. The Theory and Applications of Iteration Methods with Applications. Engineering Series, Second Edition. Boca Raton, FL: CRC Press/Taylor & Francis.
- Argyros, I.K.; George, S.; Shakhno, S.; Regmi, S.; Havdiak, M.; Argyros, M.I. 2024. "Asymptotically Newton-Type Methods Without Inverses for Solving Equations." Mathematics 12:1069. [CrossRef]
- Argyros, I.K.; George, S.; Regmi, S.; Argyros, C.I. 2024. "Hybrid Newton-like Inverse Free Algorithms for Solving Nonlinear Equations." Algorithms 17:154. [CrossRef]
- Argyros, I.K.; Shakhno, S.; Yarmola, H.; Regmi, S.; Shrestha, N. 2025. "Three step inverse free Kurchatov-like methods of convergence order close to four for equations." Eur. J. Math. Anal. 5:15-15.
- Argyros, I.K.; Shrestha, N. 2017. "Extending the local convergence analysis of Newton’s method." Commun. Appl. Nonlinear Anal. 24:49-60.
- Blum, L.; Cucker, F.; Shub, M.; Smale, S. 1998. Complexity and Real Computation. New York: Springer-Verlag.
- Burke, J.V.; Ferris, M.C. 1995. "A Gauss-Newton Method for Convex Composite Optimization." Mathematical Programming 71:179-194.
- Chen, J. 2008. "The Convergence Analysis of Inexact Gauss-Newton Methods for Nonlinear Problems." Computational Optimization and Applications 40(1):97-118.
- Chen, J.; Li, W. 2005. "Convergence of Gauss-Newton’s Method and Uniqueness of the Solution." Applied Mathematics and Computation 170(1):686-705.
- Dedieu, J.P.; Kim, M.H. 2002. "Newton’s Method for Analytic Systems of Equations with Constant Rank Derivatives." Journal of Complexity 18(1):187-209.
- Dennis Jr., J. E.; Schnabel, R.B. 1996. Numerical Methods for Unconstrained Optimization and Nonlinear Equations. Classics in Applied Mathematics, vol. 16. Philadelphia, PA: SIAM.
- Deuflhard, P.; Heindl, G. 1979. "Affine Invariant Convergence Theorems for Newton’s Method and Extensions to Related Methods." SIAM Journal on Numerical Analysis 16:1-10.
- Deuflhard, P. 2004. Newton Methods for Nonlinear Problems: Affine Invariance and Adaptive Algorithms. Berlin; Heidelberg: Springer-Verlag.
- Dontchev, A.L.; Rockafellar, R.T. 2009. Implicit Functions and Solution Mappings: A View from Variational Analysis. Springer Monographs in Mathematics. Dordrecht: Springer.
- Ferreira, O.P.; Goncalves, M.L.N.; Oliveira, P.R. 2013. "Convergence of the Gauss-Newton Method for Convex Composite Optimization Under a Majorant Condition." SIAM Journal on Optimization 23(3):1757-1783.
- Ferreira, O.P.; Goncalves, M.L.N.; Oliveira, P.R. 2011. "Local Convergence Analysis of the Gauss-Newton Method under a Majorant Condition." Journal of Complexity 27:111-125.
- Ferreira, O.P.; Goncalves, M.L.N.; Oliveira, P.R. 2011. "Local Convergence Analysis of Inexact Newton-Like Methods under Majorant Condition." Computational Optimization and Applications 48:1-21.
- Kantorovich, L.V.; Akilov, G.P. 1982. Functional Analysis. 2nd ed. Oxford: Pergamon Press. (translated from Russian by Howard L. Silcock).
- Li, C.; Ng, K.F. 2007. "Majorizing Functions and Convergence of the Gauss-Newton Method for Convex Composite Optimization." SIAM Journal on Optimization 18:613-642.
- Li, C.; Wang, X.H. 2002. "On Convergence of the Gauss-Newton Method for Convex Composite Optimization." Mathematical Programming 91:349-356.
- Li, C.; Ng, K.F. 2012. "Convergence Analysis of the Gauss-Newton Method for Convex Inclusion and Convex-Composite Optimization Problems." Journal of Mathematical Analysis and Applications 389:469-485.
- Nocedal, J.; Wright, S.J. 1999. Numerical Optimization. Springer-Verlag.
- Proinov, P.D. 2009. "General Local Convergence Theory for a Class of Iterative Processes and Its Applications to Newton’s Method." Journal of Complexity 25(1):38-62.
- Robinson, S.M. 1972. "Extension of Newton’s Method to Nonlinear Functions with Values in a Cone." Numerische Mathematik 19:341-347.
- Robinson, S.M. 1972. "Normed Convex Process." Transactions of the American Mathematical Society 174:127-140.
- Rockafellar, R.T. 1970. Convex Analysis. Princeton: Princeton University Press.
- Rockafellar, R.T. 1967. Monotone Processes of Convex and Concave Type. Memoirs of the American Mathematical Society, no. 77.
- Smale, S. 1986. "Newton’s Method Estimates from Data at One Point." In The Merging of Disciplines: New Directions in Pure, Applied, and Computational Mathematics, 185-196. New York: Springer.
- Stewart, G.W. 1969. "On the Continuity of the Generalized Inverse." SIAM Journal on Applied Mathematics 17:33-45.
- Wedin, P.A. 1973. "Perturbation Theory for Pseudo-Inverses." BIT 13:217-232.
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).