Preprint
Article

This version is not peer-reviewed.

Eigenvalue Bounds for Symmetric, Multiple Saddle-Point Matrices with SPD Preconditioners

Submitted:

02 February 2026

Posted:

06 February 2026

You are already at the latest version

Abstract
We derive eigenvalue bounds for symmetric block-tridiagonal multiple saddle-point systems preconditioned with the symmetric positive definite (SPD) preconditioner proposed by J. Pearson and A. Potschka in 2024, and further studied by L. Bergamaschi and coauthors, for double saddle point problems, with inexact Schur complement matrices. The analysis applies to an arbitrary number of blocks. Numerical experiments are carried out to validate the proposed estimates.
Keywords: 
;  ;  ;  

1. Introduction

We consider the iterative solution of a block tridiagonal multiple saddle-point linear system A x = b , where
A = A 0 B 1 0 0 B 1 A 1 B 2 0 B 2 A 2 0 B N 0 0 B N ( 1 ) N A N .
We assume that A 0 R n 0 × n 0 is symmetric positive definite, all other square block matrices A k R n k × n k are symmetric positive semi-definite and B k R n k × n k 1 have full rank (for k = 1 , , N ). We assume also that n k n k 1 for all k. These conditions are sufficient to ensure the invertibility of A .
Linear systems involving matrix A arise mostly with N = 2 (double saddle–point systems) in many scientific applications including magma–mantle dynamics [3], liquid crystal director modeling [4] or in the coupled Stokes–Darcy problem [5,6,7,8], and the preconditioning of such linear systems has been considered in [9,10,11]. In particular, block diagonal preconditioners for matrix A have been studied in [12,13,14,15,16].
Multiple saddle-point linear systems with N > 2 have recently attracted the attention of a number of researchers. Such systems often arise from modeling multiphysics processes, i.e. the simultaneous simulation of different aspects of physical systems and the interactions among them. Preconditioning of the 4 × 4 saddle-point linear system ( N = 3 ), although with a different block structure, has been addressed in [17] to solve a class of optimal control problems. A practical preconditioning strategy for multiple saddle–point linear systems, based on sparse approximate inverses of the diagonal blocks of the block diagonal Schur complement preconditioning matrix, is proposed in [18], for the solution of coupled poromechanical models and the mechanics of fractured media. Theoretical analysis of such preconditioners has been carried on in [19]. In [17], two 5 × 5 multiple saddle-point systems arising from optimal control problems constrained with either the heat or the wave equation are addressed, and iteratively solved by robust block diagonal preconditioners.
In this work we consider the preconditioner proposed in [1] (PP preconditioner in short) to develop a spectral analysis of the preconditioned matrix for an arbitrary number of blocks and in presence of inexact Schur complements, in the line of [15] in which the block diagonal preconditioner with exact Schur complements is considered. We will define a sequence of polynomials by a three-term recurrence, whose extremal roots characterize the eigenvalues of the preconditioned matrix in terms of the nonnegative eigenvalues of the preconditioned Schur complements. Using a similar technique, we will extend results, obtained in [2] for a double saddle-point, to multiple saddle-point linear system.
Arguably the most prominent Krylov subspace methods for solving the linear system A x = b are preconditioned variants of MINRES [20] and GMRES [21]. In contrast to GMRES, the previously-discovered MINRES algorithm can explicitly exploit the symmetry of A . As a consequence, MINRES features a three-term recurrence relation, which is beneficial for its implementation (low memory requirements because subspace bases need not be stored) and its purely eigenvalue-based convergence analysis (via the famous connection to orthogonal polynomials; see [22,23]). Specifically, if the eigenvalues of the preconditioned matrix are contained within [ ρ l , ρ u ] [ ρ l + , ρ u + ] , for ρ l < ρ u < 0 < ρ l + < ρ u + such that ρ u + ρ l + = ρ u ρ l , then at iteration k the Euclidean norm of the preconditioned residual r k satisfies the bound
r k r 0 2 | ρ l ρ u + | | ρ u ρ l + | | ρ l ρ u + | + | ρ u ρ l + | k / 2 .
The paper is structured as follows: In Section 2 we describe the preconditioner and write the eigenvalue problem for the preconditioned matrix. In Section 3, a sequence of polynomials is defined by recurrence, and the connection of these with the eigenvalues of the preconditioned matrix is stated. Then, in Section 4, we characterize the extremal roots of such polynomials and the dependence of such roots with two vectors of parameters, containing information on the inexact Schur complements. Section 5 is devoted to comparing the bounds with the exactly computed eigenvalues on a large set of synthetic experiments for the number of blocks ranging from 3 to 5. In Section 6 we present as a realistic test case, the 3D Mixed Finite Element discretization of the Biot model, which gives raise to a double saddle-point linear system, to again validate the bounds and to show the good performance of the PP preconditioner, also in comparison with the most known block diagonal preconditioner. Section 7 draws some conclusions.

2. The Preconditioner

Consider the exact block factorization of matrix A = U D 1 U , where
D = A 0 S 1 S 2 ( 1 ) N S N , U = A 0 B 1 S 1 B 2 S 2 B N ( 1 ) N S N ,
with S k = A k + B k S k 1 1 B k , k = 1 , , N . The preconditioner is initially obtained by removing the minus signs in D, thus obtaining a symmetric positive definite matrix P = U | D 1 | U , and then, in view of its practical application, by approximating the Schur complements with their inexact counterparts. The final expression of the preconditioner is therefore P = P U P D 1 P U , where
P D = S ^ 0 S ^ 1 S ^ 2 S ^ N , P U = S ^ 0 B 1 S ^ 1 B 2 S ^ 2 B N ( 1 ) N S ^ N ,
with
S ^ 0 A 0 S ˜ k = A k + B k S ^ k 1 1 B k S ^ k S ˜ k .
This preconditioner has been proposed by Pearson and Potschka in [1], in the framework of PDE-constrained optimization. Since P is symmetric and positive definite, this allows the use of the MINRES iterative method.
Finding the eigenvalues of P 1 A is equivalent to solving
P D 1 / 2 A P D 1 / 2 u = λ P D 1 / 2 P L P D 1 / 2 P D 1 / 2 P U P D 1 / 2 u , u = u 1 u 2 u N + 1 .
Exploiting the block components of this generalized eigenvalue problem, we obtain
E 0 R 1 R 1 E 1 R 2 0 R 2 E 2 R 3 R N 1 ( 1 ) N 1 E N 1 R N R N ( 1 ) N E N u 1 u 2 u 3 u N u N + 1 = = λ I R 1 R 1 I + R 1 R 1 R 2 0 R 2 I + R 2 R 2 R 3 ( 1 ) N R N 1 I + R N 1 R N 1 ( 1 ) N + 1 R N ( 1 ) N + 1 R N I + R N R N u 1 u 2 u 3 u N u N + 1
where
R k = S ^ k 1 / 2 B k S ^ k 1 1 / 2 , k = 1 , , N ; E k = S ^ k 1 / 2 A k S ^ k 1 / 2 , k = 0 , , N R k R k + E k = S ^ k 1 / 2 S ˜ k S ^ k 1 / 2 S ¯ k .
Componentwise, we write the eigenvalue problem (3) as
0 = ( E 0 λ I ) u 1 + ( 1 λ ) R 1 u 2 , 0 = ( 1 λ ) R 1 u 1 + ( E 1 λ ( I + R 1 R 1 ) ) u 2 + ( 1 + λ ) R 2 u 3 , 0 = ( 1 + λ ) R 2 u 2 + ( E 2 λ ( I + R 2 R 2 ) ) u 3 + ( 1 λ ) R 3 u 4 , 0 = ( 1 + ( 1 ) N 1 λ ) R N 1 u N 1 + ( ( 1 ) N 1 E N 1 λ ( I R N 1 R N 1 ) ) u N + ( 1 + ( 1 ) N λ ) R N u N + 1 , 0 = ( 1 + ( 1 ) N λ ) R N u N + ( ( 1 ) N E N λ ( I + R N R N ) ) u N + 1 .
The matrices R k R k are all symmetric positive definite. We define two indicators γ E ( k ) and γ R ( k ) using the Rayleigh quotient
α E ( k ) λ min E k , β E ( k ) λ max E k , γ E ( k ) w = w T E k w w T w α E ( k ) , β E ( k ) I E k , i = 0 , , N α R ( k ) λ min R k R k T , β R ( k ) λ max R k R k T , γ R ( k ) w = w T R k R k T w w T w α R ( k ) , β R ( k ) I R k , i = 1 , , N .
A third set of indicators, γ S ( k ) , linked to the previous ones as
γ S ( k ) ( w ) = γ E ( k ) ( w ) + γ R ( k ) ( w ) , w R n k ,
describes the field-of-values of each preconditioned Schur complement S ¯ k :
α S ( k ) λ min S ¯ k , β S ( k ) λ max S ¯ k , γ S ( k ) w = w T S ¯ w w T w α S ( k ) , β S ( k ) I S k , i = 1 , , N .
If the Schur complements are exactly inverted, then E 0 = I and R k R k + E k = I . In such a case, denoting γ R ( 0 ) = 0 , we have γ S ( k ) ( w ) = γ E ( k ) ( w ) + γ R ( k ) ( w ) 1 , j = 0 , , N , w , and the eigenvalues of (3) are either 1 or 1, with multiplicity, respectively, n 1 + n 3 + and n 0 + n 2 + (see [1], Theorem 2.1). In the following, we will often remove the argument w from the previously defined indicators. We define the two vectors of parameters:
γ E = γ E ( 0 ) , , γ E ( N ) , γ R = γ R ( 1 ) , , γ R ( N ) .
We will now make some very mild assumptions on two of these indicators:
The value 1 is strictly included in both I E 0 and I S 1 , namely
1. 
α E ( 0 ) < 1 < β E ( 0 ) ,
2. 
α S ( 1 ) < 1 < β S ( 1 ) .
These assumptions are very commonly satisfied in practice, meaning that 1 is in both the spectra of the preconditioned ( 1 , 1 ) block and of the preconditioned Schur complement S ¯ 1 .

3. Characterization of the Eigenvalues of the Preconditioned Matrix

We now recursively define a sequence of polynomials, with γ E and γ R as parameters, whose roots are strictly related with eigenvalues of P 1 A .
Definition 1.
U 0 ( λ , γ R , γ E ) = 1 , U 1 ( λ , γ R , γ E ) = λ γ E ( 0 ) , U k + 1 ( λ , γ R , γ E ) = ( λ ( 1 + γ R ( k ) ) + ( 1 ) k + 1 γ E ( k ) ) U k ( λ , γ R , γ E ) γ R ( k ) ( λ + ( 1 ) k ) 2 U k 1 ( λ , γ R , γ E ) , k 1 .
Then a sequence of matrix valued function is also defined by recurrence:
Definition 2.
Y 1 ( λ ) = λ I E 0 , Y k + 1 ( λ ) = ( 1 ) k + 1 E k + λ I + R k R k ( λ + ( 1 ) k ) 2 R k Y k ( λ ) 1 R k , k 1 , and λ s . t . 0 σ ( Y k ( λ ) ) .
Notation. We use the notation I k to denote the union of intervals bounding the roots of polynomials of the form U k over the valid range of γ E , γ R .
We recall a technical lemma, based on an idea in [24], whose proof can be found in [2].
Lemma 1.
Let Y be a symmetric matrix valued function defined in F R and
0 [ min { σ ( Y ( ξ ) ) } , max { σ ( Y ( ξ ) ) } ] for   all ξ F .
Then, for arbitrary s 0 , there exists a vector v 0 such that
s Y ( ξ ) 1 s s s = 1 γ Z with γ Z = v Y ( ξ ) v v v .
The next lemma, which will be used in the proof of the subsequent Theorem 1, links together the two sequences previously defined.
Lemma 2.
For every u 0 , there is a choice of γ for which
u Y k + 1 ( λ ) u u u = U k + 1 ( λ ) U k ( λ ) for   all λ j = 1 k I j .
Proof. 
This is shown by induction. We first define
η k ( λ ) = ( 1 ) k + 1 γ E ( k ) + λ ( 1 + γ R ( k ) ) , μ ¯ k ( λ ) = ( λ + ( 1 ) k ) 2 .
For k = 0 we have u Y 1 ( λ ) u u u = λ γ E ( 0 ) = U 1 ( λ ) U 0 ( λ ) for all λ R . If k 1 , the condition λ I k , together with the inductive hypothesis u Y k ( λ ) u u u = U k ( λ ) U k 1 ( λ ) , implies invertibility of Y k ( λ ) . Moreover, this is equivalent to the condition 0 [ min { σ ( Y ( ξ ) ) } , max { σ ( Y ( ξ ) ) } ] that guarantees the applicability of Lemma 1. Therefore, we can write
u Y k + 1 ( λ ) u u u = η k ( λ ) μ ¯ k ( λ ) u R k Y k ( λ ) 1 R k u u u = w = R k u η k ( λ ) μ ¯ k ( λ ) w Y k ( λ ) 1 w w w γ R ( k )
We then apply Lemma 1 and the inductive hypothesis to write
w Y k ( λ ) 1 w w w = U k 1 ( λ ) U k ( λ ) .
Substituting into (10) and using relation (8) we obtain
η k ( λ ) μ ¯ k ( λ ) w Y k ( λ ) 1 w w w γ R ( k ) = η k ( λ ) U k ( λ ) μ ¯ k ( λ ) γ R ( k ) U k 1 ( λ ) U k ( λ ) = U k + 1 ( λ ) U k ( λ ) .
   □
We now state the main results of this Section:
Theorem 1.
The eigenvalues of P 1 A are located in I = k = 1 N + 1 I k .
Proof. 
The proof is carried out through induction on k, that is for every k N + 1 either
( i ) λ I k or ( ii ) u k = ( 1 + ( 1 k ) λ ) Y k ( λ ) 1 R k u k + 1 ,
and for k = N + 1 only condition (i) can hold. Let u = u 1 , , u k + 1 be an eigenvector of (4). Assume that λ I E 0 (for k = 0 ), then Y 1 ( λ ) is invertible. From the first equation of (4) we obtain
( λ I E 0 ) u 1 = ( 1 λ ) R 1 u 2 Y 1 ( λ ) u 1 = ( 1 λ ) R 1 u 2 ,
whereupon inserting (11) into the second row of (4) yields
E 1 + λ ( I + R 1 R 1 ) ( λ 1 ) 2 R 1 ( λ I E 0 ) 1 R 1 Y 2 ( λ ) u 2 = ( 1 + λ ) R 2 u 3 .
Pre-multiplying the left hand side of (12) by u 2 u 2 u 2 , we obtain that
u 2 Y 2 ( λ ) u 2 u 2 u 2 = U 2 ( λ ) U 1 ( λ )
If λ is a zero of U 2 then λ I 2 , Otherwise Y 2 ( λ ) is invertible and definite, and we can write u 2 = ( 1 + λ ) Y 2 ( λ ) 1 u 3 . Assume now the inductive hypothesis holds for k 1 . If λ I k 1 , then Y k 1 ( λ ) is definite and invertible. We can write
Y k ( λ ) u k ( 1 ) k E k 1 + λ ( I + R k 1 R k 1 ) ( λ + ( 1 ) k 1 ) 2 R k 1 Y k 1 1 R k 1 u k = ( 1 + ( 1 ) k λ ) R k u k + 1 .
Since
u k Y k ( λ ) u k u k u k = U k ( λ ) U k 1 ( λ ) ,
then either λ is a zero of U k ( λ ) and hence λ I k or Y k ( λ ) is definite, in such a case we can write u k = ( 1 + ( 1 ) k λ ) Y k ( λ ) 1 u k + 1 . The induction process ends for k = N + 1 . In this case we have that
Y N + 1 ( λ ) u N + 1 = 0 ,
and the condition λ I N + 1 must hold, noticing that u N + 1 = 0 would imply that also u N = = u 1 = 0 contradicting the definition of an eigenvector.    □

4. Bounds on the Roots of { U k }

In this Section we will first characterize the set I = I k . Then we will specify the dependence of the zeros U k ( λ ) on the parameters γ E , γ R .
Proposition 1.
Since the polynomials of the sequence (8) are monic
lim λ + U k ( λ ) = + lim λ U k ( λ ) = ( 1 ) k ·
Lemma 3.
Given the sequence of polynomials (8), for all k 1
sgn ( U 2 k 1 ( 0 ) ) = sgn ( U 2 k ( 0 ) ) = ( 1 ) k .
In other words, the signs of U k ( 0 ) follow the repeating pattern , + + , , + + , starting from k = 1 . Equivalently, s g n ( U k ( 0 ) ) = ( 1 ) k 2 .
Proof. 
We prove this by induction of k. The claim holds for k = 1 :
U 1 ( 0 ) = γ E ( 0 ) < 0 , U 2 ( 0 ) = γ E ( 1 ) γ E ( 0 ) γ R ( 1 ) < 0 .
Assume that the claim holds for some k 1 , that is the sign of U 2 k 1 ( 0 ) and U 2 k ( 0 ) is s = ( 1 ) k , then
sgn ( U 2 k + 1 ( 0 ) ) = ( 1 ) 2 k + 1 sgn ( U 2 k ( 0 ) ) sgn ( U 2 k 1 ( 0 ) ) = s = ( 1 ) k + 1 . sgn ( U 2 k + 2 ( 0 ) ) = ( 1 ) 2 k + 2 sgn ( U 2 k + 1 ( 0 ) ) sgn ( U 2 k ( 0 ) ) = s = ( 1 ) k + 1 .
   □
We now show that, based on Assumptions 1, both 1 and 1 are strictly included in I , in fact α E ( 0 ) < 1 < β E ( 0 ) implies 1 is strictly included in I 1 , and α S ( 1 ) < 1 < β S ( 1 ) implies that 1 is strictly included in I 2 . In fact
U 2 ( 1 ; γ E ( 0 ) , γ E ( 1 ) , γ R ( 1 ) ) = 1 + γ E ( 0 ) 3 γ R ( 1 ) + γ E ( 0 ) ( γ R ( 1 ) γ E ( 1 ) ) γ E ( 1 ) .
Using γ E ( 0 ) = 1 , we have U 2 ( 1 ; 1 , γ E ( 1 ) , γ R ( 1 ) ) = 2 ( 1 γ E ( 1 ) γ R ( 1 ) ) = 2 ( 1 γ S ( 1 ) ) U 2 ( 1 ; 1 , γ S ( 1 ) ) . That 1 strictly belongs to I 2 comes from U 2 ( 1 ; 1 , α S ( 1 ) ) > 0 and U 2 ( 1 ; 1 , β S ( 1 ) ) < 0 .
We now state a results regarding the extremal roots of the polynomials { U k } .
Theorem 2.
Assume that the polynomial U k ( λ ) has s k distinct roots, and let us denote as ξ 1 ( k ) < ξ 2 ( k ) < ξ s k ( k ) the roots of U k ( λ ) . If ξ 1 ( k ) < 1 and ξ k ( k ) > 1 , k , then the extremal roots of the polynomials U k satisfy
ξ 1 ( k ) < ξ 1 ( k 1 ) < < ξ 1 ( 2 ) , and ξ 1 ( 1 ) < < ξ s k 1 ( k 1 ) < ξ s k ( k ) .
Proof. 
We prove the claim by induction, starting from l = 2 . Consider first the positive roots. The basis of the induction is
1 < ξ 1 ( 1 ) < ξ 2 ( 2 ) ,
which is readily proved by taking ξ 1 ( 1 ) = γ E ( 0 ) > 1 and observing that U 2 ( γ E ( 0 ) ) < 0 . Assume now that the claim holds for l 1 , that is, ξ s l 1 ( l 1 ) < ξ s l ( l ) . This implies that U l 1 ( ξ s l ( l ) ) > 0 . Then, from the recursion (8)
U l + 1 ( ξ s l ( l ) ) = γ R ( ξ s l ( l ) + ( 1 ) l ) 2 U l 1 ( ξ s l ( l ) ) < 0 ,
which implies that ξ s l ( l ) < ξ s l + 1 ( l + 1 ) .
Consider now the negative roots. If γ E ( 0 ) < 1 and γ S ( 1 ) > 1 , direct computation shows that U 2 ( 1 ) < 0 , providing ξ 1 ( 2 ) < 1 . Moreover, U 3 ( ξ 1 ( 2 ) ) = γ R ( 2 ) ( ξ 1 ( 2 ) + 1 ) 2 ( ξ 1 ( 2 ) γ E ( 0 ) ) > 0 , which implies ξ 1 ( 3 ) < ξ 1 ( 2 ) . Assume now that the claim holds for l 1 , that is ξ 1 ( l 1 ) > ξ 1 ( l ) which implies that sgn ( U l 1 ( ξ 1 ( l ) ) ) = ( 1 ) l 1 . Then, from the recursion (8)
sgn ( U l + 1 ( ξ 1 ( l ) ) ) = sgn U l 1 ( ξ 1 ( l ) ) ) = ( 1 ) l 1 = ( 1 ) l ,
which finally shows that ξ 1 ( l + 1 ) < ξ 1 ( l ) , as desired.    □
The results of this section allow us to characterize the set I in Theorem 1 like
I = [ λ L B , λ U B ] [ λ + L B , λ + U B ] ,
where
λ L B = ξ 1 ( N + 1 ) λ U B = max j { max { ξ l ( j ) , s . t . ξ l ( j ) < 0 } } λ + L B = min j { min { ξ l ( j ) , s . t . ξ l ( j ) > 0 } } λ + U B = ξ s N + 1 ( N + 1 ) .

4.1. How Zeros of { U k } Move Depending on γ E ( j ) , γ R ( j )

Now the question is: for which values of the parameters γ E ( i ) , γ R ( i ) the extremal values of the roots is attained? The following results will allow us to state that the extremal values of the zeros of U k + 1 ( λ , γ E , γ R ) are obtained at the extremal values of γ E and γ R , namely { α E ( i ) , β E ( i ) ; α R ( i ) , β R ( i ) } .
Let ξ an extremal zero of the polynomial U k + 1 (smallest/largest negative, smallest/largest positive) for a particular combination of the parameters. Then taking separately one of the parameters, γ * we can write γ * as a convex combination of its extremal values, γ * = α γ * min + ( 1 α ) γ * max and write
U k + 1 ( λ , γ * ) = s 1 ( λ ) + γ * s 2 ( λ ) = s 1 ( λ ) + ( α γ * min + ( 1 α ) γ * max ) s 2 ( λ ) = α ( s 1 ( λ ) + γ * min s 2 ( λ ) ) + ( 1 α ) ( s 1 ( λ ) + γ * max s 2 ( λ ) ) = α U k + 1 ( λ , γ * min ) + ( 1 α ) U k + 1 ( λ , γ * max ) .
Hence,
0 = U k + 1 ( ξ , γ * ) = α U k + 1 ( ξ , γ * min ) + ( 1 α ) U k + 1 ( ξ , γ * max ) ,
with ρ 1 = U k + 1 ( ξ , γ * min ) and ρ 2 = U k + 1 ( ξ , γ * max ) therefore either taking opposite signs, or being both zero. In the first case it is clear that one between ρ 1 and ρ 2 improves the sought root, in particular we select γ { γ * min , γ * max } corresponding to the ρ * { ρ 1 , ρ 2 } such that
1.
(Smallest negative root): The sign of ρ * is opposite to the sign of U k + 1 at .
2.
(Largest negative root): The sign of ρ * is opposite to the sign of U k + 1 at 0.
3.
(Smallest positive root): The sign of ρ * is opposite to the sign of U k + 1 at 0.
4.
(Largest positive root): The sign of ρ * is opposite to the sign of U k + 1 at + .
If, finally, both ρ 1 and ρ 2 are zero, for linearity, it means that the root is independent of this parameter and we can choose one of its extremal values.
We will now prove (in Theorem 3, and in Section 4.2) that we can predict whether γ E ( i ) = α E ( i ) or γ E ( i ) = β E ( i ) in order to obtain the intervals bounding the roots of U k .
Lemma 4.
Given a fixed k, there exist polynomials W l for any 0 l k such that
U k + 1 = U j W k j + 1 U j 1 W k j μ j ,
with μ j = γ R ( j ) ( λ + ( 1 ) j ) 2 .
Proof. 
The statement holds for j = k , by setting W 1 = η k and W 0 = 1 . Let the statement be true for j 2 , then,
U k + 1 = U j W k j + 1 U j 1 W k j μ j = = W k j + 1 ( η j 1 U j 1 μ j 1 U j 2 ) U j 1 W k j = = ( W k j + 1 η j 1 W k j ) W k j + 2 U j 1 W k j + 1 μ j 1 U j 2
It is also clear by induction that W k j depends only on the parameters with index greater than j.    □
We denote with e j { α E ( j ) , β E ( j ) } one choice of the extremal values of γ E ( j ) and, analogously, r j { α R ( j ) , β R ( j ) } ; moreover we denote by e j * and r j * the other extremal value of the parameter. We now consider the polynomial U k with γ E ( j ) e j , γ R ( j ) r j and also define U E U k j as the polynomial U k with γ E ( j ) e j * . Similarly we define U R U k j as the polynomial with γ R ( j ) = r j * . Then it is immediate from the recursion of U k that
U E U j + 1 j = U j + 1 + ( 1 ) j + 1 ( e j * e j ) U j
U R U j + 1 j = U j + 1 + [ λ U j ( λ + ( 1 ) j ) 2 U j 1 ] ( r j * r j ) .
Setting z j ( λ ) = λ U j ( λ + ( 1 ) j ) 2 U j 1 , the second equality can be rewritten simply as
U R U j + 1 j = U j + 1 + z j ( r j * r j ) .
Lemma 5.
We have the following relations
U E U k + 1 j = U k + 1 + ( 1 ) j + 1 W k j U j ( e j * e j )
U R U k + 1 j = U k + 1 + W k j z j ( r j * r j ) .
Proof. 
We write
U k + 1 = U j + 1 W k j U j W k j 1 μ j + 1 ,
and notice that the only term in the right hand side that depends on γ E ( j ) is U j + 1 . Thus, using (13),
U E U k + 1 j U k + 1 = W k j ( U E U j + 1 j U j + 1 ) = ( 1 ) j + 1 W k j U j ( e j * e j ) .
This proves the first formula. The second one is proved in the same way, using (14).    □
Theorem 3.
Let ξ be the largest (resp. smallest) positive or negative root of U k + 1 and let e j , r j be the parameters that realize this root. Then ξ assumes the largest (resp. smallest) value when the γ E ( j ) are alternatively the maximum or minimum.
Proof. 
It is enough to show that ( e j * e j ) ( e j 1 * e j 1 ) is negative. First we observe that if W k j ( ξ ) is zero, then ξ does not depend on γ E ( j ) . Thus we can assume W k j ( ξ ) , W k j 1 ( ξ ) 0 . We also assume, without loss of generality, that U j + 1 ( ξ ) , U j ( ξ ) 0 . With these assumptions, and recalling that μ j > 0 , we have
sgn ( U j W k j ) = sgn ( U j 1 W k j 1 ) .
Now we observe that U E U k + 1 j ( ξ ) must have a fixed sign as j varies to guarantee that the root ξ is either maximal or minimal (this sign depends on the type of of the root – smallest/largest, positive/negative – and on k but cannot depend on j). We call this sign s. Now we have, using (15) and recalling that U k + 1 ( ξ ) = 0 ,
sgn ( e j * e j ) ( e j 1 * e j 1 ) = sgn ( U E U k + 1 j ( ξ ) ( U E U k + 1 j 1 ( ξ ) ) ( 1 ) k j ( 1 ) j sgn ( W k j 1 U j ) sgn ( W j U j 1 ) = s 2 ( 1 ) 2 j + 1 = 1 .
   □

4.2. Choice of γ E ( k )

In the previous section we have shown that the γ E ( k ) assume alternatively the maximum or minimum value. It is enough to fix the value of γ E ( k ) to determine all the γ E ( j ) . We know, from (13) that sign ( e k * e k ) is equal to ( 1 ) k + 1 sign ( U k ( ξ ) U E U k + 1 k ( ξ ) . We will consider four cases
  • Let ξ = λ + U B = ξ s k + 1 ( k + 1 ) . We know that ξ is larger than all the roots of U k and U E U k + 1 k and so U k ( ξ ) , U E U k + 1 k ( ξ ) > 0 . Thus
    sign ( e k * e k ) = ( 1 ) k + 1
    and so
    sign ( e j * e j ) = ( e k * e k ) ( 1 ) k j = ( 1 ) k + 1 ( 1 ) k j = ( 1 ) j + 1 .
  • Let ξ = λ L B = ξ 1 ( k + 1 ) . The root ξ is negative and smaller than all the roots of U k and U E U k + 1 k and so
    sign ( U k ( ξ ) ) = ( 1 ) k sign ( U E U k + 1 k ( ξ ) ) = ( 1 ) k + 1 .
    Thus
    sign ( e k * e k ) = ( 1 ) k + 1 ( 1 ) k ( 1 ) k + 1 = ( 1 ) k
    and we have
    sign ( e j * e j ) = sign ( e k * e k ) ( 1 ) k j = ( 1 ) k ( 1 ) k j = ( 1 ) j .
  • Let now ξ = λ + L B and let m be the (smallest) index such that U m + 1 ( ξ ) = 0 . Then the sign of U m ( ξ ) and U E U m + 1 m must be the sign that this polynomial assumes on 0. Thus
    sign ( e m * e m ) = ( 1 ) m + 1 sign ( U m ( 0 ) ) sign ( U E U m + 1 m ( 0 ) ) = ( 1 ) m + 1 ( 1 ) m + 1 2 ( 1 ) m 2 = ( 1 ) m + 1 ( 1 ) m + 1 = +
    and
    sign ( e j * e j ) = sign ( e m * e m ) ( 1 ) m j = ( 1 ) m j .
  • Let finally ξ = λ U B and let m be the (smallest) index such that U m + 1 ( ξ ) = 0 . Reasoning as above shows that
    sign ( e m * e m ) = + and sign ( e j * e j ) = ( 1 ) m j .

4.3. Sign of r j * r j in Terms of the Sign of z j ( ξ )

Recall that we have
U E U k + 1 j ( ξ ) = ( 1 ) j + 1 W j + 1 ( ξ ) U j ( ξ ) ( e j * e j ) , U R U k + 1 j ( ξ ) = W j + 1 ( ξ ) z j ( ξ ) ( r j * r j ) .
If W j + 1 ( ξ ) = 0 , then the root is independent of γ R ( j ) , so we can assume W j + 1 ( ξ ) 0 . We also assume, without loss of generality, that U j ( ξ ) 0 . Then we have
sign ( r j * r j ) = sign ( z j ( ξ ) U R U k + 1 j ( ξ ) ) sign ( W j + 1 ) = sign ( z j ( ξ ) ) sign ( U R U k + 1 j ( ξ ) ) ( 1 ) j + 1 sign ( U E U k + 1 j ( ξ ) U j ( ξ ) ) sign ( e j * e j ) = ( 1 ) j + 1 sign ( U j ( ξ ) ) sign ( e j * e j ) sign ( z j ( ξ ) ) ,
where the last step follows from the fact that U E U k + 1 j ( ξ ) and U R U k + 1 j ( ξ ) must have the same sign. Again we distinguish four cases depending on the type of root we are considering.
  • Let ξ = λ + U B = ξ s k + 1 ( k + 1 ) . Then U j ( ξ ) > 0 and sign ( e j * e j ) = ( 1 ) j + 1 so
    sign ( r j * r j ) = ( 1 ) j + 1 ( 1 ) j + 1 sign ( z j ( ξ ) ) = sign ( z j ( ξ ) ) .
  • Let ξ = λ U B = ξ 1 ( k + 1 ) . Then
    sign ( U j ( ξ ) ) = ( 1 ) j and sign ( e j * e j ) = ( 1 ) j
    so
    sign ( r j * r j ) = ( 1 ) j + 1 ( 1 ) j ( 1 ) j sign ( z j ( ξ ) ) = ( 1 ) j + 1 sign ( z j ( ξ ) ) .
  • Let now ξ = λ + L B and m k be the (smallest) index such that U m + 1 ( ξ ) = 0 . Then
    sign ( U j ( ξ ) ) = ( 1 ) j 2 and sign ( e j * e j ) = ( 1 ) m j
    so
    sign ( r j * r j ) = ( 1 ) j + 1 ( 1 ) j 2 ( 1 ) m j sign ( z j ( ξ ) ) = ( 1 ) m + 1 ( 1 ) j 2 sign ( z j ( ξ ) ) .
  • Let finally ξ = λ U B and m k be the (smallest) index such that U m + 1 ( ξ ) = 0 . Reasoning as above shows that
    sign ( r j * r j ) = ( 1 ) m + 1 ( 1 ) j 2 sign ( z j ( ξ ) ) .
The sign of z j ( ξ ) cannot be computed in general.

4.4. Bounds for the Eigenvalues of the Preconditioned Matrix

A procedure to compute the intervals bounding the eigenvalues of P 1 A with N + 1 blocks can be described as follows:
  • Upper bound for the positive eigenvalues. In view of Theorem 2, this is given by the largest root of the polynomial U N + 1 ( λ ) . The sign of (17) is ( 1 ) j + 1 . When the sign of e j * e j is positive, this means that the wrong indicator is larger than the correct one, and therefore we must set γ E ( j ) = α E ( j ) . Hence, as sgn ( e 0 * e 0 ) = 1 , we should start with γ E ( 0 ) = β E ( 0 ) , and then alternating between the two extremal values: β E ( 0 ) , α E ( 1 ) , β E ( 2 ) , .
  • Lower bound for the negative eigenvalues. In view of Theorem 2, this is given by the smallest zero of the polynomial U N + 1 ( λ ) . The sign of (18) is ( 1 ) j . This means that we should start with γ E ( 0 ) = α E ( 0 ) (since e 0 * e 0 > 0 ) and then alternating between the two extremal values: α E ( 0 ) , β E ( 1 ) , α E ( 2 ) , .
  • Upper bounds for the negative eigenvalues. We consider evaluating the largest negative root ξ ( m ) of all polynomials U m , 2 m N + 1 . The sign in (19) is ( 1 ) m + j suggesting that we have to start with α E ( 0 ) if m is even and with β E ( 0 ) , if m is odd, and proceed as before. Finally we compute
    λ U B = max m N + 1 { ξ ( m ) } .
  • Lower bounds for the positive eigenvalues. We consider evaluating the smallest positive root ξ + ( m ) of all polynomials U m , 1 m N + 1 . The sign in (20) is again ( 1 ) m + j suggesting that we have to start with α E ( 0 ) if m is even and with β E ( 0 ) , if m is odd, and proceed as before. Finally we compute
    λ + L B = min m N + 1 { ξ + ( m ) } .
In each of the four cases, the choice of γ R ( j ) [ α R ( j ) , β R ( j ) ] must be performed by taking all the combinations of these values. Some insights on how to select the extremal values of the γ R indicators, upon additional assumptions, are given in Lemma 6, whose proof is deferred to Appendix A, and in Theorem 4.
Lemma 6.
Let ρ a positive number and assume that, for every j = 0 , , k , α E ( j ) 2 ρ and β E ( j ) 2 + ρ . Assume further that | ξ | > 1 ρ . Then z j ( ξ k + 1 ( k + 1 ) ) < 0 , and s g n z j ( ξ 1 ( k + 1 ) ) = ( 1 ) j .
Under the assumptions of Lemma 6 we are now able to determine the monotonicity of the extremal roots of U k + 1 depending on γ R ( j ) .
Theorem 4.
Under the assumptions of Lemma 6, the largest (respectively, the smallest) root of U k + 1 assumes its largest (respectively, smallest) value, in combination with γ R ( j ) = β R ( j ) , j = 1 , , k .
Proof. 
Let ξ be the largest positive root. Then the previous lemma shows that sign ( z j ( ξ ) ) = ( 1 ) . But then (21) implies
sign ( r j * r j ) = 1
and so r j = β R ( j ) . Now let ξ be the smallest negative root. The previous lemma shows that sign ( z j ( ξ ) ) = ( 1 ) j but then (22) implies
sign ( r j * r j ) = ( 1 ) j + 1 ( 1 ) j = 1
and so again r j = β R ( j ) .    □

5. Numerical Results. Randomly Generated Matrices

We now undertake numerical tests to validate the theoretical bounds of Theorem 1, and subsequent characterizations. We determine the extremal eigenvalues of P 1 A on randomly-generated linear systems. Specifically, we considered a simplified case with E i 0 , i > 0 and run a number of different test cases , combining the values of the extremal eigenvalues (Table 1) of the symmetric positive definite matrices involved. In particular
  • Case N = 2 . We have three parameters with 6 different endpoints of the corresponding intervals. Overall we run 3 6 = 729 test cases.
  • Case N = 3 . We have 4 parameters with 4 different endpoints of the corresponding intervals. Overall we run 4 4 = 256 test cases.
  • Case N = 4 . We have 5 parameters with 4 different endpoints of the corresponding intervals. Overall we run 5 4 = 1024 test cases.
From Figure 1, Figure 2 and Figure 3 we notice that three out of four bounds perfectly capture the eigenvalues of the preconditioned matrix, while the upper bounds for negative eigenvalues are not as tight for N = 2 , 4 and lower bounds for positive eigenvalues are less effective for N = 3 . All in all, our technique is able to predict the behavior of the MINRES iterative solver applied to the preconditioned saddle-point linear system.

5.1. Comparisons Against the Block Diagonal Preconditioner

The preconditioner we have studied so far, being SPD, allows the MINRES iteration for the iterative solution of the multiple saddle-point linear systems. Another, most known, preconditioner sharing the same properties is the block diagonal preconditioner P D , based on inexact Schur complements, defined in (1).
We compare the MINRES number of iterations in all the previously described test cases for N = 3 . In Figure 4 we plot the number of iterations with the PP preconditioner and the block diagonal preconditioner, together with the square root of κ , an indicator of the ill-conditioning of the overall saddle-point system, namely
κ = β E ( 0 ) α E ( 0 ) β R ( 1 ) β R ( 2 ) β R ( 3 ) α R ( 1 ) α R ( 2 ) α R ( 3 ) .
From the figure, we notice that the two preconditioners are both affected by the κ indicator, with the PP preconditioner more effective than P D for small to modest values of κ . When the Schur complements are instead not optimally approximated, the block diagonal preconditioner is to be preferred.

6. A Realistic Example: The 3-D Biot’s Consolidation Model

A numerical experiment, concerning the mixed form of Biot’s poroelasticity equations, is used to investigate the quality of the bounds for the eigenvalues of the preconditioned matrix and the effectiveness of the proposed triangular preconditioners on a realistic problem. For the PDEs and boundary conditions governing this model, we refer to the description in [19]. As a reference test case for the experiments, we considered the porous cantilever beam problem originally introduced in [25] and already used in [26]. The domain is the unit cube with no-flow boundary conditions along all sides, zero displacements along the left edge, and a uniform load applied on top. The material properties are described in ([19], Table 5,1). We considered the Mixed-Hybrid Finite Element discretization of the coupled Biot equations, which gives rise to a double saddle point linear system. The test case refers to a unitary cubic domain, uniformly discretized using a h = 0.05 mesh size in all the spatial dimensions. The size of the block matrices and the overall number of nonzeros in the double saddle-point system are reported in Table 2.

6.1. Handling the Case n 2 > n 1

The theory previously developed is based on the assumption n 0 n 1 n 2 , which is not verified in this problem, where n 2 = 25200 > n 1 = 8000 . The main consequence of this is that R 2 R 2 is singular and γ R ( 2 ) = 0 . This drawback can be circumvented by attacking the eigenvalue problem (3) in a different way. Let us write (3) for a double saddle-point linear system:
( E 0 λ I ) u 1 + ( 1 λ ) R 1 u 2 = 0 , ( 1 λ ) R 1 u 1 + ( E 1 λ ( I + R 1 R 1 ) ) u 2 + ( 1 + λ ) R 2 u 3 = 0 , ( 1 + λ ) R 2 u 2 + ( E 2 λ ( I + R 2 R 2 ) ) u 3 = 0 .
As in the proof of Theorem 1, we have, for all λ [ α E ( 0 ) , β E ( 0 ) ]
E 1 + λ ( I + R 1 R 1 ) ( λ 1 ) 2 R 1 ( λ I E 0 ) 1 R 1 u 2 = ( 1 + λ ) R 2 u 3 .
Assuming now that λ α E ( 2 ) 1 + β R ( 2 ) , β E ( 2 ) 1 + α R ( 2 ) I ¯ , which ensures that E 2 λ ( I + R 2 R 2 ) is SPD; we obtain u 3 from the third of (23)
u 3 = ( λ + 1 ) ( E 2 λ ( I + R 2 R 2 ) ) 1 R 2 u 2
Substituting this into (24) yields
E 1 + λ ( I + R 1 R 1 ) ( λ 1 ) 2 R 1 ( λ I E 0 ) 1 R 1 + ( 1 + λ ) 2 R 2 ( E 2 λ ( I + R 2 R 2 ) ) 1 R 2 u 2 = 0
Premultiplying now by u 2 and defining w = R 1 u 2 and z = R 2 u 2 , yields
( λ 1 ) 2 w ( E 0 λ I ) 1 w w w γ R ( 1 ) + λ ( 1 + γ R ( 1 ) ) + γ E ( 1 ) + ( 1 + λ ) 2 z ( E 2 λ ( I + R 2 R 2 ) ) 1 z z z γ ¯ R ( 2 ) = 0 .
Before proceeding we notice that γ ¯ R ( 2 ) , differently from definition (5), is the Rayleigh quotient of R 2 R 2 . Due to the fact that R 2 has full rank we can bound γ ¯ R ( 2 ) as 0 < α R ( 2 ) γ ¯ R ( 2 ) β R ( 2 ) . Hence, from now on, we will refer to γ ¯ R ( 2 ) as simply γ R ( 2 ) . Applying Lemma 1 to both E 0 λ I and E 2 λ ( I + R 2 R 2 ) , we rewrite the previous as
0 = γ R ( 1 ) ( λ 1 ) 2 λ γ E ( 0 ) + λ ( 1 + γ R ( 1 ) ) + γ E ( 1 ) ( 1 + λ ) 2 γ R ( 2 ) 1 λ ( 1 + γ R ( 2 ) ) γ E ( 2 ) = U 2 ( λ ) U 1 ( λ ) ( 1 + λ ) 2 γ R ( 2 ) λ ( 1 + γ R ( 2 ) ) γ E ( 2 ) = U 3 ( λ ) U 1 ( λ ) ( λ ( 1 + γ R ( 2 ) ) γ E ( 2 ) ) ,
which tells that any eigenvalue of P 1 A outside [ α E ( 0 ) , β E ( 0 ) ] I ¯ is a zero of U 3 ( λ ) , meaning that any eigenvalue is in I 1 I ¯ I 3 .

6.2. Verification of the Bounds

To approximate the ( 1 , 1 ) block, we employed the classical incomplete Cholesky factorization (IC) with fill-in based on a drop tolerance δ . The approximations S ^ 1 and S ^ 2 of the Schur complements are as in [19], which prove completely scalable with the problem size. Even if the IC preconditioner does not scale with the discretization parameter, it is helpful to conduct a spectral analysis as a function of the drop tolerance.
In Table 3 we report the extremal values of the five indicators for this problem.
We computed the relevant eigenvalues of the preconditioned matrix P 1 A , using the Matlab function eigs with a function handle for the application of the preconditioned matrix (which we could not compute explicitly, due to its size and density) as well as the four endpoints of } } I 3 , as predicted by the theory, and reported all of them in Table 4.

6.3. Improving the Upper Bounds for the Negative Eigenvalues

From Table 4 we see that the bounds perfectly capture the extremal eigenvalues, except for the upper bound for the negative eigenvalues, which is still loose. However, this can be improved by taking into account the indicator γ S ( k ) ( w ) , defined in (7). The endpoints of this indicator are, for this test case:
I S 1 = [ 0.3187 , 1.2408 ] , I S 2 = [ 0.9998 , 1.0002 ] ,
for every value of δ . In order to use these indicators in the eigenvalue analysis, we perform a change of variable, using (6), in the polynomial recurrence which reads as, for k 1 ,
U k + 1 ( λ ) = ( λ ( 1 + γ S ( k ) γ E ( k ) ) + ( 1 ) k + 1 γ E ( k ) ) U k ( λ ) + ( γ E ( k ) γ S ( k ) ) ( λ + ( 1 ) k ) 2 U k 1 ( λ ) ,
for which all the results obtained in Section 3 still hold, as well as the observation at the beginning of Section 4.1, stating that the bounds for the roots of U k + 1 are taken at the endpoints of the parameters. On the contrary, in this case, we cannot select which of the extremal value of γ E ( k ) and γ S ( k ) provides the bound. However, in view of the low degree of the polynomial we can compute its roots for every combination of the endpoints of the indicators, and take the minimum/maximum values of the roots, depending on the bounds sought. To do this we also should consider the following
Remark 1.
The definition of γ S ( k ) ( w ) , constrains this indicator to be larger than the corresponding γ E ( k ) ( w ) . Hence in correspondence of γ E ( k ) = β E ( k ) , the lower bound for γ S ( k ) is min { β E ( k ) , α S ( k ) } .
Taking this into account, we compute the 3 roots of
U 3 ( λ ; γ E ( 0 ) , γ E ( 1 ) , γ E ( 2 ) , γ S ( 1 ) , γ S ( 2 ) ) ,
for the 2 5 = 32 combinations of the extremal values of the indicators, and call ξ L B , ξ U B , ξ + L B , ξ + U B the extremal values of the roots. In view of the discussion in Section 6.3, we finally compute as the eigenvalue bounds the following expressions:
λ LB = ξ L B , λ UB = ξ U B λ + LB = min α E ( 0 ) , α E ( 2 ) 1 + β R ( 2 ) , ξ + L B , λ + UB = max β E ( 0 ) , β E ( 2 ) 1 + α R ( 2 ) , ξ + U B
which we report in Table 5, together with the exact eigenvalues.
The adherence of the bounds to the extremal eigenvalues is now impressive.

6.4. Comparisons with the Block Diagonal Preconditioner for the Biot Problem

We finally compare the block diagonal with the PP preconditioners in terms of number of iterations, for different values of the parameter δ , and report the obtained results in Table 6, showing that the preconditioner analyzed in this work represents a valid alternative to the block diagonal preconditioner, especially when the Schur complements are all well approximated. Figure 5 compare the convergence profile of MINRES with both preconditioners for δ = 10 5 , 10 6 .
A final remark is in order: In this study we have not been concerned with the cost of the application of the two preconditioners, at each MINRES iteration, which is crucial to evaluate the performance of both. This comparison is outside the scope of this paper, which is mainly concerned with eigenvalue distribution and number of iterations. However, even taking into account that the application of the PP preconditioner is from 1.5 to (slightly less than) 2 times more expensive than that of the block diagonal preconditioner, our results show that the PP preconditioner can be a viable alternative to accelerate the MINRES solution of multiple saddle-point linear systems.

7. Conclusions

We have developed an eigenvalue analysis of the preconditioned multiple saddle-point linear systems, with the SPD preconditioner (PP in short) proposed by Pearson and Potschka in [1]. This analysis is based on the (approximate) knowledge of the smallest and the largest eigenvalues of a number of SPD matrices, related to the preconditioned blocks of the original system. The obtained bounds reveal very close to the exact extremal eigenvalues of the preconditioned matrix, and therefore, this study represents a valuable tool to predict with some accuracy the number of MINRES iterations to solve the multiple saddle-point linear system. In a number of both synthetic and realistic test cases, the PP preconditioner reveals more efficient than the block diagonal preconditioner when the Schur complement matrices are well approximated.

Acknowledgments

LB acknowledges financial support under the PNRR research activity, Mission 4, Component 2, Investment 1.1, funded by the EU Next-GenerationEU – #2022AKNSE4_005 (PE1) CUP C53D2300242000. LB is member of the INdAM research group GNCS.

Appendix A. Proof of Lemma 6

Proof. 
We start with the definition of z j ( λ ) and apply the recursion:
z j ( λ ) = λ U j ( λ ) ( λ + ( 1 ) j ) 2 U j 1 ( λ ) = λ U j 1 ( λ ) ( λ ( 1 + γ R ( j 1 ) ) + ( 1 ) j γ E ( j 1 ) ) γ R ( j 1 ) ( λ + ( 1 ) j 1 ) 2 U j 2 ( λ ) ( λ + ( 1 ) j ) 2 U j 1 ( λ ) .
We now apply the definition to z j 1 ( λ ) :
z j 1 ( λ ) = λ U j 1 ( λ ) ( λ + ( 1 ) j 1 ) 2 U j 2 ( λ ) .
Combining these two expressions we get
z j ( λ ) γ R ( j 1 ) λ z j 1 ( λ ) = U j 1 ( λ ) λ 2 ( 1 + γ R ( j 1 ) ) + ( 1 ) j λ γ E ( j 1 ) ( λ + ( 1 ) j ) 2 λ 2 γ R ( j 1 ) = U j 1 ( λ ) ( 1 ) j λ γ E ( j 1 ) + 2 ( 1 ) j + 1 λ 1 = U j 1 ( λ ) ( 1 ) j + 1 λ ( 2 γ E ( j 1 ) ) + ( 1 ) j U j 1 ( λ ) ( 1 ) j + 1 t j ( λ ) .
We know that z 1 ( λ ) = t 1 ( λ ) and we have just proved that
z j ( λ ) = λ γ R ( j 1 ) z j 1 ( λ ) + ( 1 ) j + 1 U j 1 ( λ ) t j ( λ ) .
Sign of z j ( ξ ) for the largest positive root. Now let ξ ξ s k + 1 ( k + 1 ) be the largest positive root. In such a case the previous analysis shows that γ E ( j 1 ) = α E ( j 1 ) if j is even β E ( j 1 ) if j is odd . Using now the hypotheses, we have that
sgn ( t j ( ξ ) ) = sgn ( 2 γ E ( j 1 ) ) = ( 1 ) j .
Now we prove by induction that sgn ( z j ( ξ ) ) = 1 for all j. The base step is the following
sgn ( z 1 ( ξ ) ) = sgn ( t 1 ( ξ ) ) = ( 1 ) 1 = 1 .
We know that
z j ( ξ ) = λ γ R ( j 1 ) z j 1 ( ξ ) + ( 1 ) j + 1 U j 1 ( ξ ) t j ( ξ ) .
The first term is negative by induction hypothesis. As for the second term
sgn ( ( 1 ) j + 1 U j 1 ( ξ ) t j ( ξ ) = ( 1 ) j + 1 ( + 1 ) ( 1 ) j = 1 .
Thus z j is negative as well.
Sign of z j ( ξ ) for the smallest negative root. Let ξ ξ 1 ( k + 1 ) be the smallest negative root. By assumptions, ξ > 1 ρ . Moreover, we know that γ E ( j 1 ) = β E ( j 1 ) if j is even α E ( j 1 ) if j is odd . Then
sgn ( t j ( ξ ) ) = sgn ( 2 γ E ( j 1 ) ) = ( 1 ) j .
Now we prove by induction that sgn ( z j ( ξ ) ) = ( 1 ) j . The base step is the following:
sgn ( z 1 ( ξ ) ) = sgn ( t 1 ( ξ ) ) = 1 .
Now we assume that z j 1 ( ξ ) has sign ( 1 ) j 1 . We know that
z j = λ γ R ( j 1 ) z j 1 ( ξ ) + ( 1 ) j + 1 U j 1 ( ξ ) t j ( ξ ) .
The first term has sign ( 1 ) j by induction hypothesis (since ξ is negative). As for the second term
sgn ( ( 1 ) j + 1 U j 1 ( ξ ) t j ( ξ ) ) = ( 1 ) j + 1 ( 1 ) j 1 ( 1 ) j = ( 1 ) j .
Thus z j ( ξ ) has sign ( 1 ) j . □

References

  1. Pearson, J.W.; Potschka, A. On symmetric positive definite preconditioners for multiple saddle-point systems. IMA J. Numer. Anal. 2024, 44, 1731–1750.
  2. Bergamaschi, L.; Martinez, A.; Pearson, J.W.; Potschka, A. Spectral analysis of block preconditioners for double saddle-point linear systems with application to PDE-constrained optimization. Computational Optimization with Applications 2024, 91, 423–455.
  3. Rhebergen, S.; Wells, G.N.; Wathen, A.J.; Katz, R.F. Three-field block preconditioners for models of coupled magma/mantle dynamics. SIAM J. Sci. Comput. 2015, 37, A2270–A2294.
  4. Ramage, A.; Gartland, E.C. A Preconditioned Nullspace Method for Liquid Crystal Director Modeling. SIAM Journal on Scientific Computing 2013, 35, B226–B247.
  5. Greif, C.; He, Y. Block Preconditioners for the Marker-and-Cell Discretization of the Stokes-Darcy Equations. SIAM Journal on Matrix Analysis and Applications 2023, pp. 1540–1565.
  6. Chidyagwai, P.; Ladenheim, S.; Szyld, D.B. Constraint preconditioning for the coupled Stokes-Darcy system. SIAM J. Sci. Comput. 2016, 38, A668–A690.
  7. Beik, F.P.A.; Benzi, M. Preconditioning techniques for the coupled Stokes-Darcy problem: spectral and field-of-values analysis. Numer. Math. 2022, 150, 257–298.
  8. Greif, C. A BFBt preconditioner for Double Saddle-Point Systems. IMA Journal of Numerical Analysis 2026. to appear, . [CrossRef]
  9. Bakrani Balani, F.; Hajarian, M.; Bergamaschi, L. Two block preconditioners for a class of double saddle point linear systems. Applied Numerical Mathematics 2023, 190, 155 – 167.
  10. Bakrani Balani, F.; Bergamaschi, L.; Martínez, A.; Hajarian, M. Some preconditioning techniques for a class of double saddle point problems. Numerical Linear Algebra with Applications 2024, 31, e2551.
  11. Beik, F.P.A.; Benzi, M. Iterative methods for double saddle point systems. SIAM J. Matrix Anal. Appl. 2018, 39, 902–921.
  12. Sogn, J.; Zulehner, W. Schur complement preconditioners for multiple saddle point problems of block tridiagonal form with application to optimization problems. IMA Journal of Numerical Analysis 2018, 39, 1328–1359.
  13. Bradley, S.; Greif, C. Eigenvalue bounds for double saddle-point systems. IMA J. Numer. Anal. 2023, 43, 3564–3592.
  14. Pearson, J.W.; Potschka, A. Double saddle-point preconditioning for Krylov methods in the inexact sequential homotopy method. Numerical Linear Algebra with Applications 2024, 31, e2553.
  15. Bergamaschi, L.; Martinez, A.; Pearson, J.W.; Potschka, A. Eigenvalue bounds for preconditioned symmetric multiple saddle-point matrices. Linear Algebra with Applications 2026. Published online January 16, 2026, . [CrossRef]
  16. Bergamaschi, L.; Martínez, A.; Pilotto, M. Spectral Analysis of Block Diagonally Preconditioned Multiple Saddle-Point Linear Systems with Inexact Schur Complements 2026. In preparation.
  17. Beigl, A.; Sogn, J.; Zulehner, W. Robust preconditioners for multiple saddle point problems and applications to optimal control problems. SIAM Journal on Matrix Analysis and Applications 2020, 41, 1590–1615.
  18. Ferronato, M.; Franceschini, A.; Janna, C.; Castelletto, N.; Tchelepi, H. A general preconditioning framework for coupled multi-physics problems. J. Comput. Phys. 2019, 398, 108887. [CrossRef]
  19. Bergamaschi, L.; Ferronato, M.; Martínez, A. Triangular preconditioners for double saddle point linear systems arising in the mixed form of poroelasticity equations. SIAM Journal on Matrix Analysis and Applications 2026. Accepted on October 14, 2025, . [CrossRef]
  20. Paige, C.C.; Saunders, M.A. Solution of Sparse Indefinite Systems of Linear Equations. SIAM J. on Numer. Anal. 1975, 12, 617–629. [CrossRef]
  21. Saad, Y.; Schultz, M.H. GMRES: a generalized minimal residual algorithm for solving nonsymmetric linear systems. SIAM J. Sci. Statist. Comput. 1986, 7, 856–869. [CrossRef]
  22. Fischer, B. Polynomial based iteration methods for symmetric linear systems; Wiley-Teubner Series Advances in Numerical Mathematics, John Wiley & Sons, Ltd., Chichester; B. G. Teubner, Stuttgart, 1996; p. 283. [CrossRef]
  23. Greenbaum, A. Iterative methods for solving linear systems; Vol. 17, Frontiers in Applied Mathematics, Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA, 1997; pp. xiv+220. [CrossRef]
  24. Bergamaschi, L. On Eigenvalue distribution of constraint-preconditioned symmetric saddle point matrices. Numer. Linear Algebra Appl. 2012, 19, 754–772. [CrossRef]
  25. Phillips, P.J.; Wheeler, M.F. Overcoming the problem of locking in linear elasticity and poroelasticity: A heuristic approach. Computational Geosciences 2009, 13, 5–12.
  26. Frigo, M.; Castelletto, N.; Ferronato, M.; White, J.A. Efficient solvers for hybridized three-field mixed finite element coupled poromechanics. Computers and Mathematics with Applications 2021, 91, 36 – 52. [CrossRef]
Figure 1. Double saddle-point linear system. Extremal eigenvalues of the preconditioned matrix (blue dots) and bounds (red line) after 10 runs with each combination of the parameters from Table 1.
Figure 1. Double saddle-point linear system. Extremal eigenvalues of the preconditioned matrix (blue dots) and bounds (red line) after 10 runs with each combination of the parameters from Table 1.
Preprints 197224 g001
Figure 2. Multiple saddle-point linear system with N = 3 . Extremal eigenvalues of the preconditioned matrix (blue dots) and bounds (red line) after 10 runs with each combination of the parameters from Table 1.
Figure 2. Multiple saddle-point linear system with N = 3 . Extremal eigenvalues of the preconditioned matrix (blue dots) and bounds (red line) after 10 runs with each combination of the parameters from Table 1.
Preprints 197224 g002
Figure 3. Multiple saddle-point linear system with N = 4 . Extremal eigenvalues of the preconditioned matrix (blue dots) and bounds (red line) after 10 runs with each combination of the parameters from Table 1.
Figure 3. Multiple saddle-point linear system with N = 4 . Extremal eigenvalues of the preconditioned matrix (blue dots) and bounds (red line) after 10 runs with each combination of the parameters from Table 1.
Preprints 197224 g003
Figure 4. Iteration numbers of P and P D for all the tests with N = 3 . The condition number indicator κ is also displayed.
Figure 4. Iteration numbers of P and P D for all the tests with N = 3 . The condition number indicator κ is also displayed.
Preprints 197224 g004
Figure 5. 3D cantilever beam problem. Convergence profiles of MINRES preconditioner with P D and P . Drop tolerance Cholesky parameter: δ = 10 5 (left), δ = 10 6 (right).
Figure 5. 3D cantilever beam problem. Convergence profiles of MINRES preconditioner with P D and P . Drop tolerance Cholesky parameter: δ = 10 5 (left), δ = 10 6 (right).
Preprints 197224 g005
Table 1. Extremal eigenvalues of the relevant symmetric positive definite matrices used in the verification of the bounds.
Table 1. Extremal eigenvalues of the relevant symmetric positive definite matrices used in the verification of the bounds.
α E ( 0 ) , α R ( 1 ) , α R ( 2 ) 0.1 0.3 0.9 α E ( 0 ) , α R ( 1 ) , α R ( N ) 0.1 0.8
β E ( 0 ) , β R ( 1 ) , β R ( 2 ) 1.2 1.8 3 β E ( 0 ) , β R ( 1 ) , β R ( N ) 1.2 2
Case N = 2 Cases N > 2
Table 2. 3D cantilever beam problem on a unit cube: size and number of nonzeros of the test matrices.
Table 2. 3D cantilever beam problem on a unit cube: size and number of nonzeros of the test matrices.
1 / h n 0 n 1 n 2 n 0 + n 1 + n 2 nonzeros
20 27783 8000 25200 60983 2.4 × 10 6
Table 3. Extremal values of the indicators for different values of the threshold parameter δ .
Table 3. Extremal values of the indicators for different values of the threshold parameter δ .
δ = 10 4 δ = 10 5 δ = 10 6 δ = 10 7
I E 0 [ 0.0422 , 1.2036 ] [ 0.2563 , 1.2232 ] [ 0.9600 , 1.0118 ] [ 0.9976 , 1.0001 ]
I E 1 [ 5 × 10 4 , 0.0889 ]
I E 2 [ 0.9976 , 1.0001 ]
I R 1 [ 2 × 10 4 , 0.7790 ]
I R 2 [ 3 × 10 5 , 0.0023 ]
Table 4. 3D cantilever beam problem on a unit cube: eigenvalue bounds vs real eigenvalues.
Table 4. 3D cantilever beam problem on a unit cube: eigenvalue bounds vs real eigenvalues.
μ LB μ UB μ + LB μ + UB
δ = 10 4 Bounds -2.6786 -0.0012 0.0424 1.2216
Exact Eigvs -1.2408 -0.3192 0.0453 1.2055
δ = 10 5 Bounds -2.4104 -0.0012 0.2564 1.2447
Exact Eigvs -1.2408 -0.3192 0.2895 1.2252
δ = 10 6 Bounds -1.7001 -0.0012 0.9600 1.0119
Exact Eigvs -1.2408 -0.3192 0.9600 1.0118
δ = 10 7 Bounds -1.6701 -0.0012 0.9976 1.0070
Exact Eigvs -1.2408 -0.3190 0.9979 1.0013
Table 5. Refined eigenvalue bounds vs real eigenvalues.
Table 5. Refined eigenvalue bounds vs real eigenvalues.
λ LB λ UB λ + LB λ + UB
δ = 10 4 Bounds -2.8268 -0.2629 0.0422 1.2274
Exact Eigvs -1.2408 -0.3192 0.0453 1.2055
δ = 10 5 Bounds -2.4204 -0.2583 0.1809 1.2517
Exact Eigvs -1.2408 -0.3192 0.2895 1.2252
δ = 10 6 Bounds -1.2913 -0.2951 0.9593 1.0119
Exact Eigvs -1.2408 -0.3192 0.9600 1.0118
δ = 10 7 Bounds -1.2434 -0.3174 0.9979 1.0029
Exact Eigvs -1.2408 -0.3190 0.9979 1.0013
Table 6. Number of MINRES iterations for the block diagonal and the PP preconditioner, for different values of the threshold parameter δ . With δ = 0 we indicate that the exact Cholesky factor of A is used for preconditioning.
Table 6. Number of MINRES iterations for the block diagonal and the PP preconditioner, for different values of the threshold parameter δ . With δ = 0 we indicate that the exact Cholesky factor of A is used for preconditioning.
δ 10 4 10 5 10 6 10 7 0
its (diag) 132 83 64 63 63
its (PP) 99 61 31 23 10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated