Preprint
Article

This version is not peer-reviewed.

Chordal Metric Formula Between Generalized Singular Values of Grassmann Matrix Pairs by Riemannian Optimization Models

Submitted:

24 February 2025

Posted:

25 February 2025

You are already at the latest version

Abstract
In this paper, we present a new explicit formula for the sum of the chordal distance between the generalized singular values of Grassmann matrix pairs, based on Riemannian optimization models. The new formulas involve small-size unitary matrices and real orthogonal matrices derived from Riemannian optimization models. We then apply Newton’s method on Riemannian manifolds to efficiently solve the variable matrices involved. The new explicit formulas provide significant improvements over the existing theoretical and computational upper bounds.
Keywords: 
;  ;  ;  

0. Introduction

The generalized singular value decomposition (GSVD) is useful in matrix computation problems and practical applications, such as the general Gauss-Markov linear model, the Kronecker canonical form of a general matrix pencil, generalized total least squares problem, gene data analysis and so on [1,2,3,4,5,6,8,9,10,11,12]. In particular, GSVD and generalized singular values (GSVs) can provide a comparative mathematical framework for two genome-scale expression data sets as seen in [1,2,3,4,22].

0.1. GSVD and Its Applications

More precisely, GSVD induces a linear transformation of two data sets from the two genes × arrays spaces to two reduced and diagonalized “genelets” × “arraylets” spaces. A single microarray probes the relative expression levels of p genes in a single sample. A series of q 1 arrays probes the genome-scale expression levels in q 1 different samples, i.e., under q 1 different experimental conditions. Let matrix A of size p-genes × q 1 -arrays, denote the full expression data, whose k-th row is the expression of the k-th gene across the different samples that correspond to the different arrays. Let the matrix B of size t-genes × q 2 -arrays, denote the relative expression levels of t genes under q 2 = q 1 = q < max { p , t } experimental conditions that correspond one to one to the q 1 conditions underlying A. GSVD induces simultaneous linear transformation of two expression data sets A and B from two p-genes ×q-arrays and t-genes ×q-arrays spaces to two reduced q-genelets ×q-arraylets spaces.
Let A and B have the following GSVD: A = U Σ A R 1 , B = V Σ B R 1 . In these spaces the data is represented by the diagonal non-negative matrices Σ A and Σ B and denote k | Σ A | l = ε 1 , l δ k l , k | Σ B | l = ε 2 , l δ k l for all 1 k , l q , where δ k l = 1 if k = l and δ k l = 0 if k l . By k | Σ A we denote the k-th row of the matrix Σ A , which lists the expression of the k-th gene across the different samples that correspond to the different arrays. By Σ A | l we denote the l-th column of the matrix Σ A , which lists the genome-scale expression measured by the l-th array. The antisymmetric angular distance between the data sets [3], θ l = arctan ( ε 1 , l / ε 2 , l ) π / 4 indicates the relative significance of the l-th genelet, i.e., its significance in the first data set relative to that in the second in terms of the ratio of the expression information captured by this genelet in the first data set to that in the second. An angular distance of 0 indicates a genelet of equal significance in both data sets, with ε 1 , l = ε 2 , l .
Alter et al. [1] made comparative analysis of gene mRNA expression data sets by transforming those into GSVs and analyzed the degree of change for reduced and diagonalized “genelets” × “arraylets” spaces in gene expression levels from samples under different experimental conditions. This behavior is significant for guiding further experimental research. Motivated by practical applications proximity and range size between GSVs, Xu et al. [22] considered the upper bound of sum of chordal metric between generalized singular values of Grassmann matrix pairs and applied this result in comparative analysis of gene mRNA expression data.

0.2. Riemannian Optimization

Optimization on Riemannian manifolds is called Riemannian optimization. Unconstrained optimization methods on the Euclidean space, such as the steepest descent, the conjugate gradient, and Newton’s methods are generalized to those on a Riemannian manifold. The term “optimization on manifolds” is applied to algorithms that preserve manifold constraints during iterations. Generally speaking, preserving constraints has advantages in many situations, for example, when the cost functions are undefined or of little use outside the feasible region and when the algorithm is terminated prior to convergence yet a feasible solution is required. Theories and methods for optimization on manifolds date back to the 1970s, and algorithms for specific problems, e.g., various eigenvalue problems, appeared even earlier. Most of the general-purpose algorithms, however, did not appear until the 1990s [30]. While it is trivial to generate trial points in Rn along straight search lines, it is not as easy to do so in the curved manifold. A natural choice is the geodesic, which is the analogof straightline: it has the shortest length between two points in the manifold. For more details we may refer reders to [29,30,31,32].
In this paper we provide explicit formula of sum of chordal metric between generalized singular values of Grassmann matrix pairs with Riemannian optimization models, which absolutely improves the existing upper bounds in [22] in theory. The new formula involves two small-size unitary variable matrices from Riemannian optimization models. Then we use Newton methods on Riemannian manifolds to solve the involved Riemannian optimization models for deducing these two variable matrices solutions efficiently. The contributions of this paper are summarized as follows.
  • On theoretic aspect we provide a new explicit formula of sum of chordal metric between generalized singular values of Grassmann matrix pairs with Riemannian optimization models instead of the existing upper bounds.
  • On algorithm aspect the new expression formula involves only two small-size unitary variable matrices from Riemannian optimization models and Newton methods on Riemannian manifolds are given, which reduces the required computational cost.

0.3. Notations

Throughout this paper, by R , C , R n , C m × n , U n and OR n we denote the sets of real numbers, complex numbers, real vectors of order n, m × n matrices with entries in C , and n × n unitary matrices and real orthogonal matrices, respectively. The symbols | · | and Re ( · ) stand for absolute value and real part of a complex number, respectively. The symbols I n and O m × n represent the identity matrix of order n and m × n zero matrix, respectively. For a square matrix A = ( a i j ) C n × n , by A T , A H , A 1 , and tr ( A ) we denote the transpose, conjugate transpose, inverse, and trace of matrix A, respectively. We use · 2 to denote the spectral norm of a matrix. The singular value set of A is denoted by σ ( A ) . The complex number a + b ı ( a , b R ) and the real number pair ( a , b ) can establish a one-to-one correspondence with a point on the coordinate plane, so that the coordinate plane that establishes a one-to-one correspondence with the whole complex number is called Gauss plane, which is denoted by P ( 1 , 1 ) . For given matrices A , B C n × n , A < ( ) B means B A is a positive (semi-)definite matrix. As in [9,15,16,17], we also use the chordal metric on the Riemann sphere to measure the difference between two GSVs. The pair ( α , β ) can be regarded as a point in Gauss plane P ( 1 , 1 ) . For two points ( α , β ) ( 0 , 0 ) , ( γ , δ ) ( 0 , 0 ) in Gauss plane P ( 1 , 1 ) , their chordal distance is defined by
ρ ( ( α , β ) , ( γ , δ ) ) = | α δ β γ | ( | α | 2 + | β | 2 ) ( | γ | 2 + | δ | 2 )
to measure the difference between two points ( α , β ) and ( γ , δ ) . Next, we outline some definitions related to GSVD. Readers may refer to [7,9,16] for these definitions.
Definition 1. 
The matrix pair { A , B } with A , B C n × n is said to be regular if there exists ζ C such that det ( A + ζ B ) 0 . Denote by
G 1 , 2 = { ( α , β ) ( 0 , 0 ) : α , β C } .
The pair ( α , β ) G 1 , 2 is called a generalized eigenvalue of a regular pair { A , B } , if det ( β A α B ) = 0 .
The set of generalized eigenvalues of the regular pair { A , B } is denoted by λ ( A , B ) .
Definition 2. 
Let A C m × n and B C p × n . A matrix pair { A , B } is an ( m , p , n ) -Grassman matrix pair (GMP) if rank ( A T , B T ) T = n .
For GSVD, it has several formulations in the literature. In this paper we adopt the following form as in [9,16,19].
Definition 3. 
Let {A,B} be an (m,p,n)-GMP. Then there exist unitary matrices U C m × m , V C p × p , and a nonsingular matrix R C n × n such that
U H A R = Σ A , V H B R = Σ B ,
Σ A = Λ O ( m r s ) × ( n r s ) , Σ B = O ( p + r n ) × r Ω ,
where O ( m r s ) × ( n r s ) and O ( p + r n ) × r are zero matrices, and
Λ = diag ( α 1 , , α r + s ) , Ω = diag ( β r + 1 , , β n ) ,
with
1 = α 1 = = α r > α r + 1 α r + s > α r + s + 1 = = α n = 0 ,
0 = β 1 = = β r < β r + 1 β r + s < β r + s + 1 = = β n = 1 ,
and
α i 2 + β i 2 = 1 , 1 i n .
We note that by Definition 3, given another ( m , p , n ) -GMP { A ˜ , B ˜ } , there also exist unitary matrices U ˜ C m × m , V ˜ C p × p , nonsingular R ˜ C n × n , and nonnegative diagonal matrices Σ A ˜ C m × n , Σ B ˜ C p × n for the GSVD of { A ˜ , B ˜ } .
For GSVD of a real matrix pair { A , B } we adopt the following definition form as in [9,16,19].
Definition 4. 
Let A R m × n , B R p × n have rank ( A T , B T ) = n . Then there exist two orthogonal matrices U R m × m , V R p × p and a nonsingular matrix R R n × n such that
U T A R = Σ A , V T B R = Σ B ,
Σ A = Λ O ( m r s ) × ( n r s ) , Σ B = O ( p + r n ) × r Ω ,
where Λ = diag ( α 1 , , α r + s ) and Ω = diag ( β r + 1 , , β n ) with
1 = α 1 = = α r > α r + 1 α r + s > α r + s + 1 = = α n = 0 , 0 = β 1 = = β r < β r + 1 β r + s < β r + s + 1 = = β n = 1 ,
and α i 2 + β i 2 = 1 for 1 i n .

0.4. Overview of Existing Results about Chordal Metric Between GSVs of GMP

There have been a large amount of papers in various topics related to GSVD and its applications, e.g., see [5,6,8,9,10,11,12,16,17,18,19,23]. One of interesting and important topics is about chordal metric between GSVs of Grassmann matrix pairs (GMPs), which was first analyzed by Sun [16] in 1983. In [16], Sun presented the Hoffman-Wielandt theorem for GSVs of GMPs and gave bounds on perturbations of GSVs, which generalized several well-known results for the standard singular value problem. Since then, Paige and Saunders [15] in 1981 gave a bound for GSV variations. Li [9] in 1993 presented several perturbation bounds of GSVs and its associated subspace. Recently Xu et al. [22] considered sharp upper bounds of the chordal metric between generalized singular values of Grassmann matrix pairs. The following results provide perturbation bounds based on the chordal metric of GSVs.
Theorem 1 
(The Hoffman-Wielandt type theorem for GSV of GMP[16]). Let { A , B } and { A ˜ , B ˜ } be two ( m , p , n ) -GMPs, σ { A , B } = { ( α i , β i ) } i = 1 n and σ { A ˜ , B ˜ } = { ( α ˜ i , β ˜ i ) } i = 1 n . Then
i = 1 n 1 ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) 1 d 2 ( Y , Y ˜ ) ,
i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) n 1 1 d 2 ( Y , Y ˜ ) n ,
where
Y = A B , Y ˜ = A ˜ B ˜ ,
ρ ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) is given by (1) and d ( Y , Y ˜ ) = 1 | det ( Y H Y ˜ ) | 2 det ( Y H Y ) det ( Y ˜ H Y ˜ ) 1 2 .
Theorem 2 
(Explicit expression and sharp upper bound[22]). Let { A , B } and { A ˜ , B ˜ } be two ( m , p , n ) -GMPs, σ { A , B } = { ( α i , β i ) } i = 1 n and σ { A ˜ , B ˜ } = { ( α ˜ i , β ˜ i ) } i = 1 n . Then
i = 1 n 1 ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) max Ψ 1 U m , Ψ 2 U p ( 1 d 2 ( Y , Y ˜ Ψ ) ) = 1 d 2 ( Y , Y ˜ Ψ 0 ) ,
i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) n 1 max Ψ 1 U m , Ψ 2 U p 1 d 2 ( Y , Y ˜ Ψ ) n = n 1 1 d 2 ( Y , Y ˜ Ψ 0 ) n ,
where
Y = A B , Y ˜ = A ˜ B ˜ , Y ˜ Ψ = Ψ Y ˜ , Y ˜ Ψ 0 = Ψ 0 Y ˜ ,
Ψ = Ψ 1 O m × p O p × m Ψ 2 , Ψ 0 = U U ˜ H O m × p O p × m V V ˜ H ,
and Ψ 1 U m , Ψ 2 U p ; U , U ˜ , V , V ˜ are the unitary matrices in GSVD of { A , B } and { A ˜ , B ˜ } (refer to (2)-(3)). The inequality (9) is an equality if and only if ρ ( ( α 1 , β 1 ) , ( α ˜ 1 , β ˜ 1 ) ) = ρ ( ( α 2 , β 2 ) , ( α ˜ 2 , β ˜ 2 ) ) = = ρ ( ( α n , β n ) , ( α ˜ n , β ˜ n ) ) .
In this paper we continue along this line and further consider explicit expression of i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) instead of the existing upper bound in (9).

0.5. Organization

The rest of this paper is organized as follows. Section 1 is the preliminary. In Section 2, we present an explicit expression of i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) by Riemannian optimization models and the corresponding numerical methods. Finally, in Section 3 concluding remarks are drawn.

1. Preliminaries

In this section we propose some useful lemmas in order to deduce the main results.
Lemma 1. 
[21]Let A 1 , , A m C n × n with σ 1 ( A j ) σ 2 ( A j ) σ n ( A j ) 0 the singular values arranged in decreasing order and c R , j = 1 , , m . We have the following extreme values:
max U 1 , , U m U n | tr ( c I n ± j = 1 m U j A j ) | = n c + i = 1 n j = 1 m σ i ( A j ) .
Let ( A , B ) and ( A ˜ , B ˜ ) be two ( m , p , n ) GMPs and
S 1 = ( Z H Z ) 1 / 2 A H , m n , m p ; A ( Z H Z ) 1 / 2 , n m , m p ; ( Z H Z ) 1 / 2 B H , p n , p m ; B ( Z H Z ) 1 / 2 , n p , p m ;
S 2 = ( Z ˜ H Z ˜ ) 1 / 2 A ˜ H , m n , m p ; A ˜ ( Z ˜ H Z ˜ ) 1 / 2 , n m , m p ; ( Z ˜ H Z ˜ ) 1 / 2 B ˜ H , p n , p m ; B ˜ ( Z ˜ H Z ˜ ) 1 / 2 , n p , p m
and
M ( Φ ) = Φ H S 2 H S 2 Φ S 1 H S 1 , Φ U min { m , p , n } .
Let ( A , B ) and ( A ˜ , B ˜ ) be two ( m , p , n ) real matrix pairs defined in (4) and
S 1 = ( Z T Z ) 1 / 2 A T , m n , m p ; A ( Z T Z ) 1 / 2 , n m , m p ; ( Z T Z ) 1 / 2 B T , p n , p m ; B ( Z T Z ) 1 / 2 , n p , p m ;
S 2 = ( Z ˜ T Z ˜ ) 1 / 2 A ˜ T , m n , m p ; A ˜ ( Z ˜ T Z ˜ ) 1 / 2 , n m , m p ; ( Z ˜ T Z ˜ ) 1 / 2 B ˜ T , p n , p m ; B ˜ ( Z ˜ T Z ˜ ) 1 / 2 , n p , p m
and
M ( Φ ) = Φ T S 2 T S 2 Φ S 1 T S 1 , Φ OR min { m , p , n } .
We set
Q i = diag ( I i , O ( min { m , p , n } i ) × ( min { m , p , n } i ) ) , m n , m p ; n m , m p ; diag ( O ( min { m , p , n } i ) × ( min { m , p , n } i ) , I i ) , p n , p m ; n p , p m
for 1 i min { m , p , n } .
Lemma 2. 
[24] Let f ( X ) : C n × n R be an analytical function of several complex variables on the domain X X H I n . Then f ( X ) attains its maximum modulus on the characteristic manifold { X C n × n : X X H = I n } .
Lemma 3. 
Let { A , B } and { A ˜ , B ˜ } be two ( m , p , n ) -GMPs and let σ { A , B } = { ( α i , β i ) , i = 1 , 2 , n } and σ { A ˜ , B ˜ } = { ( α ˜ j , β ˜ j ) , j = 1 , 2 , n } . Then
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 max Φ U min { m , p , n } t r ( M ( Φ ) ) t r ( S 1 H S 1 + S 2 H S 2 ) + n ,
where M ( Φ ) is given by (11).
Proof. 
Let Z = A B , Z ˜ = A ˜ B ˜ , then Z H Z = R H R 1 , Z ˜ H Z ˜ = R ˜ H R ˜ 1 are nonsingular. Let
S 1 = ( Z H Z ) 1 / 2 A H = ( R H R 1 ) 1 / 2 R H Σ A H U H ,
S 2 = ( Z ˜ H Z ˜ ) 1 / 2 A ˜ H = ( R ˜ H R ˜ 1 ) 1 / 2 R ˜ H Σ A ˜ H U ˜ H S 3 = ( Z H Z ) 1 / 2 B H = ( R H R 1 ) 1 / 2 R H Σ B H V H ,
S 4 = ( Z ˜ H Z ˜ ) 1 / 2 B ˜ H = ( R ˜ H R ˜ 1 ) 1 / 2 R ˜ H Σ B ˜ H V ˜ H
and
M 1 ( Φ ) = S 1 Φ H S 2 H S 2 Φ S 1 H , Φ U m ; M 2 ( Φ ) = S 1 H Φ H S 2 S 2 H Φ S 1 , Φ U n M 3 ( Φ ) = S 3 Φ H S 4 H S 4 Φ S 3 H , Φ U p ; M 4 ( Φ ) = S 3 H Φ H S 4 S 4 H Φ S 3 , Φ U n .
We first prove
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 max Φ U m tr ( M 1 ( Φ ) ) t r ( S 1 H S 1 + S 2 H S 2 ) + n .
Using the GSVDs of { A , B } and { A ˜ , B ˜ } in Definition 3 we have
A = U Σ A R , B = V Σ B R , A ˜ = U ˜ Σ A ˜ R ˜ , B ˜ = V ˜ Σ B ˜ R ˜ ,
where U , U ˜ , V , V ˜ , Σ A , Σ B , Σ A ˜ and Σ B ˜ are given by (2) and (3). Let M 1 ( Φ ) be given by (18). Then
max Φ U m tr ( M 1 ( Φ ) ) = max Φ U m tr ( Φ H S 2 H S 2 Φ S 1 H S 1 ) = max Φ U m tr ( Φ U ˜ Σ A ˜ Σ A ˜ H U ˜ H Φ H U Σ A Σ A H U H ) .
Next we consider two cases for different m and n. Case 1: if n m , then we write
Σ A = Σ ^ A O ( m n ) × n , Σ A ˜ = Σ ^ A ˜ O ( m n ) × n , U H Φ U ˜ = W 11 W 12 W 21 W 22 , Φ U m ,
where Σ ^ A , Σ ^ A ˜ , W 11 C n × n and W 22 C ( m n ) × ( m n ) . Substituting into (20) we have
max Φ U m tr ( M 1 ( Φ ) ) = max W 11 W 11 H I n tr ( Σ ^ A H W 11 Σ ^ A ˜ Σ ^ A ˜ H W 11 H Σ ^ A ) ,
where W 11 W 11 H I n means that I n W 11 W 11 H is positive semi-definite. As we know, tr ( Σ ^ A H W 11 Σ ^ A ˜ Σ ^ A ˜ H W 11 H Σ ^ A ) as an analytical function of several complex variables on the domain W 11 W 11 H I n attains its maximum modulus on the characteristic manifold { W 11 C n × n : W 11 W 11 H = I n } (see [21]), therefore by Lemma 2 we have
max Φ U m tr ( M 1 ( Φ ) ) = max W 11 W 11 H = I n tr ( Σ ^ A H W 11 Σ ^ A ˜ Σ ^ A ˜ H W 11 H Σ ^ A ) ,
which together with Lemma 1 and (20) yields
max Φ U m tr ( M 1 ( Φ ) ) = i = 1 n α i 2 α ˜ i 2 .
Let U H Ψ U = U 11 U 12 U 21 U 22 , U ˜ H Ψ ˜ U ˜ = U ˜ 11 U ˜ 12 U ˜ 21 U ˜ 22 , Ψ , Ψ ˜ U m with U 11 , U ˜ 11 C n × n and U 22 , U ˜ 22 C ( m n ) × ( m n ) , then by Lemma 1 and (33) we have
tr ( S 1 H S 1 + S 2 H S 2 ) = tr ( Σ A Σ A H + Σ A ˜ Σ A ˜ H ) , = i = 1 n ( α i 2 + α ˜ i 2 ) .
Case 2: if m n , then we write
Σ A = ( Σ ^ A , O m × ( n m ) ) , Σ A ˜ = ( Σ ^ A ˜ , O m × ( n m ) ) .
Then by (20) and Lemma 1 we have
max Φ U m tr ( M 1 ( Φ ) ) = max Φ U m tr ( Φ U ˜ Σ A ˜ Σ A ˜ H U ˜ H Φ H U Σ A Σ A H U H ) = max Φ U m tr Φ U ˜ Σ ^ A ˜ Σ ^ A ˜ H U ˜ H Φ H U Σ ^ A Σ ^ A H U H 0 0 0 ( n m ) × ( n m ) = max Φ U m tr ( Φ U ˜ Σ ^ A ˜ Σ ^ A ˜ H U ˜ H Φ H U Σ ^ A Σ ^ A H U H ) = i = 1 m α i 2 α ˜ i 2 = i = 1 n α i 2 α ˜ i 2
since α m + 1 = = α n = α ˜ m + 1 = = α ˜ n = 0 by the definition of GSVD. It follows from Lemma 1 and (33) that
tr ( S 1 H S 1 + S 2 H S 2 ) = tr ( U Σ A Σ A H U H + U ˜ Σ A ˜ Σ A ˜ H U ˜ H ) , = i = 1 m ( α i 2 + α ˜ i 2 ) = i = 1 n ( α i 2 + α ˜ i 2 ) .
By Definition 3 we have β i 2 = 1 α i 2 , β ˜ i 2 = 1 α ˜ i 2 , and together with (20)-(24) we have
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = i = 1 n 2 α i 2 α ˜ i 2 ( α i 2 + α ˜ i 2 ) + 1 = 2 i = 1 n α i 2 α ˜ i 2 i = 1 n ( α i 2 + α ˜ i 2 ) + n = 2 max Φ U m tr ( M 1 ( Φ ) ) tr ( S 1 H S 1 + S 2 H S 2 ) + n .
By Cases 1 and 2 we deduce (35) holds true. By the similar aforementioned methods we get
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 max Φ U n tr ( M 2 ( Φ ) ) tr ( S 1 H S 1 + S 2 H S 2 ) + n .
Note that in (41)
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 i = 1 n β i 2 β ˜ i 2 i = 1 n ( β i 2 + β ˜ i 2 ) + n ,
and hence by the similar aforementioned methods we have
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 max Φ U p tr ( M 3 ( Φ ) ) tr ( S 1 H S 1 + S 2 H S 2 ) + n ,
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 max Φ U n tr ( M 4 ( Φ ) ) tr ( S 1 H S 1 + S 2 H S 2 ) + n .
In order to select the smallest dimensional variable unitary matrix by (35), (42)-(44) we have (31) holds. This completes the proof.    □
Lemma 4. 
Let Λ = d i a g ( λ 1 , , λ n ) , Δ = d i a g ( δ 1 , , δ n ) R n × n be diagonal matrices with λ 1 λ n 0 and δ 1 δ n 0 . Then we have
max Φ OR n tr ( Λ Φ T Δ Φ ) = i = 1 n λ i δ i ,
Proof. 
It follows from (1) that
max Φ OR n tr ( Λ Φ T Δ Φ ) = max Φ OR n tr ( Λ Φ H Δ Φ ) max Φ U n tr ( Λ Φ H Δ Φ ) = i = 1 n λ i δ i .
Meanwhile,
i = 1 n λ i δ i = tr ( Λ Δ ) max Φ OR n tr ( Λ Φ T Δ Φ ) .
Combing (29) and (30) yields
max Φ OR n tr ( Λ Φ T Δ Φ ) = i = 1 n λ i δ i .
This completed the proof.    □
Lemma 5. 
Let A R m × n , B R p × n have rank ( A T , B T ) = n and the GSV ( α i , β i ) of { A , B } be given in (4) and (5). Then
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 max Φ OR min { m , p , n } t r ( M ( Φ ) ) t r ( S 1 T S 1 + S 2 T S 2 ) + n ,
where M ( Φ ) is given by (13).
Proof. 
Let Z = A B , Z ˜ = A ˜ B ˜ , then Z T Z = R T R 1 , Z ˜ T Z ˜ = R ˜ T R ˜ 1 are nonsingular. Let
S 1 = ( Z T Z ) 1 / 2 A T = ( R T R 1 ) 1 / 2 R T Σ A T U T ,
S 2 = ( Z ˜ T Z ˜ ) 1 / 2 A ˜ T = ( R ˜ T R ˜ 1 ) 1 / 2 R ˜ T Σ A ˜ T U ˜ T S 3 = ( Z T Z ) 1 / 2 B T = ( R T R 1 ) 1 / 2 R T Σ B T V T ,
S 4 = ( Z ˜ T Z ˜ ) 1 / 2 B ˜ T = ( R ˜ T R ˜ 1 ) 1 / 2 R ˜ T Σ B ˜ T V ˜ T
and
M 1 ( Φ ) = S 1 Φ T S 2 T S 2 Φ S 1 T , Φ OR m ; M 2 ( Φ ) = S 1 T Φ T S 2 S 2 T Φ S 1 , Φ OR n M 3 ( Φ ) = S 3 Φ T S 4 T S 4 Φ S 3 T , Φ OR p ; M 4 ( Φ ) = S 3 T Φ T S 4 S 4 T Φ S 3 , Φ OR n .
We first prove
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 max Φ OR m tr ( M 1 ( Φ ) ) t r ( S 1 T S 1 + S 2 T S 2 ) + n .
Then
max Φ OR m tr ( M 1 ( Φ ) ) = max Φ OR m tr ( Φ T S 2 H S 2 Φ S 1 T S 1 ) = max Φ OR m tr ( Φ U ˜ Σ A ˜ Σ A ˜ T U ˜ T Φ H U Σ A Σ A H U T ) ,
which together with Lemma 1 and (20) yields
max Φ U m tr ( M 1 ( Φ ) ) = i = 1 n α i 2 α ˜ i 2 .
Let U T Ψ U = U 11 U 12 U 21 U 22 , U ˜ T Ψ ˜ U ˜ = U ˜ 11 U ˜ 12 U ˜ 21 U ˜ 22 , Ψ , Ψ ˜ OR m with U 11 , U ˜ 11 C n × n and U 22 , U ˜ 22 C ( m n ) × ( m n ) , then by Lemma 1 and (33) we have
tr ( S 1 H S 1 + S 2 H S 2 ) = tr ( Σ A Σ A H + Σ A ˜ Σ A ˜ H ) , = i = 1 n ( α i 2 + α ˜ i 2 ) .
Case 2: if m n , then we write
Σ A = ( Σ ^ A , O m × ( n m ) ) , Σ A ˜ = ( Σ ^ A ˜ , O m × ( n m ) ) .
Then by (20) and Lemma 1 we have
max Φ OR m tr ( M 1 ( Φ ) ) = max Φ OR m tr ( Φ U ˜ Σ A ˜ Σ A ˜ T U ˜ T Φ T U Σ A Σ A T U T ) = max Φ OR m tr Φ U ˜ Σ ^ A ˜ Σ ^ A ˜ T U ˜ T Φ T U Σ ^ A Σ ^ A T U T 0 0 0 ( n m ) × ( n m ) = max Φ OR m tr ( Φ U ˜ Σ ^ A ˜ Σ ^ A ˜ T U ˜ T Φ T U Σ ^ A Σ ^ A T U T ) = i = 1 m α i 2 α ˜ i 2 = i = 1 n α i 2 α ˜ i 2
since α m + 1 = = α n = α ˜ m + 1 = = α ˜ n = 0 by the definition of GSVD. It follows from Lemma 1 and (33) that
tr ( S 1 H S 1 + S 2 H S 2 ) = tr ( U Σ A Σ A T U T + U ˜ Σ A ˜ Σ A ˜ T U ˜ T ) , = i = 1 m ( α i 2 + α ˜ i 2 ) = i = 1 n ( α i 2 + α ˜ i 2 ) .
By Definition 3 we have β i 2 = 1 α i 2 , β ˜ i 2 = 1 α ˜ i 2 , and together with (20)-((24)) we have
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = i = 1 n 2 α i 2 α ˜ i 2 ( α i 2 + α ˜ i 2 ) + 1 = 2 i = 1 n α i 2 α ˜ i 2 i = 1 n ( α i 2 + α ˜ i 2 ) + n = 2 max Φ U m tr ( M 1 ( Φ ) ) tr ( S 1 H S 1 + S 2 H S 2 ) + n .
By Cases 1 and 2 we deduce (19) holds true. By the similar aforementioned methods we get
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 max Φ OR n tr ( M 2 ( Φ ) ) tr ( S 1 H S 1 + S 2 H S 2 ) + n .
Note that in (25)
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 i = 1 n β i 2 β ˜ i 2 i = 1 n ( β i 2 + β ˜ i 2 ) + n ,
and hence by the similar aforementioned methods we have
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 max Φ OR p tr ( M 3 ( Φ ) ) tr ( S 1 H S 1 + S 2 H S 2 ) + n ,
i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) = 2 max Φ OR n tr ( M 4 ( Φ ) ) tr ( S 1 H S 1 + S 2 H S 2 ) + n .
In order to select the smallest dimensional variable unitary matrix by (35), (42)-(44) we have (31) holds. This completes the proof.    □

2. Explicit Formula of i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) with Riemannian Optimization Models

We first present a new explicit formula of i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) for complex grassman matrix pair. Then we use the corresponding numerical methods to solve solutions of the involved Riemannian optimization models.

2.1. For complex grassman matrix pair

We present a new explicit formula of i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) for complex grassman matrix pair.
Lemma 6. 
Let M ( Φ ) , Q i , S 1 , S 2 be given by (10), (11) and (14). Then
max Φ U min { m , p , n } t r ( M ( Φ ) ) = t r ( M ( F 2 F 1 H ) ) ,
where F 1 , F 2 U min { m , p , n } can be obtained from Π 1 , Π 2 by solving the trace function optimization models
max Π 1 U min { m , p , n } | t r ( Π 1 H S 1 H S 1 Π 1 Q i ) | ,
max Π 2 U min { m , p , n } | t r ( Π 2 H S 2 H S 2 Π 2 Q i ) |
for every i = 1 , , n , respectively.
Proof. 
We first prove the case for m n , m p and for others cases when n m , m p ; p n , p m ; and n p , p m the corresponding results can be proved similarly. By (10), (11) and (18) S 1 , S 2 , M , F 1 , F 2 are taken as S 1 , S 2 , M 1 , F 1 , F 2 , where F 1 and F 2 can be obtained by solving
max Π 1 U min { m , p , n } | t r ( Π 1 H S 1 H S 1 Π 1 Q i ) | ,
max Π 2 U min { m , p , n } | t r ( Π 2 H S 2 H S 2 Π 2 Q i ) |
for every i = 1 , , n , respectively. Let f i be the i-th column of unitary matrix F 1 . By Lemma 1 we have for i = 1
max Π 1 U min { m , p , n } | t r ( Π 1 H S 1 H S 1 Π 1 Q 1 ) | = | t r ( F 1 H S 1 H S 1 F 1 Q 1 ) | = f 1 H S 1 H S 1 f 1 = λ 1 .
Here f 1 has some solutions and is set to satisfy S 1 H S 1 f 1 = λ 1 f 1 . Similarly, for i > 1
max Π 1 U min { m , p , n } | t r ( Π 1 H S 1 H S 1 Π 1 Q i ) | = | t r ( F 1 H S 1 H S 1 F 1 Q i ) | = f 1 H S 1 H S 1 f 1 + + f i H S 1 H S 1 f i = λ 1 + + λ i .
Here f i has some solutions and is set to satisfy S 1 H S 1 f i = λ i f i . Since F 1 is unitary, then f k H f j = 0 and f k H f k = 1 , k j ; k , j = 1 , , n . It follows that f k H S 1 H S 1 f j = λ i f k H f j = 0 and f k H S 1 H S 1 f k = λ k f k H f k = λ k , k j ; k , j = 1 , , n . Then
F 1 H S 1 H S 1 F 1 = Λ 1 ,
where Λ 1 = d i a g ( λ i ) . Similarly,
F 2 H S 2 H S 2 F 2 H = Λ 2 ,
where Λ 2 = d i a g ( χ i ) . By Lemma 1 and (47) we have for i = 1
max Π 1 U min { m , p , n } | t r ( Π 1 H S 1 H S 1 Π 1 Q 1 ) | = | t r ( F 1 H S 1 H S 1 F 1 Q 1 ) | = λ 1 = max 1 i n λ i .
For other i = 2 , , n , we have
max Π 1 U min { m , p , n } | t r ( Π 1 H S 1 H S 1 Π 1 Q i ) | = | t r ( F 1 H S 1 H S 1 F 1 Q i ) | = | t r ( Λ 1 Q i ) | = λ 1 + + λ i .
Then Λ 1 = d i a g ( λ i ) , Λ 2 = d i a g ( χ i ) have descending sorts with λ i 0 , χ i 0 , otherwise, there will be conflicts with (51) and (52). Then by Lemma 1 we have
max Φ U m tr ( M 1 ( Φ ) ) = max Φ U m tr ( Φ H S 2 H S 2 Φ S 1 H S 1 ) = max Φ U m tr ( Φ H F 2 Λ 2 F 2 H Φ F 1 Λ 1 F 1 H ) = tr ( Λ 2 Λ 1 ) .
It is easy to check that tr ( M 1 ( F 2 F 1 H ) ) = tr ( Λ 2 Λ 1 ) , which together with (53) implies max Φ U m tr ( M 1 ( Φ ) ) = tr ( M 1 ( F 2 F 1 H ) ) . For other cases for n m , m p ; p n , p m ; and n p , p m ; the desired corresponding results can be proved similarly. The proof is complete.    □
Theorem 3. 
Let M ( Φ ) , Q i , S 1 , S 2 be given by (10), (11) and (14). Let { A , B } and { A ˜ , B ˜ } be two ( m , p , n ) -GMPs and σ { A , B } = { ( α i , β i ) , i = 1 , 2 , n } and σ { A ˜ , B ˜ } = { ( α ˜ j , β ˜ j ) , j = 1 , 2 , n } . Then
i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) = t r ( S 1 H S 1 + S 2 H S 2 ) 2 max Φ U min { m , p , n } t r ( M ( Φ ) ) 2 t r ( L 1 1 2 ( I min { m , p , n } L 1 ) 1 2 L 2 1 2 ( I min { m , p , n } L 2 ) 1 2 ) = t r ( S 1 H S 1 + S 2 H S 2 ) 2 t r ( M ( F 2 F 1 H ) ) 2 t r ( L 1 1 2 ( I min { m , p , n } L 1 ) 1 2 L 2 1 2 ( I min { m , p , n } L 2 ) 1 2 ) ,
where
L 1 = F 1 H S 1 H S 1 F 1 , L 2 = F 2 H S 2 H S 2 F 2
and F 1 , F 2 U min { m , p , n } can be obtained from Π 1 , Π 2 by solving the trace function optimization models (45) and (46).
Proof. 
We first prove the case when m n , m p and for others cases when n m , m p ; p n , p m ; and n p , p m the corresponding results can be proved similarly. By (10), (11) and (18) S 1 , S 2 , M , F 1 , F 2 can be taken as S 1 , S 2 , M 1 , F 1 , F 2 . Since F 1 and F 2 are solutions of (45) and (46) for every i = 1 , , n , respectively, then S 1 H S 1 = A ( A H A + B H B ) 1 A H = U Σ A Σ A H U H . For i = 1 by (45) and Lemma 1 we have
max Π 1 U min { m , p , n } | t r ( Π 1 H A ( A H A + B H B ) 1 A H Π 1 Q 1 ) | = | t r ( F 1 H A ( A H A + B H B ) 1 A H F 1 Q 1 ) | = α 1 ,
For i = 2 we have
max Π 1 U min { m , p , n } | t r ( Π 1 H A ( A H A + B H B ) 1 A H Π 1 Q i ) | = | t r ( F 1 H A ( A H A + B H B ) 1 A H F 1 Q 2 ) | = α 1 + α 2 .
For i > 2 , we have
max Π 1 U min { m , p , n } | t r ( Π 1 H A ( A H A + B H B ) 1 A H Π 1 Q i ) | = α 1 + + α i .
Hence, we have L 1 = F 1 H S 1 H S 1 F 1 = d i a g ( α i ) , L 2 = F 2 H S 2 H S 2 F 2 = d i a g ( α ˜ i ) . It follows that
i = 1 n α i β i α ˜ i β ˜ i = t r ( L 1 1 2 ( I m L 1 ) 1 2 L 2 1 2 ( I m L 2 ) 1 2 ) ,
which together with (25) and Lemma 6 yields
i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) = n i = 1 n ( α i 2 α ˜ i 2 + β i 2 β ˜ i 2 ) 2 i = 1 n α i β i α ˜ i β ˜ i , = t r ( S 1 H S 1 + S 2 H S 2 ) 2 max Φ U m t r ( M 1 ( Φ ) ) 2 t r ( L 1 1 2 ( I m L 1 ) 1 2 L 2 1 2 ( I m L 2 ) 1 2 ) , = t r ( S 1 H S 1 + S 2 H S 2 ) 2 t r ( M 1 ( F 2 F 1 H ) ) 2 t r ( L 1 1 2 ( I m L 1 ) 1 2 L 2 1 2 ( I m L 2 ) 1 2 ) .
For the cases when n m , m p ; p n , p m ; and n p , p m ; by the similar aforementioned methods the desired results can be proved.    □
Remark 1. 
From Theorem 3 we see that the formula of i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) in (3) only refers to two small size unitary matrices F 1 U min { m , p , n } , F 2 U min { m , p , n } , which may require less computational cost for computing the explicit expression in Theorem 3.
When using Theorem 3 we first determine S 1 , S 2 by comparing m , p , n . Then by using (45) and (46) F 1 , F 2 can be computed.

2.2. Solving Riemannian Optimization Models in (45) and (46) for Complex Grassman Matrix Pair

We will use Newton methods on Riemannian manifolds to solve Riemannian optimization models in (45) and (46) of Theorem 3.
max Π 1 U min { m , p , n } | t r ( Π 1 S 1 H S 1 Π 1 H Q i ) | ,
for n m , n p and m n , m p and
max Π 2 U min { m , p , n } | t r ( Π 2 S 1 H S 1 Π 2 H Q i ) | ,
for p n , p m and n p , p m . Solving max Π 1 U min { m , p , n } | t r ( Π 1 S 2 H S 2 Π 1 H Q i ) | can be similarly discussed. For simplicity, we assume that n m and n p for (54) and n p , p m for ( ) and other cases for different relationships between m , n , p can be similarly discussed. To compute (54) and (55), we consider the following matrix optimization problems:
max f 1 ( Φ 1 ) : = 1 2 tr ( Φ 1 H Q i Φ 1 C ) s . t . Φ 1 U n
and
max f 2 ( Φ 2 ) : = 1 2 tr ( Φ 2 H Q i Φ 2 D ) s . t . Φ 2 U n
for 1 i n . We only need to focus on the following matrix optimization problems:
max f 1 ( Φ 11 ) : = 1 2 tr ( Φ 11 H C Φ 11 ) s . t . Φ 11 S t ( n , i ) : = { Φ 11 C n × i | Φ 11 H Φ 11 = I i } ,
max f 2 ( Φ 22 ) : = 1 2 tr ( Φ 22 H D Φ 22 ) s . t . Φ 22 S t ( n , n i + 1 ) ,
where C = ( A H A + B H B ) 1 / 2 A H A ( A H A + B H B ) 1 / 2 , D = ( A H A + B H B ) 1 / 2 B H B ( A H A + B H B ) 1 / 2 , Φ 11 is the first i columns of Φ 1 H and Φ 22 is the last n i + 1 columns of Φ 2 H . In the following, we focus on the solution of the matrix optimization problem (58) and the matrix optimization problem (59) can be solved in a similar way.
We note that f 1 ( Φ 11 Q ) = f 1 ( Φ 11 ) for all Φ 11 S t ( n , i ) and Q U i . Let the complex Grassmann manifold Grass ( n , i ) be the set of all i-dimensional complex subspace of C n . If [ X ] means the subspace spanned by the columns of X S t ( n , i ) , then we have [ X ] Grass ( n , i ) . In particular, for any X S t ( n , i ) , the natural projection [ X ] Grass ( n , i ) corresponds the equivalent class { X Q S t ( n , i ) | Q U i } of S t ( n , i ) . Thus, instead of problem (58), we consider the following optimization problem:
max f ˜ 1 ( [ Φ 11 ] ) : = f 1 ( Φ 11 ) s . t . [ Φ 11 ] Grass ( n , i ) .
Next, we present Newton’s method for solving the optimization problem (60). Let Φ 11 S t ( n , i ) . The tangent space of S t ( n , i ) at Φ 11 is given by [29]
T Φ 11 S t ( n , i ) = { Z C n × i | Z = Φ 11 Ω + ( Φ 11 ) K , Ω H = Ω , Ω C i , i , K C n i , i } ,
which can be endowed with the inner product
Z 1 , Z 2 = R [ tr ( Z 2 H ( I 1 2 Φ 11 Φ 11 ) H ) Z 1 ) ] , Z 1 , Z 2 T Φ 11 S t ( n , i ) , Φ 11 S t ( m , i ) .
Here, ( Φ 11 ) means that span ( ( Φ 11 ) ) is the orthogonal complement of span ( Φ 11 ) , where span ( Φ 11 ) denotes a linear space spanned by the column vectors of Φ 11 . We note that the tangent space of Grass ( n , i ) at [ Φ 11 ] is given by [29]
T [ Φ 11 ] Grass ( n , i ) = { Z C n × i | Z = ( Φ 11 ) K , K C n i , i } T Φ 11 S t ( n , i ) .
Hence, we can define a Riemannian metric on Grass ( n , i ) by
Z 1 , Z 2 = R [ tr ( Z 2 H Z 1 ) ] , Z 1 , Z 2 T [ Φ 11 ] Grass ( n , i ) , Φ 11 S t ( n , i )
with the induced norm · . Then the orthogonal projection onto T [ Φ 11 ] Grass ( m , i ) is given by
P Φ 11 Z = ( I Φ 11 Φ 11 H ) Z , Z C n × i .
We define the local cost function g : T [ Φ 11 ] Grass ( n , i ) R by
g ( Z ) = f ˜ 1 ( [ Φ 11 + Z ] ) .
It is easy to check that, for any Z T [ Φ 11 ] Grass ( n , i ) ,
g ( Z ) = f 1 ( Φ 11 ) + R tr ( Z H D Φ 11 ) + 1 2 vec ( Z ) H H Φ 11 ( Φ 11 H D Φ 11 ) T I n vec ( Z ) ,
where D Φ 11 = C Φ 11 and H Φ 11 = I i C are the derivative and Hessian of f 1 at Φ 11 , respectively.
Based on the above analysis, Newton’s method for solving the optimization problem (60) can be described as follows [29].
Algorithm 4. 
Step 0. Choose Φ 11 0 S t ( n , i ) , ρ , η ( 0 , 1 ) , σ ( 0 , 1 / 2 ] and let k : = 1 .
Step 1.
Apply the conjugate gradient (CG) method to solving
P Φ 11 k ( C Z k Z k ( Φ 11 k ) H D Φ 11 k ) = P Φ 11 k D Φ 11 k
for Z k T [ Φ 11 k ] Grass ( n , i ) such that
P Φ 11 k ( C Z k Z k ( Φ 11 k ) H D Φ 11 k ) + P Φ 11 k D Φ 11 k η k P Φ 11 k D Φ 11 k
and
P Φ 11 k D Φ 11 k , Z k η k Z k , Z k ,
where  η k = min { η , P Φ 11 k D Φ 11 k } . If (61) and (62) are not attainable, then let
Z k = P Φ 11 k D Φ 11 k .
Step 2.
Let  l k > 0  be the smallest integer m such that
f 1 ( π ( Φ 11 k + β l Z k ) ) f 1 ( Φ 11 k ) + σ ρ l P Φ 11 k D Φ 11 k , Z k
Set 
Z k + 1 = π ( Φ 11 k + β l Z k ) .
Step 3.
Replace k by  k + 1 and go to Step 1.
In Step 2 of Algorithm 4, π : C m × i Grass ( n , i ) is the projection onto Grass ( n , i ) , which is defined as follows: Let X C n × i be full column rank. Then
π ( X ) = argmin Y S t ( n , i ) X Y 2 .
As noted in [29], if the SVD of X is given by X = U Σ V H , then π ( X ) = U I n , i V H . If the QR decomposition of X is X = Q R , then π ( X ) = Q n , i .
Remark 2.
(i) An explicit expression of i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) is given in Theorem 3. As a comparison, Theorem 2 only provides an upper bound. Obviously, Theorem 3 improves the existing results in [22] in theory, see Example below. We adopt trace function optimization formula with orthogonal constraints in Theorem 3 to evaluate the explicit expression of i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) , which costs less complexity. While Theorem 3.2 in [22] needs unitary matrices U , V , U ˜ , V ˜ of the GSVDs of two matrix pairs { A , B } and { A ˜ , B ˜ } . It is known that computing GSVDs may cost much higher amount of calculations than computing small-size unitary matrice. Hence, on algorithm aspect Theorem 3 also improves Theorem 3.2 in [22].
(ii) The equality in Theorem 3 refers to matrix optimization with orthogonal constraints. Based on matrix optimization formula with orthogonal constraints we use classical Golub-Kahan bidiagonalization method to solve it efficiently. The proposed algorithms by trace function optimization under two variables are efficient for computing for A H A + B H B being not ill-posed. For A H A + B H B = W being ill-posed, by matrix decompositions (eg., Schur decomposition or QR decomposition) we compute W 1 , W 2 with W = W 1 W 2 , where W 1 , W 2 have small condition numbers. We calculate W 1 1 , W 2 1 and then compute multiplications for W 1 = W 2 1 W 1 1 , see [27,28]. In general we first compute the inverse of the matrix with small condition numbers and then multiply these two inverse matrices, which can well get the inverse of the original matrix with large condition numbers. The proposed algorithms by trace function optimization under two variables are efficient for computing for condition numbers of A H A + B H B being large, see Table 1. In practical applications, obviously, by the new explicit expression formula getting exact values of i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) are more precise than its upper bound.
Example. To test the comparison of the explicit formula in Theorem 3, and the upper bound of i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) in [22], we randomly generate grassman matrix pencils { A , B } and { A ˜ , B ˜ } by MATLAB command randn(m,n)+i*randn(m,n) for A , A ˜ and randn(p,n)+i*randn(p,n) for B , B ˜ with different random m , n , p . We first assess condition numbers of A H A + B H B by using two-norm and denoted by κ 2 ( · ) . By simple computations we present the table 1.
Table 1. Comparisons of exact values, explicit formula by Alg. 4 and the upper bound with different ( m , p , n )
Table 1. Comparisons of exact values, explicit formula by Alg. 4 and the upper bound with different ( m , p , n )
( m , p , n ) κ 2 ( · ) Exact values formula by Alg. 4 Upper Bound
(80,40,60) 30.2 27.24709031 27.24698583 41.3510
(200,500,450) 218.6 189.26904380 189.26910247 380.4629
(900,800,700) 359.4 458.42617335 458.42618210 573.0744
(60,120,140) 600.3 103.56173621 103.56172847 126.0846
(100,200,150) 400.6 79.35162440 79.35157629 118.5329
(250,500,450) O ( 10 3 ) 268.49009138 268.49008938 376.4290
(800,600,500) O ( 10 3 ) 409.05215925 409.05215570 454.7393
(2000,1900,1800) O ( 10 4 ) 1516.38636811 1516.38632996 1704.1106
(500,600,800) O ( 10 5 ) 379.23164800 379.23164655 634.2572
Here `Explicit expression’ denotes the equality in Theorem 3, `Upper Bound’ denotes the upper bound in [22]. This table obviously shows the equality in Theorem 3 is almost equal to exact value of i = 1 n ρ 2 ( ( α i , β i ) , ( α ˜ i , β ˜ i ) ) and may be much sharper than the upper bounds in [22]. This again verifies Remark (i).

3. Concluding Remarks

In this paper we give explicit expressions of sum of chordal metric between generalized singular values of Grassmann matrix pairs, which absolutely improves the existing upper bounds in [22]. For the new expression formula only two small-size singular value decompositions are required in the algorithm which dominates the computational cost. As a byproduct, we also present an interval estimate without using the singular values.

References

  1. Alter, O., Brown, P. O., and Botstein, D. 2000. "Singular value decomposition for genome-wide expression data processing and modelling." Proceedings of the National Academy of Sciences 97: 10101–10106. [CrossRef]
  2. Alter, O., Brown, P. O., and Botstein, D. 2001. "Processing and modelling gene expression data using singular value decomposition." Proceedings of SPIE 4266: 171–186. [CrossRef]
  3. Alter, O., Brown, P. O., and Botstein, D. 2003. "Generalized singular decomposition for comparative analysis of genome-scale expression data sets of two different organisms." Proceedings of the National Academy of Sciences 100: 3351–3356. [CrossRef]
  4. lter, O., Golub, G. H., Brown, P. O., and Botstein, D. 2004. "Novel genome-scale correlation between DNA replication and RNA transcription during the cell cycle in yeast is predicted by data-driven models." In Proceedings of the Miami Nature Biotechnology Winter Symposium on the Cell Cycle, Chromosomes and Cancer, vol. 15, edited by M. P. Deutscher et al., 15. University of Miami School of Medicine, Miami.
  5. Bai, Z. , and Demmel, J. W. 1993. "Computing the generalized singular value decomposition." SIAM Journal on Scientific Computing 14: 1464–1486.
  6. Bai, Z. , and Zha, H. 1993. "A new preprocessing algorithm for the computation of the generalized singular value decomposition." SIAM Journal on Scientific Computing 14: 1007–1012. [CrossRef]
  7. Bhatia, R. 1997. Matrix Analysis. New York: Springer.
  8. Sun, J.-G. 1983. "Perturbation analysis for the generalized eigenvalue problem and the generalized singular value problem." In Matrix Pencils, Lecture Notes in Mathematics, vol. 973, edited by Springer Verlag, 221–244.
  9. Li, R.-C. 1993. "Bounds on perturbations of generalized singular values and of associated subspaces." SIAM Journal on Matrix Analysis and Applications 14: 195–234. [CrossRef]
  10. Li, R.-C. 1993. "A perturbation bound for definite pencils." Linear Algebra and its Applications 179: 191–202. [CrossRef]
  11. Li, R.-C. 1994. "On perturbations of matrix pencils with real spectra." Mathematics of Computation 62: 231–265.
  12. Li, R.-C. 2002. "On perturbations of matrix pencils with real spectra, a revisit." Mathematics of Computation 72: 715–728.
  13. Mahony, R. E. 1996. "The constrained Newton method on a Lie group and the symmetric eigenvalue problem." Linear Algebra and its Applications 248(15): 67–89.
  14. Owren, B. , and Welfert, B. 2000. "The Newton iteration on Lie groups." BIT 40(1): 121–145.
  15. Paige, C. C. , and Saunders, M. A. 1981. "Towards a generalized singular value decomposition." SIAM Journal on Numerical Analysis 18: 398–405.
  16. Sun, J.-G. 1983. "Perturbation analysis for the generalized singular value problem." SIAM Journal on Matrix Analysis and Applications 20: 611–625.
  17. Sun, J.-G. 1998. "Perturbation analysis of generalized singular subspaces." Numerische Mathematik 79: 615–641.
  18. Sun, J.-G. 2000. "Condition number and backward error for the generalized singular value decomposition." SIAM Journal on Matrix Analysis and Applications 22: 323–341.
  19. Van Loan, C. F. 1976. "Generalizing the singular value decomposition." SIAM Journal on Numerical Analysis 13: 76–83.
  20. Wang, J. H., and Li, C. 2011. "Kantorovich’s theorems for Newton’s method for mappings and optimization problems on Lie groups." IMA Journal of Numerical Analysis 31:.
  21. Xu, W. W., Li, W., Zhu, L., and Huang, X. P. 2019. "The analytic solutions of a class of constrained matrix minimization and maximization problems with applications." SIAM Journal on Optimization 29: 1657–1686.
  22. Xu, W. W., Pang, H. K., Li, W., Huang, X. P., and Guo, W. J. 2018. "On the explicit expression of chordal metric between generalized singular values of Grassmann matrix pairs with applications." SIAM Journal on Matrix Analysis and Applications 39(4): 1547–1563.
  23. Zha, H. 1992. "A numerical algorithm for computing restricted singular value decomposition of matrix triplets." Linear Algebra and Its Applications 168: 1–26.
  24. Hua, L. K. 1963. Harmonic Analysis of Functions of Several Complex Variables in the Classical Domains. Providence, RI: American Mathematical Society.
  25. Golub, G. H., and Van Loan, C. F. 2013. Matrix Computations, 4th ed. Baltimore: Johns Hopkins University Press.
  26. Golub, G. H. , and Kahan, W. 1965. "Calculating the singular values and pseudo-inverse of a matrix." SIAM Journal on Numerical Analysis 2: 205–224.
  27. Ben-Israel, A., and Greville, T. N. E. 1977. Generalized Inverses: Theory and Applications. New York: Wiley.
  28. Jodár, L., Law, A. G., Rezazadeh, A., Watson, J. H., and Wu, G. 1991. "Computations for the Moore-Penrose and Other Generalized Inverses." Congressus Numerantium 80: 57–64.
  29. Manton, J. H. 2002. "Optimization algorithms exploiting unitary constraints." IEEE Transactions on Signal Processing 50: 635–650.
  30. Absil, P.-A., Mahony, R., and Sepulchre, R. 2008. Optimization Algorithms on Matrix Manifolds. Princeton: Princeton University Press.
  31. Adler, R. L., Dedieu, J. P., Margulies, J. Y., Martens, M., and Shub, M. 2002. "Newton’s method on Riemannian manifolds and a geometric model for the human spine." IMA Journal of Numerical Analysis 22: 359–390.
  32. Edelman, A., Arias, T. A., and Smith, S. T. 1998. "The geometry of algorithms with orthogonality constraints." SIAM Journal on Matrix Analysis and Applications 20: 303–353.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated