1. Introduction
The Eckart–Young–Mirsky theorem is a fundamental result in matrix approximation, stating that for a given matrix
A and rank
k, the best rank-
k approximation in the Frobenius norm (or any unitarily invariant norm) is obtained by truncating the Singular Value Decomposition (SVD) of
A.[
1,
3] Formally, if
is the SVD of
A and
is obtained from
by keeping only the
k largest singular values (and setting the rest to zero), then
is the unique minimizer of
over all rank-
k matrices
X [
2,
4].
This theorem underpins numerous applications, from image compression to principal component analysis, yet standard proofs often rely on variational arguments or operator norm inequalities that can obscure geometric intuition [
5]. In this paper, we present a more elementary proof(only using basic linear algebraic).
2. Proof
In this section, we will give an elementary and short proof of the Eckart-Young-Mirsky Theorem.
Let
A be a real matrix with
and
in a descending order be all the non-zero singular values of
A. The SVD factors
A into
where
U and
V are orthogonal matrices and
is a
diagonal matrix.
Let
be an integer. Define
with
being a
diagonal matrix.
The
Eckart-Young-Mirsky Theorem states that
where
is the Frobenius norm defined by
for any real matrix
.
In the following, we will relax the condition
to
and prove that
Let
with
M being a
matrix of
. By (
2),
Therefore, to show (
3), it suffices to prove
as
achieves the minimum.
Fix a
k-dimensional subspace
such that the column vectors of
M lie in
W. Choose an orthonormal basis
of
such that
span
W, where
. Let
where
-s and
-s are column vectors of
and
M, respectively. We have
To minimize (
5),
should be the projection of
onto
W, i.e.,
where
is the standard inner product. Then for any
M whose column vectors are in
W,
The coefficients of
-s of (
6) satisfy:
Since
are in descending order and their coefficients all belong to
with the sum being
p, to minimize the right hand side of (
6), the coefficients should concentrate to the lowest
p singular values. Therefore,
3. Conclusion
This paper offers an elementary yet powerful proof of the Eckart-Young-Mirsky theorem, which is essential for many fields, such as machine learning, image processing, and data science. By demonstrating the best rank-k approximation through a clear application of basic linear algebra techniques, the paper contributes to a deeper understanding of low-rank matrix approximation. This work simplifies the theorem’s proof, making it more accessible for those familiar with basic matrix theory and reinforcing its crucial role in real-world applications like dimensionality reduction, data compression, and statistical analysis.
Combining (
6) and (
7), we get (
4), which concludes the proof.
References
- N. Kishore Kumar and J. Schneider, Literature survey on low rank approximation of matrices, Linear and Multi-linear Algebra, vol. 65, no. 11, pp. 2212-2244, 2017.
- L. N. Trefethen and D. Bau, Numerical Linear Algebra, SIAM, 1997.
- G. H. Golub and C. F. Van Loan, Matrix Computations, Johns Hopkins University Press, 2013.
- G. W. Stewart, Matrix Algorithms: Volume I: Basic Decompositions, SIAM, 1998.
- R. A. Horn and C. R. Johnson, Matrix Analysis, Cambridge University Press, 1985.
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).