Singular value decomposition (SVD) theorem
Theorem: Singular Value Decomposition (SVD)
An arbitrary matrix
admits a decomposition of the form
![]()
where
are both orthogonal matrices, and the matrix
is diagonal:
![]()
where the positive numbers
are unique, and are called the singular values of
. The number
is equal to the rank of
, and the triplet
is called a singular value decomposition (SVD) of
. The first
columns of
(resp.
) are called left (resp. right) singular vectors of
, and satisfy
![]()
Proof: The matrix
is real and symmetric. According to the spectral theorem, it admits an eigenvalue decomposition in the form
, with
a
matrix whose columns form an orthonormal basis (that is,
), and
. Here,
is the rank of
(if
then there are no trailing zeros in
). Since
is positive semi-definite, the
‘s are non-negative, and we can define the non-zero quantities
.
Note that when
, since then
.
Let us construct an
orthogonal matrix
as follows. We set
![]()
These
-vectors are unit-norm, and mutually orthogonal since
‘s are eigenvectors of
. Using (say) the Gram-Schmidt orthogonalization procedure, we can complete (if necessary, that is in the case
) this set of vectors by
in order to form an orthogonal matrix
.
Let us check that
satisfy the conditions of the theorem, by showing that
. We have
![Rendered by QuickLaTeX.com \[\left(U^T A V\right)_{i j}=u_i^T A v_j= \begin{cases}\sigma_j u_i^T u_j & \text { if } j \leq r \\ 0 & \text { otherwise, }\end{cases}\]](https://ecampusontario.pressbooks.pub/app/uploads/quicklatex/quicklatex.com-7a7b2d7c5efd29d42ea0872e8998e058_l3.png)
where the second line stems from the fact that
when
. Thus,
, as claimed.