Skip to content

Singular Value Decomposition

Introduction.

Let f:VW be a linear map between finite dimensional vectors spaces, dim(V)=N,dim(W)=M endowed with scalar products. The Singular Value Decomposition theorem, shows us that we can find ortho-normal bases vn,wm of V,W respectively, so that the matrix representation of f with respect to vn,wn has only diagonal entries: f(vi)=σiwi, i=1min(N,M) Furthermore the bases can be arranged in a way that σ1σ2σr00.

In this note, we are going to explore an abstract "coordinate free" version of SVD that can be expressed without choices of bases. Instead, we are going to study the eigenvalues of the self-adjoint operators S=ff on V and T=ff on W.

S=ffVfWff=T

By the Spectral Decomposition Theorem, the vector spaces V,W decompose as orthogonal direct sums of the Eigenspaces of S,T respectively: V=λVλ,W=λWλ.

How does f look like in a basis that is adjusted to this decompositions?

Proposition

  • The map f induces linear maps fλ:VλWλ, i.e. f(Vλ)Wλ.
  • The map f induces linear maps fλ:WλVλ, i.e. f(Wλ)Vλ.
  • fλfλ=λIVλ, and fλfλ=λIWλ.
  • For λ>0 the maps 1λfλ, 1λfλ are inverse isometries between the Eigenspaces VλWλ.
  • For λ=0 we find V0=Ker(f) and W0=Im(f), f0=0, f0=0.
Proof

1) For vVλ, we have (Tλ)f(v)=fffvλfv=f(ffvλv)=f(0)=0. 2) For wWλ, we have (Sλ)f(v)=fffvλfv=0. 3) For vVλ we have fλfλv=ff(v)=λv by definition of Vλ. 4) Clear from 3. 5) Ker(ff)=Ker(f) by properties of the adjoint. Similarly Ker(ff)=Im(ff)=Im(f).

Corollary (SVD). We find orthogonal bases of V,W with respect to which f has only diagonal entries 0.

Proof

We recover the SVD from the Proposition, by choosing orthogonal basis Bλ for each Eigenspace Vλ. For λ>0, we get a ONB for Wλ as Cλ=1λfλUλ. Choose B0, and C0 arbitrary in V0,W0. Now for λ>0 and a basis vector bVλ we have f(b)=λc for c=1λf(v)Cλ. For λ=0 we have f(b)=0. QED.

Corollary. Choosing a basis B,C of V,W, that is adjusted to the Eigenspaces Vλ, Wλ in descending order, the linear map f is represented by a block diagonal matrix A with entries Aλ:

A=[Aλ1Aλ2AλkA0]

Where, AλAλ=λI and AλAλ=λI. In particular Aλ is an isomorphisom for λ>0.

Application - Least Squares via Duality

Regular least squares problems can commonly be written as a minimization problem of the form:

Find vV so that L(v)=ϕvb2is minimal.

Where ϕ:VW is an injective linear map, and bW.

We can translate this minimization problem on V to a minimization problem on W by considering:

Find wW so that M(w)=ϕϕwb2is minimal.

Since ϕ is injective, ϕ is surjective. So a global minimum of M=Lϕ will map to a global minimum of L under ϕ.

Now M can be explicitly minimized by noting that:

ϕϕwb2=λSpec(ϕϕ)λwλbλ=b0+λ>0λwλbλ

Where wλ,bλ are the components of w,b in Wλ. This function is minized by w^ with w^0=0W0 and w^λ=bλ/λWλ, i.e.

w^=λ>01λbλ.

We obtain the following formula for the solution v^:=ϕ(w^)=λ1λϕbλ.

We claim that v^ is given by following explicit formula v^=(ϕϕ)1ϕb.

Indeed, ϕbλVλ, where (ϕϕ) operats as λ. Hence (ϕϕ)λ>0ϕbλ/λ=λ>0ϕbλ=ϕb as ϕb0=0.