Next: Latent semantic indexing
Up: Matrix decompositions and latent
Previous: Term-document matrices and singular
Contents
Index
Low-rank approximations
We next state a matrix approximation problem that at first seems to have little to do with information retrieval. We describe a solution to this matrix problem using singular-value decompositions, then develop its application to information retrieval.
Given an
matrix
and a positive integer
, we wish to find an
matrix
of rank at most
, so as to minimize the Frobenius norm of the matrix difference
, defined to be
![\begin{displaymath}
\Vert X \Vert _F = \sqrt{\sum_{i=1}^\lsinoterms \sum_{j=1}^\lsinodocs X_{ij}^2}.
\end{displaymath}](img1744.png) |
(238) |
Thus, the Frobenius norm of
measures the discrepancy between
and
; our goal is to find a matrix
that minimizes this discrepancy, while constraining
to have rank at most
. If
is the rank of
, clearly
and the Frobenius norm of the discrepancy is zero in this case. When
is far smaller than
, we refer to
as a low-rank approximation .
The singular value decomposition can be used to solve the low-rank matrix approximation problem. We then derive from it an application to approximating term-document matrices. We invoke the following three-step procedure to this end:
- Given
, construct its SVD in the form shown in (232); thus,
.
- Derive from
the matrix
formed by replacing by zeros the
smallest singular values on the diagonal of
.
- Compute and output
as the rank-
approximation to
.
The rank of
is at most
: this follows from the fact that
has at most
non-zero values. Next, we recall the intuition of Example 18.1: the effect of small eigenvalues on matrix products is small. Thus, it seems plausible that replacing these small eigenvalues by zero will not substantially alter the product, leaving it ``close'' to
. The following theorem due to Eckart and Young tells us that, in fact, this procedure yields the matrix of rank
with the lowest possible Frobenius error.
Theorem.
![\begin{displaymath}
\min_{Z \vert \mbox{ rank}(Z)=k} \Vert\lsimatrix-Z\Vert _F = \Vert\lsimatrix-\lsimatrix_k\Vert _F = \sigma_{k+1}.
\end{displaymath}](img1749.png) |
(239) |
End theorem.
Recalling that the singular values are in decreasing order
, we learn from
Theorem 18.3 that
is the best
rank-
approximation to
, incurring an error (measured
by the Frobenius norm of
) equal to
.
Thus the larger
is, the smaller this error (and in particular, for
, the error is zero since
; provided
, then
and thus
).
To derive further insight into why the process of truncating the smallest
singular values in
helps generate a rank-
approximation of low error, we examine the form of
:
where
and
are the
th columns of
and
, respectively. Thus,
is a rank-1 matrix, so that we have just expressed
as the sum of
rank-1 matrices each weighted by a singular value. As
increases, the contribution of the rank-1 matrix
is weighted by a sequence of shrinking singular values
.
Exercises.
Next: Latent semantic indexing
Up: Matrix decompositions and latent
Previous: Term-document matrices and singular
Contents
Index
© 2008 Cambridge University Press
This is an automatically generated page. In case of formatting errors you may want to look at the PDF edition of the book.
2009-04-07