Skip to article frontmatterSkip to article content

Matrices

It is often useful to think of a matrix-vector product can be though of as a linear combination of the columns of the matrix. That is,

[a1ad][c1cd]=c1a1++cdad.\begin{bmatrix} | & & |\\ \vec{a}_1 & \cdots & \vec{a}_d\\ | & & | \end{bmatrix} \begin{bmatrix} c_1 \\ \vdots \\ c_d \end{bmatrix} = c_1 \vec{a}_1 + \cdots + c_d \vec{a}_d.

Norms

It is often useful to bound different norms with each other.

Factorizations

The above definition is sometimes called a reduced or thin SVD. By extending U\vec{U} to an orthonormal basis for Rn\R^n and appending zeros below Σ\vec{\Sigma} we obtain a full SVD.

The SVD provides the best low-rank approximation to a matrix.

Conditioning

A problem/task is said to be well-conditioned if small perturbations in problem don’t change the solution by much. If a small change to the problem can cause a large change in the solution, then the problem is said to be ill-conditioned. Ill-conditioned problems are difficult (or impossible) to solve accurately on computers, since even small errors made representing the problem on a computer can drastically change the solution.

In numerical linear algebra the condition-number of a matrix A\vec{A} is defined as

cond(A):=σmax(A)σmin(A),\cond(\vec{A}) := \frac{\smax(\vec{A})}{\smin(\vec{A})},

where σmax(A)\smax(\vec{A}) and σmin(A)\smin(\vec{A}) are the largest and smallest singular values of A\vec{A}, respectively. The conditioning of many core linear algebra tasks depends on the condition number.

Krylov subspace methods

Krylov subspace methods are an important and powerful class of algorithms in numerical linear algebra. Such algorithms make use of the information in the so-called Krylov subspace.

Computationally, it is useful to obtain an orthonormal basis for relevant subspaces. In the case of Krylov subspaces, this can be done by the Arnoldi algorithm. When applied to a symmetric matrix A\vec{A} for kk iterations, the arnoldi algorithm produces an orthonormal basis QRn×k\vec{Q}\in\R^{n\times k} for the Krylov subspace Kk(A,x)\mathcal{K}_k(\vec{A}, \vec{x}) and an upper-Hessenberg matrix HRk×k\vec{H}\in\R^{k\times k} such that H=QTAQ\vec{H} = \vec{Q}^\T\vec{A}\vec{Q}.

In the special case that A\vec{A} is symmetric, we observe that QTAQ\vec{Q}^\T\vec{A}\vec{Q} is also symmetric. This means H\vec{H} is both upper-Hessenberg and symmetric, and thus tridiagonal. This allows substantial simplifications to the Arnoldi algorithm, which is then called the Lanczos algorithm.