A tridiagonal matrix is a square matrix whose elements are zero away from the main diagonal, the subdiagonal, and the superdiagonal. In other words, it is a banded matrix with upper and lower bandwidths both equal to . It has the form

An important example is the matrix that arises in discretizating the Poisson partial differential equation by a standard five-point operator, illustrated for by

It is symmetric positive definite, diagonally dominant, a Toeplitz matrix, and an -matrix.

Tridiagonal matrices have many special properties and various algorithms exist that exploit their structure.

## Symmetrization

It can be useful to symmetrize a matrix by transforming it with a diagonal matrix. The next result shows when symmetrization is possible by similarity.

Theorem 1.If is tridiagonal and for all then there exists a diagonal with positive diagonal elements such that is symmetric, with element .

Proof.Let . Equating and elements in the matrix gives

or

As long as for all we can set and solve (2) to obtain real, positive , . The formula for the off-diagonal elements of the symmetrized matrix follows from (1) and (2).

## LU Factorization

The LU factors of a tridiagonal matrix are bidiagonal:

The equation gives the recurrence

The recurrence breaks down with division by zero if one of the leading principal submatrices , , is singular. In general, partial pivoting must be used to ensure existence and numerical stability, giving a factorization where has at most two nonzeros per column and has an extra superdiagonal. The growth factor is easily seen to be bounded by .

For a tridiagonal Toeplitz matrix

the elements of the LU factors converge as if is diagonally dominant.

Theorem 2.Suppose that has an LU factorization with LU factors (3) and that and . Then decreases monotonically and

From (4), it follows that under the conditions of Theorem 2, increases monotonically and . Note that the conditions of Theorem 2 are satisfied if is diagonally dominant by rows, since implies . Note also that if we symmetrize using Theorem 1 then we obtain the matrix , which is irreducibly diagonally dominant and hence positive definite if .

## Inverse

The inverse of a tridiagonal matrix is full, in general. For example,

Since an tridiagonal matrix depends on only parameters, the same must be true of its inverse, meaning that there must be relations between the elements of the inverse. Indeed, in any submatrix whose elements lie in the upper triangle is singular, and the submatrix is also singular. The next result explains this special structure. We note that a tridiagonal matrix is irreducible if and are nonzero for all .

Theorem 3.If is tridiagonal, nonsingular, and irreducible then there are vectors , , , and , all of whose elements are nonzero, such that

The theorem says that the upper triangle of the inverse agrees with the upper triangle of a rank- matrix () and the lower triangle of the inverse agrees with the lower triangle of another rank- matrix (). This explains the singular submatrices that we see in the example above.

If a tridiagonal matrix is reducible with then it has the block form

where , and so

in which the block is rank if . This blocking can be applied recursively until Theorem 1 can be applied to all the diagonal blocks.

The inverse of the Toeplitz tridiagonal matrix is known explicitly; see Dow (2003, Sec. 3.1).

## Eigenvalues

The most widely used methods for finding eigenvalues and eigenvectors of Hermitian matrices reduce the matrix to tridiagonal form by a finite sequence of unitary similarity transformations and then solve the tridiagonal eigenvalue problem. Tridiagonal eigenvalue problems also arise directly, for example in connection with orthogonal polynomials and special functions.

The eigenvalues of the Toeplitz tridiagonal matrix in (5) are given by

The eigenvalues are also known for certain variations of the symmetric matrix in which the and elements are modified (Gregory and Karney, 1969).

Some special results hold for the eigenvalues of general tridiagonal matrices. A matrix is derogatory if an eigenvalue appears in more than one Jordan block in the Jordan canonical form, and nonderogatory otherwise.

Theorem 4.If is an irreducible tridiagonal matrix then it is nonderogatory.

Proof.Let , for any . The matrix is lower triangular with nonzero diagonal elements and hence it is nonsingular. Therefore has rank at least for all . If were derogatory then the rank of would be at most when is an eigenvalue, so must be nonderogatory.

Corollary 5.If is tridiagonal with for all then the eigenvalues of are real and simple.

Proof.By Theorem 1 the eigenvalues of are those of the symmetric matrix and so are real. The matrix is tridiagonal and irreducible so it is nonderogatory by Theorem 4, which means that its eigenvalues are simple because it is symmetric.

The formula (6) confirms the conclusion of Corollary 5 for tridiagonal Toeplitz matrices.

Corollary 5 guarantees that the eigenvalues are distinct but not that they are well separated. The spacing of the eigenvalues in (6) clearly reduces as increases. Wilkinson constructed a symmetric tridiagonal matrix called , defined by

For example,

Here are the two largest eigenvalues of , as computed by MATLAB.

>> A = anymatrix('matlab/wilkinson',21); >> e = eig(A); e([20 21]) ans = 10.746194182903322 10.746194182903393

These eigenvalues (which are correct to the digits shown) agree almost to the machine precision.

## Notes

Theorem 2 is obtained for symmetric matrices by Malcolm and Palmer (1974), who suggest saving work in computing the LU factorization by setting for once is close enough to the limit.

A sample reference for Theorem 3 is Ikebe (1979), which is one of many papers on inverses of banded matrices.

The eigenvectors of a symmetric tridiagonal matrix satisfy some intricate relations; see Parlett (1998, sec. 7.9).

## References

This is a minimal set of references, which contain further useful references within.

- Murray Dow, Explicit Inverses of Toeplitz and Associated Matrices, ANZIAM J. 44, E185â€“E215, 2003.
- Robert Gregory and David Karney, A Collection of Matrices for Testing Computational Algorithms, Wiley, 1969.
- Yasuhiko Ikebe, On inverses of Hessenberg matrices. Linear Algebra Appl., 24:93â€“97, 1979.
- Michael A. Malcolm and John Palmer, A Fast Method for Solving a Class of Tridiagonal Linear Systems, Comm. ACM 17, 14â€“17, 1974.
- Beresford Parlett, The Symmetric Eigenvalue Problem, Society for Industrial and Applied Mathematics, Philadelphia, PA, USA, 1998.
- J. H. Wilkinson, The Algebraic Eigenvalue Problem, Oxford University Press, 1965, Section 5.45.