A tridiagonal matrix is a square matrix whose elements are zero away from the main diagonal, the subdiagonal, and the superdiagonal. In other words, it is a banded matrix with upper and lower bandwidths both equal to . It has the form
An important example is the matrix that arises in discretizating the Poisson partial differential equation by a standard five-point operator, illustrated for
by
It is symmetric positive definite, diagonally dominant, a Toeplitz matrix, and an -matrix.
Tridiagonal matrices have many special properties and various algorithms exist that exploit their structure.
Symmetrization
It can be useful to symmetrize a matrix by transforming it with a diagonal matrix. The next result shows when symmetrization is possible by similarity.
Theorem 1. If
is tridiagonal and
for all
then there exists a diagonal
with positive diagonal elements such that
is symmetric, with
element
.
Proof. Let
. Equating
and
elements in the matrix
gives
or
As long as
for all
we can set
and solve (2) to obtain real, positive
,
. The formula for the off-diagonal elements of the symmetrized matrix follows from (1) and (2).
LU Factorization
The LU factors of a tridiagonal matrix are bidiagonal:
The equation gives the recurrence
The recurrence breaks down with division by zero if one of the leading principal submatrices ,
, is singular. In general, partial pivoting must be used to ensure existence and numerical stability, giving a factorization
where
has at most two nonzeros per column and
has an extra superdiagonal. The growth factor
is easily seen to be bounded by
.
For a tridiagonal Toeplitz matrix
the elements of the LU factors converge as if
is diagonally dominant.
Theorem 2. Suppose that
has an LU factorization with LU factors (3) and that
and
. Then
decreases monotonically and
From (4), it follows that under the conditions of Theorem 2, increases monotonically and
. Note that the conditions of Theorem 2 are satisfied if
is diagonally dominant by rows, since
implies
. Note also that if we symmetrize
using Theorem 1 then we obtain the matrix
, which is irreducibly diagonally dominant and hence positive definite if
.
Inverse
The inverse of a tridiagonal matrix is full, in general. For example,
Since an tridiagonal matrix depends on only
parameters, the same must be true of its inverse, meaning that there must be relations between the elements of the inverse. Indeed, in
any
submatrix whose elements lie in the upper triangle is singular, and the
submatrix is also singular. The next result explains this special structure. We note that a tridiagonal matrix is irreducible if
and
are nonzero for all
.
Theorem 3. If
is tridiagonal, nonsingular, and irreducible then there are vectors
,
,
, and
, all of whose elements are nonzero, such that
The theorem says that the upper triangle of the inverse agrees with the upper triangle of a rank- matrix (
) and the lower triangle of the inverse agrees with the lower triangle of another rank-
matrix (
). This explains the singular submatrices that we see in the example above.
If a tridiagonal matrix is reducible with
then it has the block form
where , and so
in which the block is rank
if
. This blocking can be applied recursively until Theorem 1 can be applied to all the diagonal blocks.
The inverse of the Toeplitz tridiagonal matrix is known explicitly; see Dow (2003, Sec. 3.1).
Eigenvalues
The most widely used methods for finding eigenvalues and eigenvectors of Hermitian matrices reduce the matrix to tridiagonal form by a finite sequence of unitary similarity transformations and then solve the tridiagonal eigenvalue problem. Tridiagonal eigenvalue problems also arise directly, for example in connection with orthogonal polynomials and special functions.
The eigenvalues of the Toeplitz tridiagonal matrix in (5) are given by
The eigenvalues are also known for certain variations of the symmetric matrix in which the
and
elements are modified (Gregory and Karney, 1969).
Some special results hold for the eigenvalues of general tridiagonal matrices. A matrix is derogatory if an eigenvalue appears in more than one Jordan block in the Jordan canonical form, and nonderogatory otherwise.
Theorem 4. If
is an irreducible tridiagonal matrix then it is nonderogatory.
Proof. Let
, for any
. The matrix
is lower triangular with nonzero diagonal elements
and hence it is nonsingular. Therefore
has rank at least
for all
. If
were derogatory then the rank of
would be at most
when
is an eigenvalue, so
must be nonderogatory.
Corollary 5. If
is tridiagonal with
for all
then the eigenvalues of
are real and simple.
Proof. By Theorem 1 the eigenvalues of
are those of the symmetric matrix
and so are real. The matrix
is tridiagonal and irreducible so it is nonderogatory by Theorem 4, which means that its eigenvalues are simple because it is symmetric.
The formula (6) confirms the conclusion of Corollary 5 for tridiagonal Toeplitz matrices.
Corollary 5 guarantees that the eigenvalues are distinct but not that they are well separated. The spacing of the eigenvalues in (6) clearly reduces as increases. Wilkinson constructed a symmetric tridiagonal matrix called
, defined by
For example,
Here are the two largest eigenvalues of , as computed by MATLAB.
>> A = anymatrix('matlab/wilkinson',21); >> e = eig(A); e([20 21]) ans = 10.746194182903322 10.746194182903393
These eigenvalues (which are correct to the digits shown) agree almost to the machine precision.
Notes
Theorem 2 is obtained for symmetric matrices by Malcolm and Palmer (1974), who suggest saving work in computing the LU factorization by setting for
once
is close enough to the limit.
A sample reference for Theorem 3 is Ikebe (1979), which is one of many papers on inverses of banded matrices.
The eigenvectors of a symmetric tridiagonal matrix satisfy some intricate relations; see Parlett (1998, sec. 7.9).
References
This is a minimal set of references, which contain further useful references within.
- Murray Dow, Explicit Inverses of Toeplitz and Associated Matrices, ANZIAM J. 44, E185–E215, 2003.
- Robert Gregory and David Karney, A Collection of Matrices for Testing Computational Algorithms, Wiley, 1969.
- Yasuhiko Ikebe, On inverses of Hessenberg matrices. Linear Algebra Appl., 24:93–97, 1979.
- Michael A. Malcolm and John Palmer, A Fast Method for Solving a Class of Tridiagonal Linear Systems, Comm. ACM 17, 14–17, 1974.
- Beresford Parlett, The Symmetric Eigenvalue Problem, Society for Industrial and Applied Mathematics, Philadelphia, PA, USA, 1998.
- J. H. Wilkinson, The Algebraic Eigenvalue Problem, Oxford University Press, 1965, Section 5.45.