Matrices arising in applications often have diagonal elements that are large relative to the off-diagonal elements. In the context of a linear system this corresponds to relatively weak interactions between the different unknowns. We might expect a matrix with a large diagonal to be assured of certain properties, such as nonsingularity. However, to ensure nonsingularity it is not enough for each diagonal element to be the largest in its row. For example, the matrix
is singular because is a null vector. A useful definition of a matrix with large diagonal requires a stronger property.
A matrix is diagonally dominant by rows if
It is strictly diagonally dominant by rows if strict inequality holds in (2) for all .
is (strictly) diagonally dominant by columns if
is (strictly) diagonally dominant by rows.
Diagonal dominance on its own is not enough to ensure nonsingularity, as the matrix (1) shows. Strict diagonal dominance does imply nonsingularity, however.
Theorem 1.
If
is strictly diagonally dominant by rows or columns then it is nonsingular.
Proof. Since
is nonsingular if and only if
is nonsingular, it suffices to consider diagonal dominance by rows. For any nonzero
let
and choose
so that
. Then the
th equation of
can be written
which gives
Using (2), we have
Therefore
and so
is nonsingular.
Diagonal dominance plus two further conditions is enough to ensure nonsingularity. We need the notion of irreducibility. A matrix is irreducible if there does not exist a permutation matrix
such that
with and
square matrices. Irreducibility is equivalent to the directed graph of
being strongly connected.
Theorem 2.
If
is irreducible and diagonally dominant by rows with strict inequality in
for some
then it is nonsingular.
Proof. The proof is by contradiction. Suppose there exists
such that
. Define
The
th equation of
can be written
Hence for
,
The set
is nonempty, because if it were empty then we would have
for all
and if there is strict inequality in
for
, then putting
in (4) would give
, which is a contradiction. Hence as long as
for some
, we obtain
, which contradicts the diagonal dominance. Therefore we must have
for all
and all
. This means that all the rows indexed by
have zeros in the columns indexed by
, which means that
is reducible. This is a contradiction, so
must be nonsingular.
The obvious analogue of Theorem 2 holds for column diagonal dominance.
As an example, the symmetric tridiagonal matrix (minus the second difference matrix)
is row diagonally dominant with strict inequality in the first and last diagonal dominance relations. It can also be shown to be irreducible and so it is nonsingular by Theorem 2. If we replace or
by
, then
remains nonsingular by the same argument. What if we replace both
and
by
? We can answer this question by using an observation of Strang. If we define the rectangular matrix
then and
Since in general and
have the same nonzero eigenvalues, we conclude that
, where
denotes the spectrum. Hence
is symmetric positive definite and
is singular and symmetric positive semidefinite.
Relation to Gershgorin’s Theorem
Theorem 1 can be used to obtain information about the location of the eigenvalues of a matrix. Indeed if is an eigenvalue of
then
is singular and hence cannot be strictly diagonally dominant, by Theorem 1. So
cannot be true for all
. Gershgorin’s theorem is simply a restatement of this fact.
Theorem 3 (Gershgorin’s theorem).
The eigenvalues of
lie in the union of the
discs in the complex plane
If is symmetric with positive diagonal elements and satisfies the conditions of Theorem 1 or Theorem 2 then it is positive definite. Indeed the eigenvalues are real and so in Gershgorin’s theorem the discs are intervals and
, so
, so the eigenvalues are nonnegative, and hence positive since nonzero. This provides another proof that the matrix
in (5) is positive definite.
Generalized Diagonal Dominance
In some situations is not diagonally dominant but a row or column scaling of it is. For example, the matrix
is not diagonally dominant by rows or columns but
is strictly diagonally dominant by rows.
A matrix is generalized diagonally dominant by rows if
is diagonally dominant by rows for some diagonal matrix
with
for all
, that is, if
It is easy to see that if is irreducible and there is strictly inequality in (6) for some
then
is nonsingular by Theorem 2.
It can be shown that is generalized diagonally dominant by rows if and only if it is an
-matrix, where an
-matrix is a matrix for which the comparison matrix
, defined by
is an -matrix (see What Is an M-Matrix?).
Block Diagonal Dominance
A matrix is block diagonally dominant by rows if, for a given norm and block
partitioning
, the diagonal blocks
are all nonsingular and
is block diagonally dominant by columns if
is block diagonally dominant by rows. If the blocks are all
then block diagonal dominance reduces to the usual notion of diagonal dominance. Block diagonal dominance holds for certain block tridiagonal matrices arising in the discretization of PDEs.
Analogues of Theorems 1 and 2 giving conditions under which block diagonal dominance implies nonsingularity are given by Feingold and Varga (1962).
Bounding the Inverse
If a matrix is strictly diagonally dominant then we can bound its inverse in terms of the minimum amount of diagonal dominance. For full generality, we state the bound in terms of generalized diagonal dominance.
Theorem 4.
If
and
is strictly diagonally dominant by rows for a diagonal matrix
with
for all
, then
where
.
Proof. Assume first that
. Let
satisfy
and let
. Applying (3) gives
. The result is obtained on applying this bound to
and using
.
.
Another bound for when
is strictly diagonally dominant by rows can be obtained by writing
, where
,
, and
for
. It is easy to see that
, which gives another proof that
is nonsingular. Then
This bound implies that , so in view of its sign pattern
is an
-matrix, which essentially proves one direction of the
-matrix equivalence in the previous section. The same bound holds if
is diagonally dominant by columns, by writing
.
An upper bound also holds for block diagonal dominance.
Theorem 5.
If
is block diagonally dominant by rows then
where
.
It is interesting to note that the inverse of a strictly row diagonally dominant matrix enjoys a form of diagonal dominance, namely that the largest element in each column is on the diagonal.
Theorem 6.
If
is strictly diagonally dominant by rows then
satisfies
for all
.
Proof. For
we have
. Let
. Taking absolute values in
gives
or
, since
. This inequality holds for all
, so we must have
, which gives the result.
Historical Remarks
Theorems 1 and 2 have a long history and have been rediscovered many times. Theorem 1 was first stated by LĂ©vy (1881) with additional assumptions. In a short but influential paper, Taussky (1949) pointed out the recurring nature of the theorems and gave simple proofs (our proof of Theorem 2 is Taussky’s). Schneider (1977) attributes the surge in interest in matrix theory in the 1950s and 1960s to Taussky’s paper and a few others by her, Brauer, Ostrowski, and Wielandt. The history of Gershgorin’s theorem (published in 1931) is intertwined with that of Theorems 1 and 2; see Varga’s 2004 book for details.
Theorems 4 and 5 are from Varah (1975) and Theorem 6 is from Ostrowski (1952).
References
This is a minimal set of references, which contain further useful references within.
- David G. Feingold and Richard S. Varga, Block Diagonally Dominant Matrices and Generalizations of the Gerschgorin Circle Theorem, Pacific J. Math. 12(4), 1241–1250, 1962.
- A. M. Ostrowski, Note on Bounds for Determinants with Dominant Principal Diagonal, Proc. Amer. Math. Soc. 3, 260–30, 1952.
- Hans Schneider, Olga Taussky-Todd’s Influence on Matrix Theory and Matrix Theorists: A Discursive Personal Tribute, Linear and Multilinear Algebra 5, 197–224, 1977.
- Olga Taussky, A Recurring Theorem on Determinants, Amer. Math. Monthly 56(2), 672–676, 1949.
- J. M. Varah, A Lower Bound for the Smallest Singular Value of a Matrix, Linear Algebra Appl. 11, 3–5, 1975.
- Richard Varga, Geršgorin and His Circles, Springer-Verlag, Berlin, 2004.
Related Blog Posts
This article is part of the “What Is” series, available from https://nhigham.com/category/what-is and in PDF form from the GitHub repository https://github.com/higham/what-is.
Nick, I guess one needs the strict inequality in Equation (2).
(2) is diagonal dominance. (2) with strict inequality for all i is strict diagonal dominance.
Ok, sorry. I think I went too quickly.
Hello Nick,
Nice post. There is an important class of diagonally dominant (DD) matrices that just miss being M-matrices. I’ll refer to them as Q matrices, the name bestowed upon them by probabilists in their study of continuous-time Markov chains. Like M-matrices, the diagonal elements are positive and the off-diagonal elements are non-positive. But they are singular. Are you aware of a specific name for this class of DD matrices outside of Q matrices?
Hi Rich. What you are describing sounds like minus a transition intensity matrix, which has zero row sums and which comes up as a generator for a Markov chain.