An -matrix is a matrix of the form

Here, is the spectral radius of , that is, the largest modulus of any eigenvalue of , and denotes that has nonnegative entries. An -matrix clearly has nonpositive off-diagonal elements. It also has positive diagonal elements, which can be shown using the result that

for any consistent matrix norm:

Although the definition of an -matrix does not specify , we can set it to . Indeed let satisfy and set and . Then , since and for . Furthermore, for a nonnegative matrix the spectral radius is an eigenvalue, by the Perron–Frobenius theorem, so is an eigenvalue of and is an eigenvalue of . Hence .

The concept of -matrix was introduced by Ostrowski in 1937. -matrices arise in a variety of scientific settings, including in finite difference methods for PDEs, input-output analysis in economics, and Markov chains in stochastic processes.

An immediate consequence of the definition is that the eigenvalues of an -matrix lie in the open right-half plane, which means that -matrices are special cases of positive stable matrices. Hence an -matrix is nonsingular and the determinant, being the product of the eigenvalues, is positive. Moreover, since satisfies ,

In fact, nonnegativity of the inverse characterizes -matrices. Define

Theorem 1.A nonsingular matrix is an -matrix if and only if .

Sometimes an -matrix is *defined* to be a matrix with nonpositive off-diagonal elements and a nonnegative inverse. In fact, this condition is just one of a large number of conditions equivalent to a matrix with nonpositive off-diagonal elements being an -matrix, fifty of which are given in Berman and Plemmons (1994, Chap. 6).

It is easy to check from the definitions, or using Theorem 1, that a triangular matrix with positive diagonal and nonpositive off-diagonal is an -matrix. An example is

## Bounding the Norm of the Inverse

An -matrix can be constructed from any nonsingular triangular matrix by taking the comparison matrix. The comparison matrix associated with a general is the matrix

For a nonsingular triangular , is an -matrix, and it easy to show that

where the absolute value is taken componentwise. This bound, and weaker related bounds, can be useful for cheaply bounding the norm of the inverse of a triangular matrix. For example, with denoting the vector of ones, since is nonnegative we have

and can be computed in flops by solving a triangular system, whereas computing costs flops.

More generally, if we have an LU factorization of an -matrix then, since ,

Therefore the norm of the inverse can be computed in flops with two triangular solves, instead of the flops that would be required if were to be formed explicitly.

## Connections with Symmetric Positive Definite Matrices

There are many analogies between M-matrices and symmetric positive definite matrices. For example, every principal submatrix of a symmetric positive definite matrix is symmetric positive definite and every principal submatrix of an -matrix is an -matrix. Indeed if is a principal submatrix of a nonnegative then , which follows from for the -norm (for example). Hence on taking principal submatrices in we have with the same .

A symmetric -matrix is known as a *Stieltjes matrix*, and such a matrix is positive definite. An example of a Stieltjes matrix is minus the second difference matrix (the tridiagonal matrix arising from a central difference discretization of a second derivative), illustrated for by

## LU Factorization

Since the leading principal submatrices of an -matrix have positive determinant it follows that has an LU factorization with having positive diagonal elements. However, more is true, as the next result shows.

Theorem 2.An -matrix has an LU factorization in which and are -matrices.

Proof. We can write

The first stage of LU factorization is

where is the Schur complement of in . The first column of and the first row of are of the form required for a triangular -matrix. We have

Since it follows that . It is easy to see that , and hence Theorem shows that is an -matrix. The result follows by induction.

The question now arises of what can be said about the numerical stability of LU factorization of an -matrix. To answer it we use another characterization of -matrices, that is strictly diagonally dominant by columns for some diagonal matrix with for all , that is,

(This condition can also be written as because of the sign pattern of .) Matrices that are diagonally dominant by columns have the properties that an LU factorization without pivoting exists, the growth factor , and partial pivoting does not require row interchanges. The effect of row scaling on LU factorization is easy to see:

where is unit lower triangular, so that and are the LU factors of . It is easy to see that the growth factor bound of for a matrix diagonally dominant by columns translates into the bound

for an -matrix, which was obtained by Funderlic, Neumann, and Plemmons (1982). Unfortunately, this bound can be large. Consider the matrix

We have

so is an -matrix. The element of the LU factor of is , which means that

and this lower bound can be arbitrarily large. One can verify experimentally that numerical instability is possible when is large, in that the computed LU factors have a large relative residual. We conclude that pivoting is necessary for numerical stability in LU factorization of -matrices.

## Stationery Iterative Methods

A stationery iterative method for solving a linear system is based on a splitting with nonsingular, and has the form . This iteration converges for all starting vectors if . Much interest has focused on *regular splittings*, which are defined as ones for which and . An -matrix has the important property that for every regular splitting, and it follows that the Jacobi iteration, the Gauss-Seidel iteration, and the successive overrelaxation (SOR) iteration (with ) are all convergent for -matrices.

## Matrix Square Root

The principal square root of an -matrix is an -matrix, and it is the unique such square root. An expression for follows from :

This expression does not necessarily provide the best way to compute .

## Singular M-Matrices

The theory of -matrices extends to the case where the condition on is relaxed to in , though the theory is more complicated and extra conditions such as irreducibility are needed for some results. Singular -matrices occur in Markov chains (Berman and Plemmons, 1994, Chapter 8), for example. Let be the transition matrix of a Markov chain. Then is stochastic, that is, nonnegative with unit row sums, so . A nonnegative vector with such that is called a *stationary distribution vector* and is of interest for describing the properties of the Markov chain. To compute we can solve the singular system . Clearly, and , so is a singular -matrix.

## H-Matrices

A more general concept is that of -matrix: is an -matrix if the comparison matrix is an -matrix. Many results for -matrices extend to -matrices. For example, for an -matrix with positive diagonal elements the principal square root exists and is the unique square root that is an -matrix with positive diagonal elements. Also, the growth factor bound holds for any -matrix for which is diagonally dominant by columns.

## References

This is a minimal set of references, which contain further useful references within.

- Abraham Berman and Robert J. Plemmons, Nonnegative Matrices in the Mathematical Sciences, Society for Industrial and Applied Mathematics, Philadelphia, PA, USA, 1994.
- R. E. Funderlic and M. Neumann and R. J. Plemmons Decompositions of Generalized Diagonally Dominant Matrices, Numer. Math. 40, 57–69, 1982.
- Nicholas J. Higham, Accuracy and Stability of Numerical Algorithms, second edition, Society for Industrial and Applied Mathematics, Philadelphia, PA, USA, 2002. (Chapter 8.)
- Nicholas J. Higham, Functions of Matrices: Theory and Computation, Society for Industrial and Applied Mathematics, Philadelphia, PA, USA, 2008. (Section 6.8.3.)

## Related Blog Posts

- What Is a Matrix Square Root? (2020)
- What Is a Symmetric Positive Definite Matrix? (2020)
- What is Numerical Stability? (2020)

This article is part of the “What Is” series, available from https://nhigham.com/category/what-is and in PDF form from the GitHub repository https://github.com/higham/what-is.

Excuse me, can you prove the inequalities for the inveres M-matrix bound, i.e |A^{-1}| \leq M(A)^{-1} ? Thanks in advance

For an M-matrix, M(A) = A so that bound is trivial. The bound always holds for triangular A, as noted above, but does not always hold in general.

Thankyou for the replies, Sir. First of all, i have already read an article about M-Matrices and H-Matrices (https://www.sciencedirect.com/science/article/abs/pii/S0024379519305087). In that article (pg 5), there is a statement :

A well known relationship between an invertible H-matrix and its comparison matrix due to Ostrowski [10] states that, for an invertible H-matrix A ∈ C^{n×n}, the inequality |A^{−1}| ≤ (M(A))^{−1} holds.

I am confused to prove this statement. Any advice?

Please is there any result that shows that the inverse of a strictly diagonally dominant matrix is also strictly diagonally dominant?

I checked this numerically which holds, but could not find any analytic proof to support this.

Many thanks

It’s not true: A = full(gallery(‘tridiag’,5,1,2.5,1)) is a counterexample.

However, see Theorem 6 in https://nhigham.com/2021/04/08/what-is-a-diagonally-dominant-matrix/ for a partial result along these lines.