The determinant of an matrix
is defined by
where the sum is over all permutations
) of the sequence
and
is the number of inversions in
, that is, the number of pairs
with
. Each term in the sum is a signed product of
entries of
and the product contains one entry taken from each row and one from each column.
The determinant is sometimes written with vertical bars, as .
Three fundamental properties are
The first property is immediate, the second can be proved using properties of permutations, and the third is proved in texts on linear algebra and matrix theory.
An alternative, recursive expression for the determinant is the Laplace expansion
for any , where
denotes the
submatrix of
obtained by deleting row
and column
, and
for a scalar
. This formula is called the expansion by minors because
is a minor of
.
For some types of matrices the determinant is easy to evaluate. If is triangular then
. If
is unitary then
implies
on using (3) and (4). An explicit formula exists for the determinant of a Vandermonde matrix.
The determinant of is connected with the eigenvalues
of
via the property
. Since the eigenvalues are the roots of the characteristic polynomial
, this relation follows by setting
in the expression
For , the determinant is
but already for the determinant is tedious to write down. If one must compute
, the formulas (1) and (5) are too expensive unless
is very small: they have an exponential cost. The best approach is to use a factorization of
involving factors that are triangular or orthogonal, so that the determinants of the factors are easily computed. If
is an LU factorization, with
a permutation matrix,
unit lower triangular, and
upper triangular, then
. As this expression indicates, the determinant is prone to overflow and underflow in floating-point arithmetic, so it may be preferable to compute
.
The determinant features in the formula
for the inverse, where is the adjugate of
(recall that
has
element
). More generally, Cramer’s rule says that the components of the solution to a linear system
are given by
, where
denotes
with its
th column replaced by
. While mathematically elegant, Cramer’s rule is of no practical use, as it is both expensive and numerically unstable in finite precision arithmetic.
Inequalities
A celebrated bound for the determinant of a Hermitian positive definite matrix is Hadamard’s inequality. Note that for such
,
is real and positive (being the product of the eigenvalues, which are real and positive) and the diagonal elements are also real and positive (since
).
Theorem 1 (Hadamard’s inequality). For a Hermitian positive definite matrix
,
with equality if and only if
is diagonal.
Theorem 1 is easy to prove using a Cholesky factorization.
The following corollary can be obtained by applying Theorem 1 to or by using a QR factorization of
.
Corollary 2. For
,
with equality if and only if the columns of
are orthogonal.
Obviously, one can apply the corollary to and obtain the analogous bound with column norms replaced by row norms.
The determinant of can be interpreted as the volume of the parallelepiped
, whose sides are the columns of
. Corollary 2 says that for columns of given lengths the volume is maximized when the columns are orthogonal.
Nearness to Singularity and Conditioning
The determinant characterizes nonsingularity: is singular if and only if
. It might be tempting to use
as a measure of how close a nonsingular matrix
is to being singular, but this measure is flawed, not least because of the sensitivity of the determinant to scaling. Indeed if
is unitary then
can be given any value by a suitable choice of
, yet
is perfectly conditioned:
, where
is the condition number.
To deal with the poor scaling one might normalize the determinant: in view of Corollary 2,
satisfies and
if and only if the columns of
are orthogonal. Birkhoff (1975) calls
the Hadamard condition number. In general,
is not related to the condition number
, but if
has columns of unit
-norm then it can be shown that
(Higham, 2002, Prob. 14.13). Dixon (1984) shows that for classes of
random matrices
that include matrices with elements independently drawn from a normal distribution with mean
, the probability that the inequality
holds tends to as
for any
, so
for large
. This exponential growth is much faster than the growth of
, for which Edelman (1998) showed that for the standard normal distribution,
, where
denotes the mean value. This MATLAB example illustrates these points.
>> rng(1); n = 50; A = randn(n); >> psi = prod(sqrt(sum(A.*A)))/abs(det(A)), kappa2 = cond(A) psi = 5.3632e+10 kappa2 = 1.5285e+02 >> ratio = psi/(n^(0.25)*exp(n/2)) ratio = 2.8011e-01
The relative distance from to the set of singular matrices is equal to the reciprocal of the condition number.
Theorem 3 (Gastinel, Kahan). For
and any subordinate matrix norm,
Notes
Determinants came before matrices, historically. Most linear algebra textbooks make significant use of determinants, but a lot can be done without them. Axler (1995) shows how the theory of eigenvalues can be developed without using determinants.
Determinants have little application in practical computations, but they are a useful theoretical tool in numerical analysis, particularly for proving nonsingularity.
There is a large number of formulas and identities for determinants. Sir Thomas Muir collected many of them in his five-volume magnum opus The Theory of Determinants in the Historical Order of Development, published between 1890 and 1930. Brualdi and Schneider (1983) give concise derivations of many identities using Gaussian elimination, bringing out connections between the identities.
The quantity obtained by modifying the definition (1) of determinant to remove the term is the permanent. The permanent arises in combinatorics and quantum mechanics and is much harder to compute than the determinant: no algorithm is known for computing the permanent in
operations for a polynomial
.
References
This is a minimal set of references, which contain further useful references within.
- Sheldon Axler, Down With Determinants!, Amer. Math. Monthly 102, 139–154, 1995.
- Garrett Birkhoff, Two Hadamard Numbers for Matrices, Comm. ACM 18, 25–29, 1975.
- Richard Brualdi and Hans Schneider, Determinantal Identities: Gauss, Schur, Cauchy, Sylvester, Kronecker, Jacobi, Binet, Laplace, Muir, and Cayley, Linear Algebra Appl. 52–53, 769–791, 1983.
- John Dixon, How Good is Hadamard’s Inequality for Determinants?, Canadian Math. Bulletin 27, 260–264, 1984.
- Alan Edelman, Eigenvalues and Condition Numbers of Random Matrices, SIAM J. Matrix Anal. Appl. 9(4), 543–560, 1988.
- Nicholas J. Higham, Accuracy and Stability of Numerical Algorithms, second edition, Society for Industrial and Applied Mathematics, Philadelphia, PA, USA, 2002.
Related Blog Posts
- The Strange Case of the Determinant of a Matrix of 1s and -1s by Nick Higham and Alan Edelman (2017)
- What Is a Vandermonde Matrix? (2021)
- What Is the Adjugate of a Matrix? (2020)
This article is part of the “What Is” series, available from https://nhigham.com/category/what-is and in PDF form from the GitHub repository https://github.com/higham/what-is.