Does a symmetric positive definite matrix remain positive definite when we set one or more elements to zero? This question arises in thresholding, in which elements of absolute value less than some tolerance are set to zero. Thresholding is used in some applications to remove correlations thought to be spurious, so that only statistically significant ones are retained.
We will focus on the case where just one element is changed and consider an arbitrary target value rather than zero. Given an symmetric positive definite matrix
we define
to be the matrix resulting from adding
to the
and
elements and we ask when is
positive definite. We can write
where is the
th column of the identity matrix. The perturbation
has rank
, with eigenvalues
,
, and
repeated
times. Hence we can write
in the form
, where
and
. Adding
to
causes each eigenvalue to increase or stay the same, while subtracting
decreases or leaves unchanged each eigenvalue. However, more is true: after each of these rank-
perturbations the eigenvalues of the original and perturbed matrices interlace, by Weyl’s theorem. Hence, with the eigenvalues of
ordered as
, we have (Horn and Johnson, Cor. 4.3.7)
Because is positive definite these inequalities imply that
, so
has at most one negative eigenvalue. Since
is the product of the eigenvalues of
this means that
is positive definite precisely when
.
There is a simple expression for , which follows from a lemma of Chan (1984), as explained by Georgescu, Higham, and Peters (2018):
where . Hence the condition for
to be positive definite is
We can factorize
so for
where the endpoints are finite because , like
, is positive definite and so
.
The condition for to remain positive definite when
is set to zero is
, or equivalently
. To check either of these conditions we need just
,
, and
. These elements can be computed without computing the whole inverse by solving the equations
for
, for the
th column
of
, making use of a Cholesky factorization of
.
As an example, we consider the Lehmer matrix, which has
element
for
:
The smallest eigenvalue of is
. Any off-diagonal element except the
element can be zeroed without destroying positive definiteness, and if the
element is zeroed then the new matrix has smallest eigenvalue
. For
and
, the following plot shows in red
and in blue
; the black dots are the endpoints of the closure of the interval
and the vertical black line is the value
. Clearly,
lies outside
, which is why zeroing this element causes a loss of positive definiteness. Note that
also tells us that we can increase
to any number less than
without losing definiteness.
Given a positive definite matrix and a set of elements to be modified we may wish to determine subsets (including a maximal subset) of
for which the modifications preserve definiteness. Efficiently determining these subsets appears to be an open problem.
In practical applications thresholding may lead to an indefinite matrix. Definiteness must then be restored to obtain a valid correlation matrix. One way to do this is to find the nearest correlation matrix in the Frobenius norm such that the zeroed elements remain zero. This can be done by the alternating projections method with a projection to keep the zeroed elements fixed. Since the nearest correlation matrix is positive semidefinite, it is also desirable to to incorporate a lower bound on the smallest eigenvalue, which corresponds to another projection. Both these projections are supported in the algorithm of Higham and Strabić (2016), implemented in the code at https://github.com/higham/anderson-accel-ncm. For the Lehmer matrix, the nearest correlation matrix with zero
element and eigenvalues at least
is (to four significant figures)
A related question is for what patterns of elements that are set to zero is positive definiteness guaranteed to be preserved for all positive definite ? Clearly, setting all the off-diagonal elements to zero preserves definiteness, since the diagonal of a positive definite matrix is positive. Guillot and Rajaratnam (2012) show that the answer to the question is that the new matrix must be a symmetric permutation of a block diagonal matrix. However, for particular
this restriction does not necessarily hold, as the Lehmer matrix example shows.
References
- Tony F. Chan, On the Existence and Computation of
-Factorizations with Small Pivots, Math. Comp. 42(166), 535–547, 1984.
- Dan I. Georgescu, Nicholas J. Higham, and Gareth W. Peters, Explicit Solutions to Correlation Matrix Completion Problems, with an Application to Risk Management and Insurance, Roy. Soc. Open Sci. 5(3), 1–11, 2018
- Dominique Guillot and Bala Rajaratnam, Retaining Positive Definiteness in Thresholded Matrices, Linear Algebra Appl. 436(11), 4143–4160, 2012.
- Nicholas J. Higham and Nataša Strabić, Anderson Acceleration of the Alternating Projections Method for Computing the Nearest Correlation Matrix, Numer. Algorithms 72(4), 1021–1042, 2016. Sections 3.1, 3.2.
- Roger A. Horn and Charles R. Johnson, Matrix Analysis, second edition, Cambridge University Press, 2013. My review of the second edition.