Schur complement
In linear algebra and the theory of matrices, the Schur complement of a block matrix is defined as follows.
Suppose p, q are nonnegative integers, and suppose A, B, C, D are respectively p × p, p × q, q × p, and q × q matrices of complex numbers. Let [math]\displaystyle{ M = \left[\begin{matrix} A & B \\ C & D \end{matrix}\right] }[/math] so that M is a (p + q) × (p + q) matrix.
If D is invertible, then the Schur complement of the block D of the matrix M is the p × p matrix defined by [math]\displaystyle{ M/D := A - BD^{-1}C. }[/math] If A is invertible, the Schur complement of the block A of the matrix M is the q × q matrix defined by [math]\displaystyle{ M/A := D - CA^{-1}B. }[/math] In the case that A or D is singular, substituting a generalized inverse for the inverses on M/A and M/D yields the generalized Schur complement.
The Schur complement is named after Issai Schur who used it to prove Schur's lemma, although it had been used previously.[1] Emilie Virginia Haynsworth was the first to call it the Schur complement.[2] The Schur complement is a key tool in the fields of numerical analysis, statistics, and matrix analysis.
Background
The Schur complement arises when performing a block Gaussian elimination on the matrix M. In order to eliminate the elements below the block diagonal, one multiplies the matrix M by a block lower triangular matrix on the right as follows: [math]\displaystyle{ \begin{align} &M = \begin{bmatrix} A & B \\ C & D \end{bmatrix} \quad \to \quad \begin{bmatrix} A & B \\ C & D \end{bmatrix} \begin{bmatrix} I_p & 0 \\ -D^{-1}C & I_q \end{bmatrix} = \begin{bmatrix} A - BD^{-1}C & B \\ 0 & D \end{bmatrix}, \end{align} }[/math] where Ip denotes a p×p identity matrix. As a result, the Schur complement [math]\displaystyle{ M/D = A - BD^{-1}C }[/math] appears in the upper-left p×p block.
Continuing the elimination process beyond this point (i.e., performing a block Gauss–Jordan elimination), [math]\displaystyle{ \begin{align} &\begin{bmatrix} A - BD^{-1}C & B \\ 0 & D \end{bmatrix} \quad \to \quad \begin{bmatrix} I_p & -BD^{-1} \\ 0 & I_q \end{bmatrix} \begin{bmatrix} A - BD^{-1}C & B \\ 0 & D \end{bmatrix} = \begin{bmatrix} A - BD^{-1}C & 0 \\ 0 & D \end{bmatrix}, \end{align} }[/math] leads to an LDU decomposition of M, which reads [math]\displaystyle{ \begin{align} M &= \begin{bmatrix} A & B \\ C & D \end{bmatrix} = \begin{bmatrix} I_p & BD^{-1} \\ 0 & I_q \end{bmatrix}\begin{bmatrix} A - BD^{-1}C & 0 \\ 0 & D \end{bmatrix}\begin{bmatrix} I_p & 0 \\ D^{-1}C & I_q \end{bmatrix}. \end{align} }[/math] Thus, the inverse of M may be expressed involving D−1 and the inverse of Schur's complement, assuming it exists, as [math]\displaystyle{ \begin{align} M^{-1} = \begin{bmatrix} A & B \\ C & D \end{bmatrix}^{-1} ={} &\left(\begin{bmatrix} I_p & BD^{-1} \\ 0 & I_q \end{bmatrix} \begin{bmatrix} A - BD^{-1}C & 0 \\ 0 & D \end{bmatrix} \begin{bmatrix} I_p & 0 \\ D^{-1}C & I_q \end{bmatrix} \right)^{-1} \\ ={} &\begin{bmatrix} I_p & 0 \\ -D^{-1}C & I_q \end{bmatrix} \begin{bmatrix} \left(A - BD^{-1}C\right)^{-1} & 0 \\ 0 & D^{-1} \end{bmatrix} \begin{bmatrix} I_p & -BD^{-1} \\ 0 & I_q \end{bmatrix} \\[4pt] ={} &\begin{bmatrix} \left(A - BD^{-1}C\right)^{-1} & -\left(A - BD^{-1}C\right)^{-1} BD^{-1} \\ -D^{-1}C\left(A - BD^{-1}C\right)^{-1} & D^{-1} + D^{-1}C\left(A - BD^{-1}C\right)^{-1}BD^{-1} \end{bmatrix} \\[4pt] ={} &\begin{bmatrix} \left(M/D\right)^{-1} & -\left(M/D\right)^{-1} B D^{-1} \\ -D^{-1}C\left(M/D\right)^{-1} & D^{-1} + D^{-1}C\left(M/D\right)^{-1} B D^{-1} \end{bmatrix}. \end{align} }[/math] The above relationship comes from the elimination operations that involve D−1 and M/D. An equivalent derivation can be done with the roles of A and D interchanged. By equating the expressions for M−1 obtained in these two different ways, one can establish the matrix inversion lemma, which relates the two Schur complements of M: M/D and M/A (see "Derivation from LDU decomposition" in Woodbury matrix identity § Alternative proofs).
Properties
- If p and q are both 1 (i.e., A, B, C and D are all scalars), we get the familiar formula for the inverse of a 2-by-2 matrix:
- [math]\displaystyle{ M^{-1} = \frac{1}{AD-BC} \left[ \begin{matrix} D & -B \\ -C & A \end{matrix}\right] }[/math]
- provided that AD − BC is non-zero.
- In general, if A is invertible, then
- [math]\displaystyle{ \begin{align} M &= \begin{bmatrix} A&B\\C&D \end{bmatrix} = \begin{bmatrix} I_p & 0 \\ CA^{-1} & I_q \end{bmatrix}\begin{bmatrix} A & 0 \\ 0 & D - CA^{-1}B \end{bmatrix}\begin{bmatrix} I_p & A^{-1}B \\ 0 & I_q \end{bmatrix}, \\[4pt] M^{-1} &= \begin{bmatrix} A^{-1} + A^{-1} B (M/A)^{-1} C A^{-1} & - A^{-1} B (M/A)^{-1} \\ - (M/A)^{-1} CA^{-1} & (M/A)^{-1} \end{bmatrix} \end{align} }[/math]
- whenever this inverse exists.
- (Schur's formula) When A, respectively D, is invertible, the determinant of M is also clearly seen to be given by
- [math]\displaystyle{ \det(M) = \det(A) \det\left(D - CA^{-1} B\right) }[/math], respectively
- [math]\displaystyle{ \det(M) = \det(D) \det\left(A - BD^{-1} C\right) }[/math],
- which generalizes the determinant formula for 2 × 2 matrices.
- (Guttman rank additivity formula) If D is invertible, then the rank of M is given by
- [math]\displaystyle{ \operatorname{rank}(M) = \operatorname{rank}(D) + \operatorname{rank}\left(A - BD^{-1} C\right) }[/math]
- (Haynsworth inertia additivity formula) If A is invertible, then the inertia of the block matrix M is equal to the inertia of A plus the inertia of M/A.
- (Quotient identity) [math]\displaystyle{ A/B = ((A/C)/(B/C)) }[/math].[3]
- The Schur complement of a Laplacian matrix is also a Laplacian matrix.[4]
Application to solving linear equations
The Schur complement arises naturally in solving a system of linear equations such as[5]
[math]\displaystyle{ \begin{bmatrix} A & B \\ C & D \end{bmatrix}\begin{bmatrix} x \\ y \end{bmatrix} = \begin{bmatrix} u \\ v \end{bmatrix} }[/math].
Assuming that the submatrix [math]\displaystyle{ A }[/math] is invertible, we can eliminate [math]\displaystyle{ x }[/math] from the equations, as follows.
[math]\displaystyle{ x = A^{-1} (u - By) }[/math].
Substituting this expression into the second equation yields
- [math]\displaystyle{ \left(D - CA^{-1}B\right)y = v-CA^{-1}u }[/math].
We refer to this as the reduced equation obtained by eliminating [math]\displaystyle{ x }[/math] from the original equation. The matrix appearing in the reduced equation is called the Schur complement of the first block [math]\displaystyle{ A }[/math] in [math]\displaystyle{ M }[/math]:
- [math]\displaystyle{ S \ \overset{\underset{\mathrm{def}}{}}{=}\ D - CA^{-1}B }[/math].
Solving the reduced equation, we obtain
- [math]\displaystyle{ y = S^{-1} \left(v-CA^{-1}u\right) }[/math].
Substituting this into the first equation yields
- [math]\displaystyle{ x = \left(A^{-1} + A^{-1} B S^{-1} C A^{-1}\right) u - A^{-1} B S^{-1} v }[/math].
We can express the above two equation as:
- [math]\displaystyle{ \begin{bmatrix} x \\ y \end{bmatrix} = \begin{bmatrix} A^{-1} + A^{-1} B S^{-1} C A^{-1} & -A^{-1} B S^{-1} \\ -S^{-1} C A^{-1} & S^{-1} \end{bmatrix} \begin{bmatrix} u \\ v \end{bmatrix} }[/math].
Therefore, a formulation for the inverse of a block matrix is:
- [math]\displaystyle{ \begin{bmatrix} A & B \\ C & D \end{bmatrix}^{-1} = \begin{bmatrix} A^{-1} + A^{-1} B S^{-1} C A^{-1} & - A^{-1} B S^{-1} \\ -S^{-1} C A^{-1} & S^{-1} \end{bmatrix} = \begin{bmatrix} I_p & -A^{-1}B\\ & I_q \end{bmatrix}\begin{bmatrix} A^{-1} & \\ & S^{-1} \end{bmatrix}\begin{bmatrix} I_p & \\ -CA^{-1} & I_q \end{bmatrix} }[/math].
In particular, we see that the Schur complement is the inverse of the [math]\displaystyle{ 2,2 }[/math] block entry of the inverse of [math]\displaystyle{ M }[/math].
In practice, one needs [math]\displaystyle{ A }[/math] to be well-conditioned in order for this algorithm to be numerically accurate.
In electrical engineering this is often referred to as node elimination or Kron reduction.
Applications to probability theory and statistics
Suppose the random column vectors X, Y live in Rn and Rm respectively, and the vector (X, Y) in Rn + m has a multivariate normal distribution whose covariance is the symmetric positive-definite matrix
- [math]\displaystyle{ \Sigma = \left[\begin{matrix} A & B \\ B^\mathrm{T} & C\end{matrix}\right], }[/math]
where [math]\displaystyle{ A \in \mathbb{R}^{n \times n} }[/math] is the covariance matrix of X, [math]\displaystyle{ C \in \mathbb{R}^{m \times m} }[/math] is the covariance matrix of Y and [math]\displaystyle{ B \in \mathbb{R}^{n \times m} }[/math] is the covariance matrix between X and Y.
Then the conditional covariance of X given Y is the Schur complement of C in [math]\displaystyle{ \Sigma }[/math]:[6]
- [math]\displaystyle{ \begin{align} \operatorname{Cov}(X \mid Y) &= A - BC^{-1}B^\mathrm{T} \\ \operatorname{E}(X \mid Y) &= \operatorname{E}(X) + BC^{-1}(Y - \operatorname{E}(Y)) \end{align} }[/math]
If we take the matrix [math]\displaystyle{ \Sigma }[/math] above to be, not a covariance of a random vector, but a sample covariance, then it may have a Wishart distribution. In that case, the Schur complement of C in [math]\displaystyle{ \Sigma }[/math] also has a Wishart distribution.[citation needed]
Conditions for positive definiteness and semi-definiteness
Let X be a symmetric matrix of real numbers given by [math]\displaystyle{ X = \left[\begin{matrix} A & B \\ B^\mathrm{T} & C\end{matrix}\right]. }[/math] Then
- If A is invertible, then X is positive definite if and only if A and its complement X/A are both positive definite:[1]:34
- [math]\displaystyle{ X \succ 0 \Leftrightarrow A \succ 0, X/A = C - B^\mathrm{T} A^{-1} B \succ 0. }[/math]
- If C is invertible, then X is positive definite if and only if C and its complement X/C are both positive definite:
- [math]\displaystyle{ X \succ 0 \Leftrightarrow C \succ 0, X/C = A - B C^{-1} B^\mathrm{T} \succ 0. }[/math]
- If A is positive definite, then X is positive semi-definite if and only if the complement X/A is positive semi-definite:[1]:34
- [math]\displaystyle{ \text{If } A \succ 0, \text{ then } X \succeq 0 \Leftrightarrow X/A = C - B^\mathrm{T} A^{-1} B \succeq 0. }[/math]
- If C is positive definite, then X is positive semi-definite if and only if the complement X/C is positive semi-definite:
- [math]\displaystyle{ \text{If } C \succ 0,\text{ then } X \succeq 0 \Leftrightarrow X/C = A - B C^{-1} B^\mathrm{T} \succeq 0. }[/math]
The first and third statements can be derived[5] by considering the minimizer of the quantity [math]\displaystyle{ u^\mathrm{T} A u + 2 v^\mathrm{T} B^\mathrm{T} u + v^\mathrm{T} C v, \, }[/math] as a function of v (for fixed u).
Furthermore, since [math]\displaystyle{ \left[\begin{matrix} A & B \\ B^\mathrm{T} & C \end{matrix}\right] \succ 0 \Longleftrightarrow \left[\begin{matrix} C & B^\mathrm{T} \\ B & A \end{matrix}\right] \succ 0 }[/math] and similarly for positive semi-definite matrices, the second (respectively fourth) statement is immediate from the first (resp. third) statement.
There is also a sufficient and necessary condition for the positive semi-definiteness of X in terms of a generalized Schur complement.[1] Precisely,
- [math]\displaystyle{ X \succeq 0 \Leftrightarrow A \succeq 0, C - B^\mathrm{T} A^g B \succeq 0, \left(I - A A^{g}\right)B = 0 \, }[/math] and
- [math]\displaystyle{ X \succeq 0 \Leftrightarrow C \succeq 0, A - B C^g B^\mathrm{T} \succeq 0, \left(I - C C^g\right)B^\mathrm{T} = 0, }[/math]
where [math]\displaystyle{ A^g }[/math] denotes a generalized inverse of [math]\displaystyle{ A }[/math].
See also
- Woodbury matrix identity
- Quasi-Newton method
- Haynsworth inertia additivity formula
- Gaussian process
- Total least squares
References
- ↑ 1.0 1.1 1.2 1.3 Zhang, Fuzhen (2005). Zhang, Fuzhen. ed. The Schur Complement and Its Applications. Numerical Methods and Algorithms. 4. Springer. doi:10.1007/b105056. ISBN 0-387-24271-6.
- ↑ Haynsworth, E. V., "On the Schur Complement", Basel Mathematical Notes, #BNB 20, 17 pages, June 1968.
- ↑ Crabtree, Douglas E.; Haynsworth, Emilie V. (1969). "An identity for the Schur complement of a matrix" (in en). Proceedings of the American Mathematical Society 22 (2): 364–366. doi:10.1090/S0002-9939-1969-0255573-1. ISSN 0002-9939. https://www.ams.org/proc/1969-022-02/S0002-9939-1969-0255573-1/.
- ↑ Devriendt, Karel (2022). "Effective resistance is more than distance: Laplacians, Simplices and the Schur complement" (in en). Linear Algebra and Its Applications 639: 24–49. doi:10.1016/j.laa.2022.01.002. https://linkinghub.elsevier.com/retrieve/pii/S0024379522000039.
- ↑ 5.0 5.1 Boyd, S. and Vandenberghe, L. (2004), "Convex Optimization", Cambridge University Press (Appendix A.5.5)
- ↑ von Mises, Richard (1964). "Chapter VIII.9.3". Mathematical theory of probability and statistics. Academic Press. ISBN 978-1483255385. https://archive.org/details/mathematicaltheo0057vonm.
Original source: https://en.wikipedia.org/wiki/Schur complement.
Read more |