Matrix Equations

Go to: Introduction, Notation, Index



In all the equations below, x, y, z, X, Y and Z are the unknown vectors or matrices.


Conic Equation

A conic, or conic section, is the locus of points satisfying the quadratic equation [x; 1]TS[x; 1]=0 where x[2#1] is a 2-dimentional real vector and S[3#3] is a real symmetric matrix; all non-zero multiples of S result in the same conic. It is convenient to partition S as S=[A[2#2] b; bT c]=[RDRT  b; bT c] where R is a 2#2 rotation matrix and  D=diag(d) where d contains the eigenvalues of A ordered so that (d1-d2)det(S) ≥ 0. The pseudoinverse, D+, is formed by inverting all the non-zero elements of D.

The conic equation can be classified as one of the following cases (where k is an arbitrary multiple):

  Case det(S) det(A) tr(A)det(S) A bTb - tr(A)c c Conic type
1.  ≠0 <0 Hyperbola
  2. ≠0 0     Parabola
3. ≠0 >0 <0 kI   Circle: radius √(bTb/det(A) - 2c/tr(A)) with centre at x=-A-1b.
4. ≠0 >0 <0 kI   Ellipse: Centre at x=-A-1b. Eccentricity e=√(1-d2/d1). Semi-major axis √((bTA-1b - c) /d2). Semi-minor axis √((bTA-1b - c) /d1). Major axis ([I; bTA-1]R[1; 0])T [x; 1]=0.  Minor axis ([I; bTA-1]R[0; 1])T [x; 1]=0.
5. ≠0 >0 ≥0 [Empty]
  6. 0 <0         Two intersecting lines: ([I; bTA-1]R[+√|d1|; ±√|d1|])T [x; 1]=0. The lines intersect at x=-A-1b.
  7. 0 >0         A single point: x=-A-1b
  8. 0 0   0 <0   [Empty]
9. 0 0 0 0 A single line:  [Rd; bTRD+d]T[x; 1]=0.
10. 0 0 0 >0 Two parallel lines:  [Rd; bTRD+d]T[x; 1]=±√(bTb - tr(A)c).
11. 0 0 0 0 ≠0 [Empty]
12. 0 0 0 0 0 Entire Plane
  13. 0 0   0 >0   A single line: [2b; c]T [x; 1]=0

det(S)=det(A)c-bT ADJ(A) b. If det(A)≠0, this equals det(A)(c-bTA-1b) which implies that, for cases 6 and 7 above, c=bTA-1b.

If det(A)=0 (case 2 and cases 8 to 13) then AppT for some p; it follows that the diagonal elements of A cannot have opposite signs and that det(S)=±det([p b])2 so det(S)=0  (cases 8 to 13)  iff either b=0 or A=kbbT for some k.

If  lT[x; 1]=0 and mT[x; 1]=0 are two lines then their intersection is at (l × m)/det([l1 l2; m1 m2]) unless the determinant is zero in which case the lines are parallel.

According to Wikipedia-Ellipse, semi-major/minor axes are -√(-2 det(S)(tr(A)±√(tr(A)2-4 det(A))))/(2det(A))

The conic is called degenerate iff det(S)=0 (cases 6 to 13). Additional details about each case are:

  1. Hyperbola (e.g. S=[1 0 0; 0 -1 0; 0 0 1]).
  2. Parabola (e.g. S=[1 0 0; 0 0 1; 0 1 0]).
  3. Circle (e.g. S=[1 0 0; 0 1 0; 0 0 -1]).
  4. Ellipse (e.g. S=[1 0 0; 0 4 0; 0 0 -1]).
  5. [Empty] (e.g. S=[1 0 0; 0 1 0; 0 0 1]). The conic equation has no solutions.
  6. Two intersecting lines (e.g. S=[1 0 0; 0 -1 0; 0 0 0]). In this case, c=bTA-1b
  7. Single Point (e.g. S=[1 0 0; 0 1 0; 0 0 0]). In this case, c=bTA-1b and the only solution to the conic equation is x=-A-1b.
  8. [Empty] (e.g. S=[1 0 0; 0 0 0; 0 0 1]). The conic equation has no solutions.
  9. Single Line (e.g. S=[1 0 0; 0 0 0; 0 0 0]). The equation of the line is [Rd; dTD+RTb]T[x; 1]=0
  10. Two Parallel Lines (e.g. S=[1 0 0; 0 0 0; 0 0 -1]). The equation of the lines are [Rd; dTD+RTb]T[x; 1]==±√ (bTb - tr(A)c)..
  11. [Empty] (e.g. S=[0 0 0; 0 0 0; 0 0 1]). The conic equation has no solutions.
  12. Entire Plane (S=0). All values of x satisfy the conic equation when S=0.
  13. Single Line (e.g. S=[0 0 0; 0 0 1; 0 1 0]). The equation of the line is [2b; c]T[x; 1]=0.

Standard Form

 Using the rotation matrix, R, from the eigendecomposition A=RDRT, we can perform a rotation+translation coordinate transform [y; 1]=T-1[x; 1] for T=[R t; 0T 1]=[I t; 0T 1][R 0; 0T 1] and  T-1=[RT -RTt; 0T 1]=[RT 0; 0T 1][I -t; 0T 1] with  t = -A+b = -RD+RTb where the pseudoinverse, D+, is formed by inverting all the non-zero elements of the diagonal matrix D. This transformation is an isometry (preserves lengths and angles). The transformation preserves det(S), det(A), tr(A) and possibly either (a) (bTA-1b - c) when det(A)≠0 (cases 1 and 3 to 7) or (b) (bTb - tr(A)c) when det(S)=det(A)=0 (cases 8 to 13).

The conic equation now becomes  [y; 1]TP[y; 1]=0 where P = TTST = [RT 0; tT 1][A b; bT c][R t; 0T 1] = [D e; eT f] where e = DRTt+RTb = (I-DD+)RTb and f = tTRDRTt+2bTt+c = c - bTRD+RTb. Note that if A is nonsingular then DD+=I and so e=0. The inverse equation is S = T-TPT-1 = [A b; bT c] where A = RDRT, b = R(e-DRTt) and c = tTRDRTt-2tTRe+f.

 If lT[x; 1]=0 is a line then, the corresponding transformed line is mT[y; 1]=0 where m=TTl.

The matrix P has at most three distinct non-zero elements (together with symmetric counterparts):


Discrete-time Lyapunov Equation

The discrete-time Lyapunov equation is AXAH - X + Q = 0 where Q is hermitian. This is a special case of the Stein equation.

The equivalent equation for continuous-time systems is the Lyapunov equation.


Discrete Riccati Equation

The discrete Riccati equation is the quadratic equation [A, X: n#n; B: n#m; C: m#n; R, Q: hermitian] X = AHXA - (C+BHXA)H(R+BHXB)-1(C+BHXA) + Q


Quadratic Form Optimization

Suppose H[n#n]=UDUH is hermitian, U is unitary and D=diag(d)=diag(eig(H)) contains the eigenvalues in decreasing order. Then the corresponding quadratic form is the real-valued expression xHHx.

We can generalize the Rayleigh-Ritz theorem to multiple dimensions in either of two ways which surprisingly turn out to be equivalent. If W is +ve definite Hermitian and B is Hermitian, then

where d are the eigenvalues of W-1B sorted into decreasing order and these bounds are attained by taking the columns of X to be the corresponding eigenvectors.

Linear Discriminant Analysis (LDA): If vectors x are randomly generated from a number of classes with B the covariance of the class means and W the average covariance within each class, then tr((XHWX)-1 XHBX) and det((XHWX)-1 XHBX) are two alternative measures of class separability. We can find a dimension-reducing transformation that maximizes separability by taking y = ATx where the columns of A[k#n] are the eigenvectors of W-1B corresponding to the k largest eigenvalues. This choice maximizes both separability measures for any given k.


Linear Equation

A linear equation has the form Ax - b = 0.

Exact Solution

Least Squares solutions

If there is no exact solution, we can find the x that minimizes d = ||Ax-b|| = (Ax - b)H(Ax - b) .

Recursive Least Squares

We can express the least squares solution to the augmented equation [A; U]y - [b; v] = 0 in terms of the least squares solution to Ax - b = 0.

[rank(Am#n)=n] The least squares solution to the is y = x + K(v-Ux) where x is the least squares solution to Ax-b=0 and K = (AHA)-1UH(I+U(AHA)-1UH)-1. The inverse of the augmented grammian is given by ([A; U]H[A; U])-1 = (AHA)-1-KU(AHA)-1. Thus finding the least squares solution of the augmented equation requires the inversion of a matrix, (I+U(AHA)-1UH), whose dimension equals the number of rows of U instead of the number of rows of  [A; U]. The process is particularly simple if U has only one row. The computation may be reduced at the expense of numerical stability by calculating (AHA)-1UH as (U(AHA)-1)H.


Lyapunov Equation

The (continuous) Lyapunov equation is AX + XAH + Q = 0 where Q is hermitian. This is a special case of the Sylvester equation.

The equivalent equation for discrete-time systems is the Stein equation.


Riccati Equation

The (continuous) Riccati equation is the quadratic equation [A, X, C, D: n#n; C, D: hermitian] XDX + XA + AHX - C = 0


Stein Equation

A Stein equation has the form AXB - X + Q = 0.


Sylvester Equation

The Sylvester equation is AX + XB + Q = 0


This page is part of The Matrix Reference Manual. Copyright © 1998-2022 Mike Brookes, Imperial College, London, UK. See the file gfl.html for copying instructions. Please send any comments or suggestions to "mike.brookes" at "imperial.ac.uk".
Updated: $Id: equation.html 11291 2021-01-05 18:26:10Z dmb $