Specht's theorem

From HandWiki
Revision as of 22:00, 6 February 2024 by JTerm (talk | contribs) (link)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Short description: Gives a necessary and sufficient condition for two complex matrices to be unitarily equivalent

In mathematics, Specht's theorem gives a necessary and sufficient condition for two complex matrices to be unitarily equivalent. It is named after Wilhelm Specht, who proved the theorem in 1940.[1]

Two matrices A and B with complex number entries are said to be unitarily equivalent if there exists a unitary matrix U such that B = U *AU.[2] Two matrices which are unitarily equivalent are also similar. Two similar matrices represent the same linear map, but with respect to a different basis; unitary equivalence corresponds to a change from an orthonormal basis to another orthonormal basis.

If A and B are unitarily equivalent, then tr AA* = tr BB*, where tr denotes the trace (in other words, the Frobenius norm is a unitary invariant). This follows from the cyclic invariance of the trace: if B = U *AU, then tr BB* = tr U *AUU *A*U = tr AUU *A*UU * = tr AA*, where the second equality is cyclic invariance.[3]

Thus, tr AA* = tr BB* is a necessary condition for unitary equivalence, but it is not sufficient. Specht's theorem gives infinitely many necessary conditions which together are also sufficient. The formulation of the theorem uses the following definition. A word in two variables, say x and y, is an expression of the form

[math]\displaystyle{ W(x,y) = x^{m_1} y^{n_1} x^{m_2} y^{n_2} \cdots x^{m_p}, }[/math]

where m1, n1, m2, n2, …, mp are non-negative integers. The degree of this word is

[math]\displaystyle{ m_1 + n_1 + m_2 + n_2 + \cdots + m_p. }[/math]

Specht's theorem: Two matrices A and B are unitarily equivalent if and only if tr W(A, A*) = tr W(B, B*) for all words W.[4]

The theorem gives an infinite number of trace identities, but it can be reduced to a finite subset. Let n denote the size of the matrices A and B. For the case n = 2, the following three conditions are sufficient:[5]

[math]\displaystyle{ \operatorname{tr} \, A = \operatorname{tr} \, B, \quad \operatorname{tr} \, A^2 = \operatorname{tr} \, B^2, \quad\text{and}\quad \operatorname{tr} \, AA^* = \operatorname{tr} \, BB^*. }[/math]

For n = 3, the following seven conditions are sufficient:

[math]\displaystyle{ \begin{align} &\operatorname{tr} \, A = \operatorname{tr} \, B, \quad \operatorname{tr} \, A^2 = \operatorname{tr} \, B^2, \quad \operatorname{tr} \, AA^* = \operatorname{tr} \, BB^*, \quad \operatorname{tr} \, A^3 = \operatorname{tr} \, B^3, \\ &\operatorname{tr} \, A^2 A^* = \operatorname{tr} \, B^2 B^*, \quad \operatorname{tr} \, A^2 (A^*)^2 = \operatorname{tr} \, B^2 (B^*)^2, \quad\text{and}\quad \operatorname{tr} \, A^2 (A^*)^2 A A^* = \operatorname{tr} \, B^2 (B^*)^2 B B^*. \end{align} }[/math]  [6]

For general n, it suffices to show that tr W(A, A*) = tr W(B, B*) for all words of degree at most

[math]\displaystyle{ n \sqrt{\frac{2n^2}{n-1} + \frac14} + \frac{n}2 - 2. }[/math]  [7]

It has been conjectured that this can be reduced to an expression linear in n.[8]

Notes

  1. (Specht 1940)
  2. (Horn Johnson), Definition 2.2.1
  3. (Horn Johnson), Theorem 2.2.2
  4. (Horn Johnson), Theorem 2.2.6
  5. (Horn Johnson), Theorem 2.2.8
  6. (Sibirskiǐ 1976), p. 260, quoted by (Đoković Johnson)
  7. (Pappacena 1997), Theorem 4.3
  8. (Freedman Gupta), p. 160

References