Welch bounds

From HandWiki

In mathematics, Welch bounds are a family of inequalities pertinent to the problem of evenly spreading a set of unit vectors in a vector space. The bounds are important tools in the design and analysis of certain methods in telecommunication engineering, particularly in coding theory. The bounds were originally published in a 1974 paper by L. R. Welch.[1]

Mathematical statement

If [math]\displaystyle{ \{x_1,\ldots,x_m\} }[/math] are unit vectors in [math]\displaystyle{ \mathbb{C}^n }[/math], define [math]\displaystyle{ c_\max = \max_{i\neq j} |\langle x_i, x_j \rangle| }[/math], where [math]\displaystyle{ \langle\cdot,\cdot\rangle }[/math] is the usual inner product on [math]\displaystyle{ \mathbb{C}^n }[/math]. Then the following inequalities hold for [math]\displaystyle{ k=1,2,\dots }[/math]:[math]\displaystyle{ (c_\max)^{2k} \geq \frac{1}{m-1} \left[ \frac{m}{\binom{n+k-1}{k}}-1 \right]. }[/math]Welch bounds are also sometimes stated in terms of the averaged squared overlap between the set of vectors. In this case, one has the inequality[2][3][4][math]\displaystyle{ \frac{1}{m^2}\sum_{i,j=1}^m |\langle x_i,x_j\rangle|^{2k} \ge \frac{1}{\binom{n+k-1}{k}}. }[/math]

Applicability

If [math]\displaystyle{ m\leq n }[/math], then the vectors [math]\displaystyle{ \{x_i\} }[/math] can form an orthonormal set in [math]\displaystyle{ \mathbb{C}^n }[/math]. In this case, [math]\displaystyle{ c_\max=0 }[/math] and the bounds are vacuous. Consequently, interpretation of the bounds is only meaningful if [math]\displaystyle{ m\gt n }[/math]. This will be assumed throughout the remainder of this article.

Proof for k = 1

The "first Welch bound," corresponding to [math]\displaystyle{ k=1 }[/math], is by far the most commonly used in applications. Its proof proceeds in two steps, each of which depends on a more basic mathematical inequality. The first step invokes the Cauchy–Schwarz inequality and begins by considering the [math]\displaystyle{ m\times m }[/math] Gram matrix [math]\displaystyle{ G }[/math] of the vectors [math]\displaystyle{ \{x_i\} }[/math]; i.e.,

[math]\displaystyle{ G=\left[ \begin{array}{ccc} \langle x_1, x_1 \rangle & \cdots & \langle x_1, x_m \rangle \\ \vdots & \ddots & \vdots \\ \langle x_m, x_1 \rangle & \cdots & \langle x_m, x_m \rangle \end{array}\right] }[/math]

The trace of [math]\displaystyle{ G }[/math] is equal to the sum of its eigenvalues. Because the rank of [math]\displaystyle{ G }[/math] is at most [math]\displaystyle{ n }[/math], and it is a positive semidefinite matrix, [math]\displaystyle{ G }[/math] has at most [math]\displaystyle{ n }[/math] positive eigenvalues with its remaining eigenvalues all equal to zero. Writing the non-zero eigenvalues of [math]\displaystyle{ G }[/math] as [math]\displaystyle{ \lambda_1,\ldots,\lambda_r }[/math] with [math]\displaystyle{ r\leq n }[/math] and applying the Cauchy-Schwarz inequality to the inner product of an [math]\displaystyle{ r }[/math]-vector of ones with a vector whose components are these eigenvalues yields

[math]\displaystyle{ (\mathrm{Tr}\;G)^2 = \left( \sum_{i=1}^r \lambda_i \right)^2 \leq r \sum_{i=1}^r \lambda_i^2 \leq n \sum_{i=1}^m \lambda_i^2 }[/math]

The square of the Frobenius norm (Hilbert–Schmidt norm) of [math]\displaystyle{ G }[/math] satisfies

[math]\displaystyle{ ||G||^2 = \sum_{i=1}^{m} \sum_{j=1}^m |\langle x_i , x_j \rangle|^2 = \sum_{i=1}^m \lambda_i^2 }[/math]

Taking this together with the preceding inequality gives

[math]\displaystyle{ \sum_{i=1}^m \sum_{j=1}^m |\langle x_i , x_j \rangle|^2\geq \frac{(\mathrm{Tr}\;G)^2}{n} }[/math]

Because each [math]\displaystyle{ x_i }[/math] has unit length, the elements on the main diagonal of [math]\displaystyle{ G }[/math] are ones, and hence its trace is [math]\displaystyle{ \mathrm{Tr}\;G = m }[/math]. So,

[math]\displaystyle{ \sum_{i=1}^{m} \sum_{j=1}^m |\langle x_i , x_j \rangle|^2 = m+\sum_{i\neq j} |\langle x_i , x_j \rangle|^2 \geq \frac{m^2}{n} }[/math]

or

[math]\displaystyle{ \sum_{i\neq j} |\langle x_i , x_j \rangle|^2 \geq \frac{m(m-n)}{n} }[/math]

The second part of the proof uses an inequality encompassing the simple observation that the average of a set of non-negative numbers can be no greater than the largest number in the set. In mathematical notation, if [math]\displaystyle{ a_{\ell}\geq 0 }[/math] for [math]\displaystyle{ \ell=1,\ldots, L }[/math], then

[math]\displaystyle{ \frac{1}{L}\sum_{\ell=1}^L a_{\ell} \leq \max a_{\ell} }[/math]

The previous expression has [math]\displaystyle{ m(m-1) }[/math] non-negative terms in the sum, the largest of which is [math]\displaystyle{ c_\max^2 }[/math]. So,

[math]\displaystyle{ (c_\max)^2\geq \frac{1}{m(m-1)}\sum_{i\neq j} |\langle x_i , x_j \rangle|^2\geq\frac{m-n}{n(m-1)} }[/math]

or

[math]\displaystyle{ (c_\max)^2\geq \frac{m-n}{n(m-1)} }[/math]

which is precisely the inequality given by Welch in the case that [math]\displaystyle{ k=1 }[/math].

Achieving the Welch bounds

In certain telecommunications applications, it is desirable to construct sets of vectors that meet the Welch bounds with equality. Several techniques have been introduced to obtain so-called Welch Bound Equality (WBE) sets of vectors for the [math]\displaystyle{ k=1 }[/math] bound.

The proof given above shows that two separate mathematical inequalities are incorporated into the Welch bound when [math]\displaystyle{ k=1 }[/math]. The Cauchy–Schwarz inequality is met with equality when the two vectors involved are collinear. In the way it is used in the above proof, this occurs when all the non-zero eigenvalues of the Gram matrix [math]\displaystyle{ G }[/math] are equal, which happens precisely when the vectors [math]\displaystyle{ \{x_1,\ldots,x_m\} }[/math] constitute a tight frame for [math]\displaystyle{ \mathbb{C}^n }[/math].

The other inequality in the proof is satisfied with equality if and only if [math]\displaystyle{ |\langle x_i, x_j \rangle| }[/math] is the same for every choice of [math]\displaystyle{ i\neq j }[/math]. In this case, the vectors are equiangular. So this Welch bound is met with equality if and only if the set of vectors [math]\displaystyle{ \{x_i\} }[/math] is an equiangular tight frame in [math]\displaystyle{ \mathbb{C}^n }[/math].

Similarly, the Welch bounds stated in terms of average squared overlap, are saturated for all [math]\displaystyle{ k\le t }[/math] if and only if the set of vectors is a [math]\displaystyle{ t }[/math]-design in the complex projective space [math]\displaystyle{ \mathbb{CP}^{n-1} }[/math].[4]


See also

References

  1. Welch, L. (1974-05-01). "Lower bounds on the maximum cross correlation of signals (Corresp.)". IEEE Transactions on Information Theory 20 (3): 397–399. doi:10.1109/TIT.1974.1055219. ISSN 1557-9654. https://ieeexplore.ieee.org/document/1055219. 
  2. Klappenecker, Andreas; Roetteler, Martin (2005-02-11). Mutually Unbiased Bases are Complex Projective 2-Designs. 
  3. Belovs, Aleksandrs; Smotrovs, Juris (2008-07-22). A Criterion for Attaining the Welch Bounds with Applications for Mutually Unbiased Bases. 
  4. 4.0 4.1 Datta, Somantika; Howard, Stephen; Cochran, Douglas (2012-05-29). Geometry of the Welch Bounds.