Physics:Linear entropy

From HandWiki

In quantum mechanics, and especially quantum information theory, the linear entropy or impurity of a state is a scalar defined as

[math]\displaystyle{ S_L \, \dot= \, 1 - \mbox{Tr}(\rho^2) \, }[/math]

where ρ is the density matrix of the state.

The linear entropy can range between zero, corresponding to a completely pure state, and (1 − 1/d), corresponding to a completely mixed state. (Here, d is the dimension of the density matrix.)

The linear entropy is trivially related to the purity [math]\displaystyle{ \gamma \, }[/math] of a state by

[math]\displaystyle{ S_L \, = \, 1 - \gamma \, . }[/math]

Motivation

The linear entropy is a lower approximation to the (quantum) von Neumann entropy S, which is defined as

[math]\displaystyle{ S \, \dot= \, -\mbox{Tr}(\rho \ln \rho) = -\langle \ln \rho \rangle \, . }[/math]

The linear entropy then is obtained by expanding ln ρ = ln (1−(1−ρ)), around a pure state, ρ2=ρ; that is, expanding in terms of the non-negative matrix 1−ρ in the formal Mercator series for the logarithm,

[math]\displaystyle{ - \langle \ln \rho \rangle = \langle 1- \rho \rangle + \langle (1- \rho )^2 \rangle/2 + \langle (1- \rho)^3 \rangle /3 + ... ~, }[/math]

and retaining just the leading term.

The linear entropy and von Neumann entropy are similar measures of the degree of mixing of a state, although the linear entropy is easier to calculate, as it does not require diagonalization of the density matrix.

Alternate definition

Some authors[1] define linear entropy with a different normalization

[math]\displaystyle{ S_L \, \dot= \, \tfrac{d}{d-1} (1 - \mbox{Tr}(\rho^2) ) \, , }[/math]

which ensures that the quantity ranges from zero to unity.

References

  1. Nicholas A. Peters; Tzu-Chieh Wei; Paul G. Kwiat (2004). "Mixed state sensitivity of several quantum information benchmarks". Physical Review A 70 (5): 052309. doi:10.1103/PhysRevA.70.052309. Bibcode2004PhRvA..70e2309P.