Classical capacity

From testwiki
Jump to navigation Jump to search

Template:Short description In quantum information theory, the classical capacity of a quantum channel is the maximum rate at which classical data can be sent over it error-free in the limit of many uses of the channel. Holevo, Schumacher, and Westmoreland proved the following least upper bound on the classical capacity of any quantum channel 𝒩:

χ(𝒩)=maxρXAI(X;B)𝒩(ρ)

where ρXA is a classical-quantum state of the following form:

ρXA=xpX(x)|xx|XρxA,

pX(x) is a probability distribution, and each ρxA is a density operator that can be input to the channel 𝒩.

Achievability using sequential decoding

We briefly review the HSW coding theorem (the statement of the achievability of the Holevo information rate I(X;B) for communicating classical data over a quantum channel). We first review the minimal amount of quantum mechanics needed for the theorem. We then cover quantum typicality, and finally we prove the theorem using a recent sequential decoding technique.

Review of quantum mechanics

In order to prove the HSW coding theorem, we really just need a few basic things from quantum mechanics. First, a quantum state is a unit trace, positive operator known as a density operator. Usually, we denote it by ρ, σ, ω, etc. The simplest model for a quantum channel is known as a classical-quantum channel:

xρx.

The meaning of the above notation is that inputting the classical letter x at the transmitting end leads to a quantum state ρx at the receiving end. It is the task of the receiver to perform a measurement to determine the input of the sender. If it is true that the states ρx are perfectly distinguishable from one another (i.e., if they have orthogonal supports such that Tr{ρxρx}=0 for xx), then the channel is a noiseless channel. We are interested in situations for which this is not the case. If it is true that the states ρx all commute with one another, then this is effectively identical to the situation for a classical channel, so we are also not interested in these situations. So, the situation in which we are interested is that in which the states ρx have overlapping support and are non-commutative.

The most general way to describe a quantum measurement is with a positive operator-valued measure (POVM). We usually denote the elements of a POVM as {Λm}m. These operators should satisfy positivity and completeness in order to form a valid POVM:

Λm0    m
mΛm=I.

The probabilistic interpretation of quantum mechanics states that if someone measures a quantum state ρ using a measurement device corresponding to the POVM {Λm}, then the probability p(m) for obtaining outcome m is equal to

p(m)=Tr{Λmρ},

and the post-measurement state is

ρm=1p(m)ΛmρΛm,

if the person measuring obtains outcome m. These rules are sufficient for us to consider classical communication schemes over cq channels.

Quantum typicality

The reader can find a good review of this topic in the article about the typical subspace.

Gentle operator lemma

The following lemma is important for our proofs. It demonstrates that a measurement that succeeds with high probability on average does not disturb the state too much on average:

Lemma: [Winter] Given an ensemble {pX(x),ρx} with expected density operator ρxpX(x)ρx, suppose that an operator Λ such that IΛ0 succeeds with high probability on the state ρ:

Tr{Λρ}1ϵ.

Then the subnormalized state ΛρxΛ is close in expected trace distance to the original state ρx:

𝔼X{ΛρXΛρX1}2ϵ.

(Note that A1 is the nuclear norm of the operator A so that A1Tr{AA}.)

The following inequality is useful for us as well. It holds for any operators ρ, σ, Λ such that 0ρ,σ,ΛI:

The quantum information-theoretic interpretation of the above inequality is that the probability of obtaining outcome Λ from a quantum measurement acting on the state ρ is upper bounded by the probability of obtaining outcome Λ on the state σ summed with the distinguishability of the two states ρ and σ.

Non-commutative union bound

Lemma: [Sen's bound] The following bound holds for a subnormalized state σ such that 0σ and Tr{σ}1 with Π1, ... , ΠN being projectors: Tr{σ}Tr{ΠNΠ1 σ Π1ΠN}2i=1NTr{(IΠi)σ},

We can think of Sen's bound as a "non-commutative union bound" because it is analogous to the following union bound from probability theory:

Pr{(A1AN)c}=Pr{A1cANc}i=1NPr{Aic},

where A1,,AN are events. The analogous bound for projector logic would be

Tr{(IΠ1ΠNΠ1)ρ}i=1NTr{(IΠi)ρ},

if we think of Π1ΠN as a projector onto the intersection of subspaces. Though, the above bound only holds if the projectors Π1, ..., ΠN are commuting (choosing Π1=|++|, Π2=|00|, and ρ=|00| gives a counterexample). If the projectors are non-commuting, then Sen's bound is the next best thing and suffices for our purposes here.

HSW theorem with the non-commutative union bound

We now prove the HSW theorem with Sen's non-commutative union bound. We divide up the proof into a few parts: codebook generation, POVM construction, and error analysis.

Codebook Generation. We first describe how Alice and Bob agree on a random choice of code. They have the channel xρx and a distribution pX(x). They choose M classical sequences xn according to the IID\ distribution pXn(xn). After selecting them, they label them with indices as {xn(m)}m[M]. This leads to the following quantum codewords:

ρxn(m)=ρx1(m)ρxn(m).

The quantum codebook is then {ρxn(m)}. The average state of the codebook is then

where ρ=xpX(x)ρx.

POVM Construction . Sens' bound from the above lemma suggests a method for Bob to decode a state that Alice transmits. Bob should first ask "Is the received state in the average typical subspace?" He can do this operationally by performing a typical subspace measurement corresponding to {Πρ,δn,IΠρ,δn}. Next, he asks in sequential order, "Is the received codeword in the mth conditionally typical subspace?" This is in some sense equivalent to the question, "Is the received codeword the mth transmitted codeword?" He can ask these questions operationally by performing the measurements corresponding to the conditionally typical projectors {Πρxn(m),δ,IΠρxn(m),δ}.

Why should this sequential decoding scheme work well? The reason is that the transmitted codeword lies in the typical subspace on average:

𝔼Xn{Tr{Πρ,δ ρXn}}=Tr{Πρ,δ 𝔼Xn{ρXn}}
=Tr{Πρ,δ ρn}
1ϵ,

where the inequality follows from (\ref{eq:1st-typ-prop}). Also, the projectors Πρxn(m),δ are "good detectors" for the states ρxn(m) (on average) because the following condition holds from conditional quantum typicality:

𝔼Xn{Tr{ΠρXn,δ ρXn}}1ϵ.

Error Analysis. The probability of detecting the mth codeword correctly under our sequential decoding scheme is equal to

Tr{ΠρXn(m),δΠ^ρXn(m1),δΠ^ρXn(1),δ Πρ,δn ρxn(m) Πρ,δn Π^ρXn(1),δΠ^ρXn(m1),δΠρXn(m),δ},

where we make the abbreviation Π^IΠ. (Observe that we project into the average typical subspace just once.) Thus, the probability of an incorrect detection for the mth codeword is given by

1Tr{ΠρXn(m),δΠ^ρXn(m1),δΠ^ρXn(1),δ Πρ,δn ρxn(m) Πρ,δn Π^ρXn(1),δΠ^ρXn(m1),δΠρXn(m),δ},

and the average error probability of this scheme is equal to

11MmTr{ΠρXn(m),δΠ^ρXn(m1),δΠ^ρXn(1),δ Πρ,δn ρxn(m) Πρ,δn Π^ρXn(1),δΠ^ρXn(m1),δΠρXn(m),δ}.

Instead of analyzing the average error probability, we analyze the expectation of the average error probability, where the expectation is with respect to the random choice of code:

Our first step is to apply Sen's bound to the above quantity. But before doing so, we should rewrite the above expression just slightly, by observing that

1=𝔼Xn{1MmTr{ρXn(m)}}
=𝔼Xn{1MmTr{Πρ,δnρXn(m)}+Tr{Π^ρ,δnρXn(m)}}
=𝔼Xn{1MmTr{Πρ,δnρXn(m)Πρ,δn}}+1MmTr{Π^ρ,δn𝔼Xn{ρXn(m)}}
=𝔼Xn{1MmTr{Πρ,δnρXn(m)Πρ,δn}}+Tr{Π^ρ,δnρn}
𝔼Xn{1MmTr{Πρ,δnρXn(m)Πρ,δn}}+ϵ

Substituting into (Template:EquationNote) (and forgetting about the small ϵ term for now) gives an upper bound of

𝔼Xn{1MmTr{Πρ,δnρXn(m)Πρ,δn}}
𝔼Xn{1MmTr{ΠρXn(m),δΠ^ρXn(m1),δΠ^ρXn(1),δ Πρ,δn ρXn(m) Πρ,δn Π^ρXn(1),δΠ^ρXn(m1),δΠρXn(m),δ}}.

We then apply Sen's bound to this expression with σ=Πρ,δnρXn(m)Πρ,δn and the sequential projectors as ΠρXn(m),δ, Π^ρXn(m1),δ, ..., Π^ρXn(1),δ. This gives the upper bound 𝔼Xn{1Mm2[Tr{(IΠρXn(m),δ)Πρ,δnρXn(m)Πρ,δn}+i=1m1Tr{ΠρXn(i),δΠρ,δnρXn(m)Πρ,δn}]1/2}. Due to concavity of the square root, we can bound this expression from above by

2[𝔼Xn{1MmTr{(IΠρXn(m),δ)Πρ,δnρXn(m)Πρ,δn}+i=1m1Tr{ΠρXn(i),δΠρ,δnρXn(m)Πρ,δn}}]1/2
2[𝔼Xn{1MmTr{(IΠρXn(m),δ)Πρ,δnρXn(m)Πρ,δn}+imTr{ΠρXn(i),δΠρ,δnρXn(m)Πρ,δn}}]1/2,

where the second bound follows by summing over all of the codewords not equal to the mth codeword (this sum can only be larger).

We now focus exclusively on showing that the term inside the square root can be made small. Consider the first term:

𝔼Xn{1MmTr{(IΠρXn(m),δ)Πρ,δnρXn(m)Πρ,δn}}
𝔼Xn{1MmTr{(IΠρXn(m),δ)ρXn(m)}+ρXn(m)Πρ,δnρXn(m)Πρ,δn1}
ϵ+2ϵ.

where the first inequality follows from (Template:EquationNote) and the second inequality follows from the gentle operator lemma and the properties of unconditional and conditional typicality. Consider now the second term and the following chain of inequalities:

im𝔼Xn{Tr{ΠρXn(i),δ Πρ,δn ρXn(m) Πρ,δn}}
=imTr{𝔼Xn{ΠρXn(i),δ} Πρ,δn 𝔼Xn{ρXn(m)} Πρ,δn}
=imTr{𝔼Xn{ΠρXn(i),δ} Πρ,δn ρn Πρ,δn}
im2n[H(B)δ] Tr{𝔼Xn{ΠρXn(i),δ} Πρ,δn}

The first equality follows because the codewords Xn(m) and Xn(i) are independent since they are different. The second equality follows from (Template:EquationNote). The first inequality follows from (\ref{eq:3rd-typ-prop}). Continuing, we have

im2n[H(B)δ] 𝔼Xn{Tr{ΠρXn(i),δ}}
im2n[H(B)δ] 2n[H(B|X)+δ]
=im2n[I(X;B)2δ]
M 2n[I(X;B)2δ].

The first inequality follows from Πρ,δnI and exchanging the trace with the expectation. The second inequality follows from (\ref{eq:2nd-cond-typ}). The next two are straightforward.

Putting everything together, we get our final bound on the expectation of the average error probability:

1𝔼Xn{1MmTr{ΠρXn(m),δΠ^ρXn(m1),δΠ^ρXn(1),δ Πρ,δn ρXn(m) Πρ,δn Π^ρXn(1),δΠ^ρXn(m1),δΠρXn(m),δ}}
ϵ+2[(ϵ+2ϵ)+M 2n[I(X;B)2δ]]1/2.

Thus, as long as we choose M=2n[I(X;B)3δ], there exists a code with vanishing error probability.

See also

References

Template:Quantum computing