Lecture 10
Lecture 10
In this lecture, we will study how classical information can be encoded into quantum
states. Specifically, we will consider a d-dimensional quantum system and define a classical
channel by first preparing it in some quantum state associated to a classical message, and
then measuring the system with some POVM. We will then prove Holevo’s theorem giving
an upper bound for the classical capacity of any such channel. It will turn out that this
capacity is always less or equal to log(d), and we conclude that a d-dimensional quantum
system cannot store more than log(d) classical bits of information reliably.
1 Pinsker’s inequality
The following theorem will be proved in the exercises:
Theorem 1.1 (Pinsker’s inequality). For any quantum states ρ, σ ∈ D(H) we have
1
D (ρkσ) > kρ − σk21 .
2 ln(2)
Pinsker’s inequality can be seen as a refinement of Klein’s inequality (see exercises), and
it shows that the relative entropy is faithful, i.e., D(ρkσ) = 0 holds if and only if ρ = σ.
Proof. Exercise.
Consequences of the previous lemma are the following elementary properties of the quan-
tum mutual information: For any ρAB ∈ D(HA ⊗ HB ) we have
• I(A : B)ρAB > 0 with equality if and only if ρAB = ρA ⊗ ρB .
• I(A : B)(idA ⊗T )(ρAB ) 6 I(A : B)ρAB for any quantum channel T : B(HA ) → B(HC ).
We can reformulate the strong subadditivity inequality in the following way:
Theorem 2.3. For any quantum state ρABC ∈ D (HA ⊗ HB ⊗ HC ) we have
Proof. Exercises.
1
3 Convex structure of quantum measurements
Let H denote a complex Euclidean space and consider the set of all POVMs with at most
N
N ∈ outcomes, which we denote by
N
X
MN = {µ : {1, . . . , N } → B(H)+ : µ(n) = 1H }.
n=1
C
Since the set D( N ⊗ H) is compact and the restriction to the diagonal blocks is continuous,
we conclude that MN is compact as the image of a compact set under a continuous map.
Lemma 3.2. Let H denote a complex Euclidean space and µ : {1, . . . , N } → B(H)+ a
POVM. If µ ∈ Ext (MN ) is extremal, then we have
|{n ∈ {1, . . . , N } : µ(n) 6= 0}| 6 dim (H)2 .
Proof. Assume that
|{n ∈ {1, . . . , N } : µ(n) 6= 0}| > dim (H)2 .
Then, the operators {µ(1), . . . , µ(N )} are linearly dependent and there exist α1 , . . . , αN ∈ R,
not all of which are zero, such that
N
X
αn µ(n) = 0.
n=1
2
4 The setting of Holevo’s theorem
To explain the setting of Holevo’s theorem, we consider a set of quantum states
{ρx : x ∈ ΣA } ⊂ D(H),
labeled by some alphabet ΣA . For any alphabet ΣB and any POVM µ : ΣB → B(H)+ , we
may define a classical communication channel Nµ : ΣA → P (ΣB ) by
Note that we use the notation N (y|x) for a classical channel N : ΣA → P (ΣB ) to denote the
probability of y ∈ ΣB with respect to the probability distribution N (x) (as in Lecture 1).
Recall the mutual information of a joint probability distribution pAB ∈ P (ΣA × ΣB ) given
by
I (A : B)pAB = H(pA ) + H(pB ) − H(pAB ),
where pA and pB denote the marginals of pAB . In Lecture 1, we stated Shannon’s channel
coding theorem:
Theorem 4.1 (Shannon’s channel coding theorem). For alphabets ΣA and ΣB let N : ΣA →
P(ΣB ) denote a communication channel. The capacity of N is given by
where
pN
AB (x, y) = pA (x)N (y|x),
R < C(N ).
Definition 4.2 (Accessible information). The accessible information of an ensemble {p(x), ρx }x∈ΣA ,
where p ∈ P(ΣA ) and ρx ∈ D(H) for all x ∈ ΣA is given by
where the supremum goes over all POVMs µ : ΣB → B(H)+ and all alphabets ΣB .
The accessible information quantifies the highes possible mutual information between
the input variable x ∈ ΣA and the measurement outcome y ∈ ΣB for any choice of the
measurement. For any given set of quantum states {ρx : x ∈ ΣA } ⊂ D(H) we conclude
that
sup C(Nµ ) = sup Iacc ({p(x), ρx }) .
µ pA ∈P(ΣA )
This quantity equals the supremum of achievable rates for information transmission through
the preparation-measurement process for an optimal choice of measurement operators. In
the following, we will analyze the accessible information in more detail:
3
5 Accessible information is attained
We will start by showing that the supremum in the definition of the accessible information
is always attained by some measurement with a finite number of measurement outcomes. To
show this, we will start by showing that the optimization in the accessible information Iacc
is over a convex function.
Lemma 5.1. Let {p(x), ρx }x∈ΣA denote an ensemble for p ∈ P(ΣA ) and quantum states
ρx ∈ D(H). For some finite alphabet ΣB , any pair of POVMs µ0 , µ1 : ΣB → B(H)+ , and
any λ ∈ [0, 1] we have
where
µ = (1 − λ)µ0 + λµ1 .
Proof. Recall from the exercises that
and that the function N 7→ I(A : B)pN is convex (by joint convexity of the classical relative
AB
entropy). The statement of the lemma is an immediate consequence of this fact.
Now, we can show that the supremum in the definition of Iacc is always attained by an
extremal POVM.
Theorem 5.2. Let {p(x), ρx }x∈ΣA denote an ensemble for p ∈ P(ΣA ) and quantum states
ρx ∈ D(H). There exists an alphabet ΣB with |ΣB | 6 dim(H)2 and a POVM µ : ΣB →
B(H)+ such that
Iacc ({p(x), ρx }) = I(A : B)pNµ .
AB
Proof. Recall the set ML ⊂ B(H)L of POVMs with at most L outcomes. It will be conve-
nient to introduce the quantities
By renaming the elements y ∈ {1, . . . , L} for which µopt (y) 6= 0, we can identify µopt with a
POVM in Mdim(H)2 . This shows that
I(A : B) Nµopt 6 Iacc {p(x), ρx }, dim (H)2 .
pAB
4
Combining the previous statements shows that
Iacc ({p(x), ρx }) = Iacc {p(x), ρx }, dim (H)2 = I(A : B) Nµ ,
pABopt
for some probability distribution p ∈ P(ΣA ), quantum states σx ∈ D(HB ), and where
(cq)
{|xi}x∈ΣA denotes the computational basis. The physical interpretation of the state ρAB
is that system ‘A’ is in a classical state x with probability p(x). The classical state x is
represented by the pure state |xihx|A in the computational basis. Formally, the system ‘A’
is still a quantum system, but since the classical information is represented in a fixed basis
it can be accessed deterministically (by the PVM {|xihx|}x∈ΣA ).
To any ensemble {p(x), ρx }x∈ΣA with p ∈ P(ΣA ) and quantum states ρx ∈ D(H) we may
associate the classical-quantum state
(cq)
X
ρCA = p(x)|xihx|C ⊗ ρAx.
x∈ΣA
which is non-negative by concavity of the von Neumann entropy. Intuitively, one might
expect that this mutual information (which generalizes the classical mutual information)
could somehow quantify the information that the quantum system ‘A’ has about the classical
state x. This intuition is indeed correct, and as a result, the above quantity got its own name:
Definition 6.1 (Holevo information). For any ensemble {p(x), ρx }x∈ΣA with p ∈ P(ΣA )
and quantum states ρx ∈ D(H), we define
X X
χ ({p(x), ρx }) = H px ρx − px H(ρx ).
x∈ΣA x∈ΣA
5
such that
χ ({p(x), ρx }) = I(C : A)σCA = D(σCA kσC ⊗ σA ).
Next, consider a POVM µ : {1, . . . , L} → B(H)+ and define a quantum channel Mµ :
C
B(H) → B( L ) by
L
X
Mµ (X) = hµ(y), XiHS |yihy|,
y=1
Corollary 6.3. For any ensemble {p(x), ρx }x∈ΣA with p ∈ P(ΣA ) and quantum states ρx ∈
D(H), we have
Iacc ({p(x), ρx }) 6 log(dim(H)).
Although any quantum system can be in a continuum of different quantum states, the
previous corollary shows that classical messages can only be send reliably at rates below
log (dim(H)) by the process of preparation and measurement. This can be interpreted as
saying that a qubit can only store 1 bit of information reliably.