Coding Theorems of Classical and Quantum Information Theory Second Edition Parthasarathy download
Coding Theorems of Classical and Quantum Information Theory Second Edition Parthasarathy download
https://ebookname.com/product/coding-theorems-of-classical-and-
quantum-information-theory-second-edition-parthasarathy/
https://ebookname.com/product/classical-and-quantum-information-
theory-an-introduction-for-the-telecom-scientist-1st-edition-
emmanuel-desurvire/
https://ebookname.com/product/fundamentals-of-information-theory-
and-coding-design-1st-edition-roberto-togneri/
https://ebookname.com/product/quantum-information-theory-and-the-
foundations-of-quantum-mechanics-2nd-edition-christopher-g-
timpson/
https://ebookname.com/product/principles-of-cell-
proliferation-1st-edition-j-r-heath/
Lead Acid Battery Technologies Fundamentals Materials
and Applications Electrochemical Energy Storage and
Conversion 1st Edition Joey Jung (Editor)
https://ebookname.com/product/lead-acid-battery-technologies-
fundamentals-materials-and-applications-electrochemical-energy-
storage-and-conversion-1st-edition-joey-jung-editor/
https://ebookname.com/product/the-private-lives-of-trees-1st-
edition-alejandro-zambra/
https://ebookname.com/product/energy-capitals-local-impact-
global-influence-1st-edition-joseph-a-pratt/
https://ebookname.com/product/weight-gain-during-pregnancy-
reexamining-the-guidelines-1-har-com-edition-committee-to-
reexamine-iom-pregnancy-weight-guidelines/
https://ebookname.com/product/warhammer-fantasy-roleplay-4th-
edition-dominic-mcdowall/
The Cinematic Whitechapel Documents of Contemporary Art
Campany
https://ebookname.com/product/the-cinematic-whitechapel-
documents-of-contemporary-art-campany/
~ TEXTS Ai'iD READINGS 45
~ IN j\IATHEMATICS
~HINDUSTAN
U UlJUBOOKAGENCY
TEXTS AND READINGS
IN MATHEMATICS
45
Co ding Theorems of
Classical and Quantum
Information Theory
Second Edition
Texts and Readings in Mathematics
Advisory Editor
C. S. Seshadri, Chennai Mathematical Institute, Chennai.
Managing Editor
Rajendra Bhatia, Indian Statistical Institute, New Delhi.
Editors
V. Balaji, Chennai Mathematical Institute, Chennai.
R. B. Bapat, Indian Statistical Institute, New Delhi.
V S. Borkar, Tata Inst. of Fundamental Research, Mumbai.
Probal Chaudhuri, Indian Statistical Institute, Kolkata.
Coding Theorems of
Classical and Quantum
Information Theory
Second Edition
K. R. Parthasarathy
Indian Statistical Institute
NewDelhi
[Ldgl@ HINDUSTAN
U U!J UBOOK AGENCY
Published by
email: [email protected]
www.hindbook.com
All export rights for this edition vest exclusively with Hindustan Book Agency
(India) . Unauthorized export is a violation of Copyright Law and is subject
to legal action.
Shyama
Preface
The logarithmic connection between entropy and probability was first enun-
ciated by L.E. Boltzmann (1844-1906) in his kinetic theory of gases. His
famous formula for entropy S is S = k log W (as engraved on his tombstone in
Vienna) where k is a constant and W is the number of possible microstates
corresponding to the macroscopic state of a system of particles in agas.
Ignoring the constant k and replacing log W by -log P(E) where P(E) is the
probability of an event E in the probability space (n, F, P) of a statistical
experiment, C. E. Shannon (1916- 2001) looked upon -logP(E) as a measure
of the information gained about the probability space from the occurrence of
E. If X is a simple random variable on this probability space assuming the
values al , a2, ... , ak from a finite set with P( X = aj) = Pj for each j then the
famous Shannon entropy H(X) = - Lj Pj log Pj is the expected information
about (n, F, P) gained from observing X. Cent red around this idea of
entropy a mathematical theory of communication was woven by Shannon in
a celebrated pair of papers in the 1948 volume of the Bell System Technical
Journal. Here Shannon established two fundamental coding theorems about
the optimal compressibility of a text in its storage and the optimal capacity of
a channel in communicating a text after encoding.
The essential feature of the revised edition is the inclusion of a new chapter
devoted to the Knill-Laflamme theory of quantum error correction and its
consequences in the construction of t-error correcting quantum codes. Our
approach is based on the unification of classical and quantum error correcting
codes through imprimitivity systems for finite group actions.
Many typographical error corrections and some minor changes have been
made in the text of the first edition.
Preface vii
Bibliography 111
Index 115
Chapter 1
Entropy of Elementary
Information Sources
U Ar
00
S(A) =
r=l
and for any WI = XIX2 ... Xnl E Anl, W2 = YIY2 ... Yn 2 E An 2 define the product
word WIW2 by WIW2 = XIX2.·· Xnl YIY2· .. Yn 2· Thus l(WIW2) = l(wI) + l(w2).
Clearly, this multiplication is associative.1t makes S(A) a semigroup without an;
identity element. We call S(A) the free semigroup or word semigroup generated
by the alphabet A.
Let A, B be alphabets of sizes a, b respectively. A one-to-one (or injective)
map f : A ~ S(B) is called a code with message alphabet A and encoding
alphabet B. When B is the two point set {O, I} such a code f is called a binary
code. Any word in the range of a code f is called a basic code word. Start with
a code f : A ~ S(B) and extend it uniquely to a map j : S(A) ~ S(B) by
putting
2 Chapter 1. Entropy of Elementary Information Sources
for any word w = XIX2 ... X n in S(A) . Then f is said to be a uniquely deci-
pherable code if its extension j is also one to one. The code f is said to be
irreducible if for any two letters x and y in A , f(y) #- f(x) and f(y) cannot
be expressed as f(y) = f(x)w for any word w in S(B). A simple examination
shows that an irreducible code is uniquely decipherable.
We shall now establish a necessary condition for a code f : A -+ S(B) to
be uniquely decipherable.
Proof. Let
L b-l(f(x» L L
L
b-l(f(x»
xEA r=l x :l(f(x»=r
r=l
Define
N(k) = # {j(w) Iw E S(A), l(j(w)) = k} ,
the cardinality of the set of all code words of length k. Clearly, N(k) :::; bk for
k = 1,2, .. . Thus the power series
L N(k)zk
00
F(z) = 1 +
k=l
where b = #B. Then there exists an irreducible (and hence uniquely decipher-
able) code! : A -t S(B) such that m(x) = l(J(x)) '<I xE A.
Proof. Define
L maxm(x),
xEA
Ar {x I m(x) = r},
Cr #A r , 1:'S r :'S L.
Then A = UrA r is a partition of A into disjoint sets and Cl + C2 + ... + CL = a,
the size of A. Then (1.1.2) can be expressed as
This implies
Definition 1.1.4 An elementary information source (EIS) is a pair (A, J.L) where
Ais an alphabet and J.L is a probability distribution on A, Le., J.L : A -+ [0,1] is
a map satisfying E J.L(x) = 1.
xEA
As an example we may consider A to be the union of the set of all 26 let-
ters a, b, . . . ,Z of the English language, the set of all punctuation marks like full
stop ., comma , , question mark ?, . . . etc and a symbol for the 'space' between
successive words. By making an analysis of the frequencies with which letters
from A appear in different pages of books one can construct a model distribu-
tion J.L so that the EIS (A, J.L) is an approximation of the English language. Such
a model can be constructed for any language after a suitable alphabetization
of its script.
Let now (A, J.L) be an EIS and let B be an encoding alphabet with #A = a,
#B = b. Consider any uniquely decipherable code f : A -+ S(B). Then l(J(x))
considered as a function of x E A is a positive integer-valued random variable
on the probability space (A, J.L) with expectation
which is called the mean code word length of the code f. Write
L(J.L) = min{I(J) I f : A -+ S(B) is a uniquely decipherable code} .
Theorem 1.1.5 Let (A, J-l) be an EIS and let b be the size of the eneoding
alphabet B. Then
- LJ-l(x)logv(x)
xEA
= - LJ-l(x)logJ-l(x) -log
xEA
rr
xEA
V(X)){l(X)
(~
J-l
Sinee a weighted geometrie mean does not exeeed the eorresponding weighted
arithmetic mean we have
rr
xEA
(V(X)){l(X) S L
J-l(x) xEA
v(x) J-l(x)
J-l(x)
= 1.
Thus
-L J-l(x) log v(x) 2': - L J-l(x) 10gJ-l(x).
xEA xEA
xEA x
6 Chapter 1. Entropy of Elementary Information Sources
Hence by Theorem 1.1.2 there exists an irreducible (and hence uniquely deci-
pherable) code f : A -+ 8(B) for which l(f(x)) = m(x) V x. Hence
Z(f) = L m(x)p,(x)
xEA
< ~
L...J
(1 _ log p,(X)) p,(x)
log b
xEA
- LXEA p,(x) logp,(x)
= . log b + 1,
proving the right hand part of (1.1.4). o
Remark 1.1.6 We can express the inequality (1.1.4) as
is also a very useful quantity in the development of our subject. In fact the
randorn variable - log2 p,(.), its expectation H (p,) and its standard deviation 0"Jl.
play an important role in the understanding of coding theorems of information
theory.
Exercise 1.1.7 For any alphabet A let P(A) denote the set of all probability
distributions on A. Then the following holds :
(i) 0 ~ H(p,) ~ log #A V p, E P(A) . Here equality obtains on the left hand
side if and only if p, is degenerate, i.e., p,(x) = 1 for so me xE A. Equality
obtains on the right hand side if and only if p, is the uniform distribution
on A, i.e., p,(x) = (#A)-l V xE A.
(ii) For any pre-assigned value H for H{p,) the maximum of 0"Jl. is attained at
a distribution of the form
p if x = Xo,
p,{x) = { 2=L.f .../..
#A-l 1 X I Xo
1.1. Uniquely decipherable and irreducible codes 7
where q = 1 - p. Furthermore,
max(Jp.
p.
(Hint: For both (i) and (ii) use the method of Lagrange multipliers.)
Exercise 1.1.8 Let A = {O, I} and let f.tp(O) = 1- p, f.tp(1) = p where 0 :s: p :s: 1.
Then (A, f.tp) is called a Bernoulli source. Its entropy H(f.tp) = -p log2 P -
(1 - p) log2(1 - p) is a continuous function of p in the unit interval [O,IJ.
Here we interpret x log2 x to be 0 when x = O. Furthermore H(f.tp) increases
monotonically from 0 to 1 as p increases from 0 to 1/2. Thus, for every 0 :s:
() :s: 1, there exists a unique p(()) in the interval [0, 1/2J such that H(f.tp((J») = ().
See figure below:
Figure 1.1
8 Chapter 1. Entropy of Elementary Information Sources
Exercise 1.1.9 If (A, J.l) and (B, v) are two elementary information sources define
their product (A x B, J.l ® v) by
A = {O, l} lh J+1
0lhJ
J.l=J.l I /2 ®J.lp({h})
where J.lp(O) and J.lI/2 are as in Exercise 1.1.8. Then H(J.l) = h. In other words,
for any 0 < h < 00 there exists a product of Bernoulli sources with Shannon
entropy h .
We now conclude this section with a heuristic discussion. The important
inequality (1.1.7) teIls us that the Shannon entropy H(J.l) measures the amount
of space (in terms of codeword length) needed for encoding or storing a 'mes-
sage' from the EIS (A, J.l). We may interpret H(J.l) as the expected information
from (A, J.l). Now consider a general probability space (n, F, J.l). Such a proba-
bility space describes a statistical experiment. In such an experiment watch the
occurrence of an event E E F. The occurrence of such an event throws light
on the prob ability space and provides 'information'. Suppose this information
is measured by a nonnegative quantity I(E) . If the event E is certain to occur,
i.e., J.l(E) = 1 no information is gained by watehing E. In other words I(E) = 0
whenever J.l(E) = 1. The rarer an event, its occurrence throws greater light on
the probability space. In other words I(E) is a monotonie decreasing function
of J.l(E). If EI and E 2 are two independent events it is desirable to have the
property I(E I n E 2 ) = I(E I ) + I(E2 ). These properties are fulfilled if we put
I(E) == -k log J.l(E) for some positive constant k. As a normalization we choose
I(E) = 1 when J.l(E) = 1/ 2. This suggests that I(E) = -log2 J.l(E) is a suitable
measure of the information provided by watehing the occurrence of the event
E.
When the sampie space n is partitioned into disjoint events Ej E F ,
j = 1, 2, .. . , n so that n = U']=I E j then (n, F, J.l) is approximated by (n, A, J.l)
where A is the algebra generated by the events E j , j = 1,2, . . . ,n. This
data gives rise to the elementary information source ({E I ,E2 , ... ,En },J.l)
with the probability distribution J.l(Ej ) , j = 1, 2, ... , n and Shannon entropy
n
- L J.l( E j ) log2 J.l( E j ) which is the average information .provided by the ap-
j=l
proximate experiment (n, A, J.l) .
1.2. The HuJJman code 9
Proposition 1.2.1 Let f : {I, 2, . . . ,a} --* S( {O, I})" be an irreducible code with
a
l(J(i» = Li, 1 S; i S; a such that L(J..L) =L liPi. Then lt S; h ::; . .. S; la-2 S;
i=1
la-I = La ·
Proof. Clearly, lt S; l2 S; .. . S; La . Indeed, if this were not so and li > li+1 for
some i we can interchange the code words f (i) and f (i + 1) and construct a new
irreducible code 9 for which g(i) = f(i + 1), g(i + 1) = f(i) and g(j) = f(j) for
all j different from i and i + 1. Then Z(g) < Z(J), a contradiction. Now suppose
that la-I< La. Then the word f(a) can be expressed as f(a) = WIW2 where
r)
l(WI) = la-I, l(w = la -la-I· Since f is irreducible WI is not an extension of
any of the words f(i) , i = 1,2, ... , a - 1. By dropping W2 we can construct a
new irreducible code 9 for which g(i) = f(i) Vi S; a - 1 and g(a) = WI . Then
Z(g) < Z(J), a contradiction. 0
Theorem 1.2.2 Consider the contracted EIS ({ 1,2, ... , a-l}, v) where v( i) = Pi,
1 S; i S; a - 2, v(a -1) = Pa-I + Pa where PI ~ P2 ~ .. . ~ Pa as in Proposition
1.2.1. Suppose 9 : {I, 2, ... , a - I} --* S( {O, I}) is an irreducible code for which
Z(g) = L(L') . Now define fo : {I, 2, ... ,a} --* S( {O, I}) by fo( i) = g( i) if 1 S; i S;
a - 2, fo(a - 1) = g(a - 1)0, fo(a) = g(a - 1)1. Then fo is an irreducible code
für which l(fo) = L(J..L) where J..L(i) = Pi , 1 ::; i ~ a .
Proof. Let f : {I, 2, ... , a} --* S( {O, I}) be an irreducible code for which Z(J) =
L(J..L) . By Proposition 1.2.1, lt S; h S; .. . S; la-I = la where li = l(J(i». By
Theorem 1.1.1 and the fact that irreducibility implies unique decipherability
we have
2- 11 + 2- 12 + ... + 2- la - 2 + 2- la - 1 +1 ::; 1.
By Theorem 1.1.2 there exists an irreducible code
h: {1,2, .. . ,a -I} --* S({O, I}) for which l(h(i» = Li if 1 S; i S; a - 2 and
l(h(a - 1» = la-I - 1. Thus
a-2
L liPi + (la-I - 1) (Pa-I + Pa) ~ L(v) = Z(g)
i=1
a-2
L l(g(i»Pi + l(g(a - 1»(Pa-1 + Pa).
i=1
(1.2.1)
10 Chapter 1. Entropy of Elementary Information Sources
a-2
IUo) = Ll(g(i))Pi + (l(g(a -1)) + I)(Pa-l + Pa)
i=l
a-2
= LliPi + la-l(Pa-l + Pa). (1.2.2)
i=l
a-2
L(Ji-) L l(g(i))pi + (l(g(a - 1)) + I)(Pa-l + Pa)
i=l
IUo).
°
reducible binary code (i.e., a code with encoding alphabet {O, I}) for an EIS
({1,2, ... ,a},Ji-) where Ji-(i) = Pi, Pl ~ P2 ~ ... ~ Pa > reduces to the
construction of an optimal irreducible binary code for the contracted or re-
duced EIS ({1,2, ... ,G - 1},v) where v(i) = Pi, i = 1,2, ... ,a - 2 and
v(a - 1) = Pa-l + Pa. However, it may be necessary to rearrange the points
of the reduced alphabet in order of decreasing probabilities. This reduces the
problem inductively and finally to an EIS with a two-point alphabet. We shall
illustrate by concrete examples.
°
Example 1.2.4 If A = {I, 2, 3}, Ji-(i) = Pi, Pl ~ P2 ~ P3 > consider the reduced
EIS ({I, 2}, v) with v(l) = Pl, v(2) = P2 + P3. Then we have two cases :
A 1 2 3 4 5 6 7
0.3 0.2 0.15 0.15 0.1 0.05 0.05
f-t 0 0
0
O~'
, __
A
o. ~ ___ >_ ~
0.2
-- > -
Figure 1.2
The code word corresponding to 6 i.e., 10(6) is to be read along the dotted line
6
from left to right. Note that I: 2 fo (i) = 1.
i=l
Exercise 1.2.7 When A = {O, 1,2, ... , 2n - I} and p, is the uniform distribution
then the binary expansion map f : A -+ {O, l}n is a Ruffman code. In this case
L(p,) = H(p,) = n.
Theorem 1.1.5, Remark 1.1.6 and the heuristic discussion at the end of Section
1.1 indicate that the Shannon entropy can be interpreted as a reasonable mea-
sure of the 'information content' of the EIS (A, J1). We shall now make a more
detailed investigation of the properties of H. Hereafter by the symbol 'log' we
shall understand that it is the logarithm with respect to the base 2. Thus the
naturallogarithm will be expressed as loge' Consider the function
cp
(t) = {
°
-t loge t if
if
0 < t :S 1,
t = 0.
(1.3.2)
in the interval [0,1]. Since cp'(t) = -(1 + loge t) and cp"(t) = -Cl it follows
that cp is a continuous and strictly concave function which increases from to°
°
e- l in [0, e- l ], attains its maximum value e- l at e- l and decreases from e- l
to in [e- l , 1]. See Figure 1.3 below :
Figure 1.3
1.3. Entropy and conditional entropy 13
Proposition 1.3.1 The map H : P(A) -+ [0,00) is continuous and for any Pi > 0,
i= 1,2, ... , k with LPi = 1 and any J-Li E P(A), i = 1,2, ... , k the following
holds:
(1.3.4)
where equality on the left obtains if and only if J-Ll = J-L2 = ... = J-Lk.
Furthermore, for any J-L E P(A), 0 :::; H(J-L) :::; log #A where (i) H(J-L) = 0
if and only if J-L is degenerate at a point; (ii) H(J-L) = log #A if and only if J-L is
the uniform distribution on A.
Proof. The continuity of H(J-L) in the variable J-L E P(A) is immediate from the
continuity of the function cP and (1.3.3). The strict concavity of cP implies that
multiplying both sides by Pj J-Lj (x) and adding over j and x we get the right
hand part of (1.3.4). Clearly, H(J-L) 2: 0 and H(J-L) = 0 if and only if cp(J-L(x)) = 0
\:j x E A, i.e., J-L(x) = 0 or 1 \:j x, i.e., J-L is degenerate at some point Xo in A. To
prove the last part of the proposition we shall repeat the argument involving
the weighted geometrie and arithmetic means in the proof of Theorem 1.1.5.
Without loss of generality we assurne that J-L(x) > 0 \:j xE A. Let v be another
probability distribution on A. Then
rr
xEA
[V(X)]JL(X) :::; LJ-L(X)v(x) = 1
J-L(x) xEA J-L(x)
where equality holds if and only if v(x) == J-L(x). Taking log we get
H(J-L):::; - LJ-L(x)logv(x).
xEA
14 Chapter 1. Entropy of Elementary Information Sources
°
definitions of P(Ei I A) on a set of probability we may assurne that the
following properties hold:
(i) P(Ei I A)(w) = P(Ei,w) is an A-measurable random variable with values
in [0,1] for each i = 1,2, ... , k.
k
(ii) I: P(Ei ,w) = 1 'VwEn.
i=I
Our website is not just a platform for buying books, but a bridge
connecting readers to the timeless values of culture and wisdom. With
an elegant, user-friendly interface and an intelligent search system,
we are committed to providing a quick and convenient shopping
experience. Additionally, our special promotions and home delivery
services ensure that you save time and fully enjoy the joy of reading.
ebookname.com