Lecture Notes On The Theory of Open Quantum Systems
Lecture Notes On The Theory of Open Quantum Systems
Daniel A. Lidar
Departments of Electrical Engineering, Chemistry, and Physics & Astronomy
Center for Quantum Information Science & Technology
University of Southern California, Los Angeles, California 90089, USA
This is a self-contained set of lecture notes covering various aspects of the theory of open quantum system, at
a level appropriate for a one-semester graduate course. The main emphasis is on completely positive maps and
master equations, both Markovian and non-Markovian.
CONTENTS
arXiv:1902.00967v2 [quant-ph] 21 Feb 2020
IX. Derivation of the Lindblad equation from a short time expansion of quantum maps 45
A. Derivation 45
B. The Markovian evolution operator as a one-parameter semigroup 47
C. Proof that the solution of the Lindblad Equation is a CP map 47
D. Examples 48
1. Just H for a single qubit: the Bloch equations 48
2. Phase Damping for a single qubit 49
3. Amplitude damping / Spontaneous Emission for a single qubit 49
XIV. Derivation of the Lindblad equation from the cumulant expansion and coarse graining 70
A. Cumulant expansion 70
B. The second order cumulant 72
C. Why the first order cumulant can be made to vanish 72
D. Derivation of the Lindblad equation 73
E. Complete positivity 74
F. LE from the cumulant expansion and coarse-graining 74
G. Illustration using the spin-boson model for phase damping 75
XV. First-principles derivation of the Lindblad equation from the Born, Markov, and rotating wave approximations 76
A. Setting up 76
B. Born approximation 77
C. Markov approximation and Redfield equation 78
D. Going to the frequency domain 79
E. Rotating Wave Approximation 79
F. The missing steps 81
1. From Born-Markov [Eq. (524)] to the RWA-LE [Eq. (536)] 81
2. Proof of Eq. (539) 81
3. Transformation back to the Schrödinger picture 82
4. Proof that γ(ω) > 0 83
XVI. The Kubo-Martin-Schwinger (KMS) condition and the Gibbs state as a stationary state Lindblad equation 85
A. The KMS condition 85
B. The Gibbs state is a stationary state of the RWA-LE 86
C. Return to equilibrium, quantum detailed balance, and ergodicity under the RWA-LE 88
References 131
5
The theory of open quantum systems is the backbone of nearly all modern research in quantum mechanics and its applications.
The reason is simple: the idealization of an isolated quantum system obeying perfectly unitary quantum dynamics is just that:
an idealization. In reality every system is open, meaning that it is coupled to an external environment. Sometimes these open
system effects are small, but they can almost never be neglected. This is particularly relevant in the field of quantum information
processing, where the existence of a quantum advantage over classical information processing is often derived first from the
idealized, closed system perspective, and must then be re-scrutinized in the realistic, open system setting.
These lecture notes provide a fairly comprehensive and self-contained introduction to the theory of open quantum systems.
They are based on lectures I gave at the University of Southern California as part of a one-semester graduate course on the
topic taught in Fall 2006, Spring 2013, Spring 2017, and Fall 2018. There are several excellent textbooks and monographs
either devoted to or containing the same subject, and these notes are in parts heavily influenced by these works, in particular
the invaluable books by Heinz-Peter Breuer and Francesco Petruccione [1] and by Robert Alicki and Karl Lendi [2]. The
notes do fill in many details not found in the original sources (at times tediously so!), and also draw on various articles and
unpublished materials. I therefore hope that these notes will serve as a useful companion to the textbooks, and will help students
and researchers interested in entering the field in a semester of dedicated study.
The notes were originally typeset by students serving as scribes during the lectures given in 2013 and 2017, and have under-
gone extensive editing and additions since then. I am extremely grateful to all the students who participated in this effort: Chao
Cao, Rajit Chatterjea, Yi-Hsiang Chen, Jan Florjanczyk, Jose Raul Gonzalez Alonso, Anastasia Gunina, Drew Henry, Kung-
Chuan Hsu, Zhihao Jiang, Joshua Job, Hannes Leipold, Milad Marvian, Anurag Mishra, Nicolas Moure Gomez, Siddharth
Muthu Krishnan, Shayne Sorenson, Georgios Styliaris, Christopher Sutherland, Subhasish Sutradhar, Walter Unglaub, Ka Wa
Yip, and Yicong Zheng. I am also very grateful to the students in the 2018 course, who offered numerous additional feedback:
Namit Anand, Mojgan Asadi, Brian Barch, Matthew Kowalsky, Lawrence Liu, Humberto Munoz Bauza, Adam Pearson, Bibek
Pokharel, Evangelos Vlachos, Aaron Wirthwein, Haimeng Zhang, and Zihan Zhao. Finally, I wish to warmly thank Dr. Tameem
Albash and Dr. Jenia Mozgunov, who filled in for me on various occasions, and whose notes I relied on as well.
Of course, all errors, typos, and omissions are mine. The reader is strongly encouraged to send me any corrections at li-
[email protected]. The notes will be updated regularly to reflect these corrections, as well as new material of interest. I apologize
in advance to all the numerous authors whose contributions I did not cite; the field is vast and the intent of these notes is not to
serve as a comprehensive review article. I have certainly not done justice to the literature.
The completion of this work was (partially) supported by the Office of the Director of National Intelligence (ODNI), Intelli-
gence Advanced Research Projects Activity (IARPA), via the U.S. Army Research Office contract W911NF-17-C-0050.
6
The introductory material presented here is based on the approach of the excellent textbook by Nielsen & Chuang [3]. There
are four main postulates on which Quantum Mechanics can be built. These four postulates are:
1. Where things happen: Hilbert space
2. Combining Quantum Systems: Tensor product of vectors, matrices and of Hilbert spaces
3. Time Evolution (Dynamics): Schrödinger equation
4. Information extraction: Measurements
A. Postulate 1
“To every Quantum system is associated a state space, i.e, a Hilbert space H.”
A Hilbert space is a vector space equipped with an inner product. The vector spaces that we will be working with, H, can be
defined in the following way, in which C is the field of complex numbers:
H = Cd (1a)
⎛ v0 ⎞
⎜ v ⎟
v=⎜ 1
= {⃗ ⎟ ∣vi C} (1b)
⎜ ⋮ ⎟
⎝ vd−1 ⎠
Thus for our purposes a vector space can be defined as the set of d-dimensional vectors v⃗, each element of which, vi , is a
complex number. Recall that a vector space has a couple of properties. First, for all vectors v⃗ ∈ H, a⃗ v + bv⃗′ ∈ H, with a, b ∈ C,
i.e., any linear combination of vectors z⃗ is also an element of the vector space V . Second, the vector space must contain the zero
vector, an element that satisfies the condition 0⃗ + v⃗ = v⃗ ∀⃗
v ∈ C.
The postulate means that physical states of a quantum system can be associated to a vector v⃗ ∈ H. We shall use Dirac notation,
in which column vectors are denoted by “kets”: v⃗ ↦ ∣v⟩. In what follows, we will usually assume that the dimension of H is
d−1
finite, and find an orthonormal basis for it. That is, if dim(H) = d, then denote a such a basis {∣k⟩}k=0 . A good (but obviously
non-unique) choice is the standard basis
⎛0⎞
⎜⋮⎟
⎜ ⎟
∣k⟩ = ⎜1⎟ ← k + 1th position (2)
⎜ ⎟
⎜⋮⎟
⎝0⎠
Any vector in the Hilbert space can be expanded in an orthonormal basis as a linear combination
d−1
∣v⟩ = ∑ vk ∣k⟩ , vk ∈ C, (3)
k=0
which quantum physicists often call a superposition. The coefficients vk are called probability amplitudes. The reason is that
2
the probability of a quantum system “being” in a specific state ∣k⟩ is ∣vk ∣ . This latter statement is part of the postulate. The
different orthonormal basis vector ∣k⟩ represent mutually exclusive possibilities, such as the discrete positions of a particle on a
line, or different spin configurations.
2
Of course, in order for the set {∣vk ∣ } to be a proper probability distribution, the probabilities must sum to one. This is the
reason that we need to endow the vector space H with an inner product, i.e., work with Hilbert spaces. To define the inner
product function we first introduce the dual of a ket, called a “bra”. In Dirac notation, row vectors (or bras) are written as ⟨v∣,
†
where by definition ⟨v∣ = ∣v⟩ , where the dagger denotes Hermitian conjugation, i.e., transpose and complex conjugation. Thus
if ∣v⟩ is written as in Eq. (1b) then ⟨v∣ = {v1∗ , v2∗ , . . . , vn∗ }. One reason that Dirac notation is convenient because we can represent
the inner product as a “braket”, i.e.,
⎛ w0 ⎞ d−1
⟨v∣w⟩ ≡ (v0∗ , ⋯, vd−1
∗
) ⎜ ⋮ ⎟ = ∑ vk∗ wk . (4)
⎝wd−1 ⎠ k=0
7
Additional linear algebra and Dirac notation facts are collected in Appendix A.
B. Postulate 2
“Given two quantum systems with respective Hilbert spaces H1 and H2 the combined quantum system has
associated with it a Hilbert space given by H = H1 ⊗ H2 .”
1 −1 2 −1
Let us define H1 to be the span of {∣vi ⟩}di=0 , and similarly H2 to be the span of {∣wj ⟩}dj=0 . Then we have H defined as the
1 −1,d2 −1
span of {∣vi ⟩ ⊗ ∣wj ⟩}di=0,j=0 . For two states ∣ψ⟩ ∈ H1 , ∣ϕ⟩ ∈ H2 , the tensor product is given by
⎛ ψ0 ϕ0 ⎞
⎛ ψ0 ⎞ ⎛ ϕ0 ⎞ ⎜ ⋮ ⎟
⎜ ⎟
∣ψ⟩ ⊗ ∣ϕ⟩ = ⎜ ⋮ ⎟ ⊗ ⎜ ⋮ ⎟ = ⎜ ψ0 ϕd2 −1 ⎟ (7)
⎜
⎝ψd1 −1 ⎠ ⎝ϕd2 −1 ⎠ ⎜ ⎟
⋮ ⎟
⎝ψd1 −1 ϕd2 −1 ⎠
Note that the underlying Hilbert spaces could represent entirely different physical systems, e.g., the first could be the space of
electron spins, whereas the second could be the space of photon polarizations. We can also define the tensor product between
matrices, i.e., if
then
C. Postulate 3
“∃ a unitary operator U (t) such that the time evolution of a state is given by
Equivalently, the state vector of the system satisfies the Schrödinger equation
i
∣ψ̇(t)⟩ = − ̵ H ∣ψ(t)⟩ (11)
h
with H being a Hermitian operator known as the Hamiltonian.”
̵ = 1, which means that the units of energy and frequency will be the same.
The dot denotes ∂/∂t, and in this course we will set h
We shall show below that the equivalence holds provided U (t) = exp (−iHt) when H is t-independent. In the time dependent
case, the situation is more complicated, and we have
U (t) = T+ e−i ∫0
T
H(t′ )dt′
(12)
I = A† A = I = (∑ λ∗a ∣a⟩ ⟨a∣) (∑ λa′ ∣a′ ⟩ ⟨a′ ∣) = ∑ λ∗a λa ∣a⟩ ⟨a∣ = ∑ ∣λa ∣ ∣a⟩ ⟨a∣ = ∑ ∣a⟩ ⟨a∣ ,
2
(13)
a a′ a a a
where the last equality is the spectral representation of the identity operator I (all its eigenvalues are 1). Thus the eigen-
values of a unitary operator are all phases: λa = eiθa where θa ∈ R.
We now define functions of normal operators. If we have a function f ∶ C → C, then we can extend it to the case of normal
operators by defining
d At d A2 t2 A3 t3 A2 A3 2 A2 t2
(e ) = (I + At + + + ...) = A + (2t) + (3t ) + ... = A(I + At + + ...) = AeAt (16)
dt dt 2! 3! 2! 3! 2!
Now for the other direction, we start with writing the spectral decomposition of the Hamiltonian as H = ∑a λa ∣a⟩ ⟨a∣, and note
also that from the definition (14) we have:
Now, since the eigenvectors of H are a basis (again, from the spectral theorem), we can decompose ∣ψ(t)⟩ in this basis and write
∣ψ(t)⟩ = ∑a ψa (t) ∣a⟩, so that the left hand side of Eq. (11) becomes
d d
∣ψ(t)⟩ = ∑ ψa (t) ∣a⟩. (18)
dt a dt
9
For Eq. (11) to hold, these two need to be equal term by term (from orthonormality of the basis), so that we find
d
ψa (t) = −iλa ψa (t) ⇒ ψa (t) = e−iλa t ψa (0) (20)
dt
Plugging this result into ∣ψ(t)⟩ = ∑a ψa (t) ∣a⟩ and using orthonormality once more we now have:
∣ψ(t)⟩ = ∑ e−iλa t ψa (0) ∣a⟩ = (∑ e−iλa t ∣a⟩ ⟨a∣)(∑ ψa′ (0) ∣a′ ⟩) = e−iHt ∣ψ(0)⟩ = U (t) ∣ψ(0)⟩. (21)
a a a′
D. Postulate 4
This is the most controversial postulate, but we will not discuss those issues here and simply assume its validity.
This postulate has two parts: measuring states and measuring operators.
N
1. Measuring States: Quantum measurements are described by a set {Mk }k=1 of measurement operators satisfying the
constraint ∑k Mk† Mk = I.
Given a state ∣ψ⟩ ∈ H, instantaneously after the measurement it becomes,
Mk ∣ψ⟩
∣ψ⟩ ↦ √ ≡ ∣ψk ⟩ , (22)
pk
with probability
pk = ⟨ψ∣Mk† Mk ∣ψ⟩ = ∥Mk ∣ψ⟩ ∥2 ≥ 0. (23)
The measurement outcome is the index k of the state that resulted. The constraint listed in the postulate has the following
origin. Notice that ∑k pk = 1 must be true since pk is a probability, which implies ⟨ψ∣ ∑k Mk† Mk ∣ψ⟩ = 1. Since this is
true for arbitrary ∣ψ⟩ the sum rule follows
∑ Mk Mk = I.
†
(24)
k
2. To every physically measurable quantity is associated an observable, i.e., a Hermitian operator A. A has a spectral
decomposition (since it is Hermitian and hence normal),
A = ∑ λa ∣a⟩ ⟨a∣ , (25)
a
with λa ∈ R since A is Hermitian. The λa ’s, the eigenvalues, are the outcomes of the measurement (hence need to be
real).1
The set of eigenvectors, {∣a⟩} are an orthonormal set. Hence, {Pa ≡ ∣a⟩ ⟨a∣} are projectors (defined below). These are the
measurement operators corresponding to the measurement of this observable. Hence, if the system is in state ∣ψ⟩ before the
observable A is measured, according to Eq. (23) the probability of outcome λa is given by pa = ⟨ψ∣Pa† Pa ∣ψ⟩ = ⟨ψ∣Pa2 ∣ψ⟩ =
⟨ψ∣Pa ∣ψ⟩ = ∣ ⟨ψ∣ a⟩∣2 . Moreover, according to Eq. (22) the state after the measurement is performed and outcome λa is
observed, becomes ∣ψa ⟩ = P√ap∣ψ⟩
a
⟨a∣ψ⟩
= ∣⟨a∣ψ⟩∣ ∣a⟩ = eiθ ∣a⟩, where eiθ is the phase associated with the complex number ⟨a∣ ψ⟩.
We next consider several important special cases of the generalized measurements defined above.
1 It is interesting to ask why physically measurable quantities should be associated with Hermitian operators. Intuitively, since physical measurements pro-
duce real numbers we want to associate an operator with a real spectrum as a physically observable quantity. Moreover, we would like states with dif-
ferent eigenvalues (or different results from the measurement) to be orthogonal. A Hermitian operator satisfies both of these requirements. However,
these justifications admittedly leave something to be desired. For more details see https://physics.stackexchange.com/questions/39602/
why-do-we-use-hermitian-operators-in-qm.
10
Projective measurements are a special case of generalized measurements, in which the measurement operators, Mk are Her-
mitian operators called projectors. That is, Mk = Pk , where Pk Pl = δk,l Pk and Pk† = Pk . In particular, Pk2 = Pk . Using this, we
can see that the probability of outcome k, pk = ⟨ψ∣Mk† Mk ∣ψ⟩ = ⟨ψ∣Pk ∣ψ⟩.
1 0
Example: Let ∣ψ⟩ = a ∣0⟩ + b ∣1⟩ where ∣0⟩ = ( ) , ∣1⟩ = ( ) and a, b ∈ C. That is, ∣ψ⟩ ∈ C2 and {∣0⟩ , ∣1⟩} is the standard basis
0 1
for the space. Such a ∣ψ⟩ is called a qubit.
Now, we define measurement operators, M0 = P0 = ∣0⟩⟨0∣ and M1 = P1 = ∣1⟩⟨1∣. We can see that P0,1 2
= P0,1 and P0 P1 = 0.
Hence, this is a set of projective measurements. Thus, the probabilities of outcomes are,
2
p0 = ⟨ψ∣P0 ∣ψ⟩ = ⟨ψ∣0⟩ ⟨0∣ψ⟩ = ∣a∣ , (26a)
2
p1 = ⟨ψ∣P1 ∣ψ⟩ = ⟨ψ∣1⟩ ⟨1∣ψ⟩ = ∣b∣ . (26b)
This shows that the absolute value squared of the amplitudes of a wavefunction when expanding it in an orthonormal basis
provide the probabilities of observing the outcomes corresponding to those basis states. This is sometimes called the Born rule
in quantum mechanics.
Also, using Postulate 4, we can see that the state transformation in the above measurement would be:
⎧
⎪ P0 ∣ψ⟩ 2
with probability p0 = ∣a∣ ,
⎪
∣ψ⟩ ↦ ⎨ P1∣a∣∣ψ⟩ (27)
⎪ 2
⎩ ∣b∣ with probability p1 = ∣b∣ .
⎪
We can easily see that P0 ∣ψ⟩ = ∣0⟩⟨0∣(a ∣0⟩ + b ∣1⟩) = a ⟨0∣0⟩ ∣0⟩ + b ⟨0∣1⟩ ∣0⟩ = a ∣0⟩, where in the last step we have used that
{∣0⟩ , ∣1⟩} is an orthonormal set. Similarly, P1 ∣ψ⟩ = b ∣1⟩. Hence the transformation (27) becomes,
⎧
⎪ 2
⎪ ∣a∣ ∣0⟩ = e a ∣0⟩ with probability p0 = ∣a∣ ,
a iθ
∣ψ⟩ ↦ ⎨ b , (28)
⎪
⎪
2
∣1⟩ = eiθb ∣1⟩ with probability p1 = ∣b∣ .
⎩ ∣b∣
where θa and θb are the arguments of the complex numbers a and b respectively. We can see that the phase factors eiθa,b are
completely arbitrary since they have no influence on the probabilities of the measurement outcomes.
Thus, quantum states are equivalent up to a global phase factor. Because of this, quantum states are rays in a Hilbert space,
since they are not just one vector but an equivalence class of vectors: equivalent up to a global phase.
• Pauli matrices: The Pauli matrices and their properties are reviewed in Appendix A 8. Consider, e.g., measuring the Pauli
matrix Z on a qubit ∣ψ⟩ = a ∣0⟩ + b ∣1⟩. Writing the spectral decomposition of Z,
we can see that the set of measurement operators for this observable is {P0 = ∣0⟩⟨0∣, P1 = ∣1⟩⟨1∣}, with outcomes as the
2
corresponding eigenvalues {λ0 = +1, λ1 = −1}. Thus, we obtain λ0 with probability p0 = ⟨ψ∣P0 ∣ψ⟩ = ∣a∣ , and obtain λ1
2
with probability p1 = ⟨ψ∣P0 ∣ψ⟩ = ∣a∣ .
Hence, the action of measuring Z takes ∣ψ⟩ to eiθa ∣0⟩ if the outcome was λ0 ; and to eiθb ∣1⟩ if the outcome was λb .
• Measuring Energy: When measuring energy, the observable we use is simply the Hamiltonian H of the system. Since H
is Hermitian it has a spectral decomposition, We can write H as,
where Ea denotes the energy and ∣a⟩ the associated energy eigenstate. So, in our experiment, we measure an energy of
Ea with probability pa = ⟨ψ∣Pa ∣ψ⟩, where Pa = ∣a⟩ ⟨a∣. The post-measurement state is ∣ψa ⟩ = P√ap∣ψ⟩
a
= ⟨a∣ψ⟩
√
pa
∣a⟩, i.e., the
new state is an eigenstate of the Hamiltonian, also sometimes called an energy eigenstate.
11
H = ωx σx + ωz σz . (31)
What happens when we measure it in the state ∣ψ⟩? To find √ out we need H’s spectral decomposition, i.e., we need to
diagonalize H. The eigenvalues are easily found to be E± = ± ωx2 + ωz2 , so that H can be written in diagonal form as
where ∣E± ⟩ are the corresponding eigenvectors. When H is measured, we find E± with probability p± = ⟨ψ∣E± ⟩⟨E± ∣ψ⟩ =
∣ ⟨ψ∣ E± ⟩∣2 . For example, if the system is prepared in the ground state ∣E− ⟩ (the state with the lower energy), then p− = 1
but p+ = 0. Or,√ if the system is prepared in a uniform superposition of the ground state ∣E− ⟩ and the excited state ∣E+ ⟩,
i.e., ∣ψ⟩ = 1/ 2(∣E− ⟩ + ∣E+ ⟩), then p− = p+ = 1/2.
Given an observable A = ∑a λa ∣a⟩ ⟨a∣, since we obtain λa with probability pa , we can naturally define an expectation value
of this observable in the state ∣ψ⟩ as
The last equality can be proved as follows: first create an orthonormal basis for the Hilbert space with ∣ψ⟩ as one of the elements
of the basis (say by using the Gram-Schmidt procedure [3][p.66]. That is,
where d is the dimension of the Hilbert space and all vectors in the basis are orthonormal. Now,
d−1
Tr(A ∣ψ⟩ ⟨ψ∣) = ∑ ⟨φi ∣A∣ψ⟩ ⟨ψ∣φi ⟩ (35a)
i=0
d−1
= ∑ ⟨φi ∣A∣φ0 ⟩ ⟨φ0 ∣φi ⟩ (35b)
i=0
= ⟨φ0 ∣A∣φ0 ⟩ (35c)
= ⟨ψ∣A∣ψ⟩ , (35d)
where in Eq. (35b) we used the fact that ∣ψ⟩ = ∣φ0 ⟩ is the first element in our basis.
Likewise we can define the standard deviation as follows:
√
2
∆A = ⟨(A − ⟨A⟩ψ ) ⟩ (36)
ψ
In quantum mechanics, an important property of pairs of non-commuting observables is that they cannot be measured with
arbitrary precision simultaneously. What this means is that if we measure the Hermitian operators C and D on ∣ψ⟩ then they
obey the Heisenberg Uncertainty Principle:
1
(∆C) (∆D) ≥ ∣⟨ψ∣[C, D]∣ψ⟩∣ . (37)
2
12
Let us now show this. Define the Hermitian operators A = C −⟨C⟩, and B = D −⟨D⟩. We can always decompose the expectation
value as a complex number:
Note that 2AB = {A, B} + [A, B], where {A, B} is Hermitian (purely real eigenvalues) and [A, B] is anti-Hermitian (purely
imaginary eigenvalues). Therefore
∗
⟨ψ∣{A, B}∣ψ⟩ = ⟨ψ∣{A, B}† ∣ψ⟩ = ⟨ψ∣{A, B}∣ψ⟩ (39a)
∗
⟨ψ∣[A, B]∣ψ⟩ = ⟨ψ∣[A, B] ∣ψ⟩ = − ⟨ψ∣[A, B]∣ψ⟩ ,
†
(39b)
which means that ⟨ψ∣{A, B}∣ψ⟩ is real while ⟨ψ∣[A, B]∣ψ⟩ is purely imaginary. Hence the following must be true:
Therefore, by using the Cauchy-Schwarz inequality (see Appendix A) in the third line:
2 2
4x2 + 4y 2 = ∣⟨ψ∣{A, B}∣ψ⟩∣ + ∣⟨ψ∣[A, B]∣ψ⟩∣ (41a)
2
= 4 ∣⟨ψ∣AB∣ψ⟩∣ (41b)
≤ 4 ⟨ψ∣A A∣ψ⟩ ⟨ψ∣B B∣ψ⟩
† †
(41c)
= 4 ⟨ψ∣A ∣ψ⟩ ⟨ψ∣B ∣ψ⟩
2 2
(41d)
2 2
= 4 ⟨ψ∣ (C − ⟨C⟩) ∣ψ⟩ ⟨ψ∣ (D − ⟨D⟩) ∣ψ⟩ . (41e)
2
Obviously ∣⟨ψ∣{A, B}∣ψ⟩∣ ≥ 0, and hence:
2 2 2
4 ⟨ψ∣ (C − ⟨C⟩) ∣ψ⟩ ⟨ψ∣ (D − ⟨D⟩) ∣ψ⟩ ≥ ∣⟨ψ∣[A, B]∣ψ⟩∣ (42a)
2
= ∣⟨ψ∣[C, D]∣ψ⟩∣ (42b)
Given a generalized measurement with measurement operators {Mk } we define the elements of a POVM via
Ek = Mk† Mk . (43)
†
The normalization condition then becomes ∑k Ek = I. Clearly, Ek† = Mk† (Mk† ) = Ek , so that the POVM elements are
Hermitian. It is easy to show that the Ek s are moreover positive operators, i.e., that ⟨ψ∣Ek ∣ψ⟩ ≥ 0 is true for every ∣ψ⟩ (for more
2
details on positive operators see Appendix A 7). Indeed, ⟨ψ∣Ek ∣ψ⟩ = ⟨ψ∣Mk† Mk ∣ψ⟩ = ∥Mk ∣ψ⟩∥ ≥ 0. Note that the probability
of outcome k is simply pk = ⟨ψ∣Ek ∣ψ⟩. How about the effect of the measurement Ek on a state ∣ψ⟩? Suppose we are given
an arbitrary set of positive operators {Ek } that satisfy ∑k Ek = I. How do we extend the measurement postulate in this case?
√ It is true that for any operator A, we can always
The answer to this is to use the so called polar decomposition of the operator.
find a unitary U and a positive operator P such that A = U P with P = A† A. If the operator A is invertible, then such a
decomposition is unique and U = AP −1 . In our case, we could use the given POVMs and define for every k
√
Mk = U k E k , (44)
where the Uk ’s are just arbitrary unitaries. In other words, since only the Ek are specified (by assumption), we are free to choose
the Uk ’s, and for every such choice we get a different set of Mk ’s. Hence, we can now write the state after the measurement as
√
Uk Ek ∣ψ⟩
∣ψ⟩ ↦ ∣ψk ⟩ = √ with probability pk = ⟨ψ∣Ek ∣ψ⟩ . (45)
pk
Since Uk ’s are arbitrary (again, since only the Ek ’s were specified), this unitary freedom is a generalization of the freedom to
leave the overall phase of a state unspecified.
13
To see why POVMs are relevant let’s consider the following example. Suppose we have to play a game. Alice always gives
us one of these two states:
∣ψ1 ⟩ = ∣0⟩ , or (46a)
1
∣ψ2 ⟩ = √ (∣0⟩ + ∣1⟩) ≡ ∣+⟩ . (46b)
2
We do not, a priori, know which state has been handed to us. We do know that it is one of these two states. Our task is
to perform measurements and decide which of the two states we were given. Also, we are not allowed to make an error in
identification, i.e., if we provide an answer, it has to be right. However, we are allowed to proclaim ignorance if we don’t
know the answer. Moreover, we must treat both states equally, i.e., we cannot preferentially identify only one of the states
and proclaim ignorance on the other. What is our strategy? Since these are non-orthogonal states, there is no way distinguish
these two states with complete certainty [3][Box 2.3, p.87]. Suppose we try to do it with projective measurements. Let’s take
the measurement set to be {Mk } = {P0 , P1 }, where Pi = ∣i⟩ ⟨i∣ , i = 0, 1. Suppose that the outcome is the index 0. This can
happen in either of two ways: Alice prepared ∣ψ1 ⟩ or she prepared ∣ψ2 ⟩. The probability that the outcome is 0 given that she
prepared ∣ψ1 ⟩ is p(0∣ψ1 ) = ⟨ψ1 ∣P0 ∣ψ1 ⟩ = ⟨0∣0⟩ ⟨0∣0⟩ = 1. And, the probability that the outcome is 0 given that she prepared ∣ψ2 ⟩
is p(0∣ψ2 ) = ⟨ψ2 ∣P0 ∣ψ2 ⟩ = 1/2. This means that if the outcome is 0 then we cannot know for sure which of the two states Alice
prepared, since both occur with non-vanishing probability. Therefore in this case we must proclaim ignorance. However, note
that it also follows that p(1∣ψ1 ) = 0 and p(1∣ψ2 ) = 1/2 which means that given outcome 1 we know with certainty that Alice
prepared ∣ψ2 ⟩. Thus we cannot satisfy the condition of treating the two states equally. As is easily checked, this will always be
the case with a projective measurement.
Now, let’s try with an intelligent choice of POVMs. Define,
E1 = α∣1⟩⟨1∣, (47a)
E2 = α ∣−⟩ ⟨−∣ , (47b)
E3 = I − E1 − E2 . (47c)
where, ∣−⟩ ≡ √12 (∣0⟩ − ∣1⟩) and α > 0 is an arbitrary parameter which we can optimize later, and which must be chosen so that
E3 > 0. If we do so then this clearly is a set of POVMs, since ∑k Ek = I, and for suitable α, all the Ek ’s are positive. Let us
now compute the probabilities of the 3 possible outcomes,
p(1∣ψ1 ) = ⟨ψ1 ∣E1 ∣ψ1 ⟩ = 0, (48a)
α
p(1∣ψ2 ) = ⟨ψ2 ∣E1 ∣ψ2 ⟩ = , (48b)
2
α
p(2∣ψ1 ) = ⟨ψ1 ∣E2 ∣ψ1 ⟩ = , (48c)
2
p(2∣ψ2 ) = ⟨ψ2 ∣E2 ∣ψ2 ⟩ = 0, (48d)
α
p(3∣ψ1 ) = ⟨ψ1 ∣E3 ∣ψ1 ⟩ = 1 − , (48e)
2
α
p(3∣ψ2 ) = ⟨ψ2 ∣E3 ∣ψ2 ⟩ = 1 − . (48f)
2
So, if we get outcome 1, we can say with certainty that the given state was ∣ψ2 ⟩ and if we get outcome 2, we can say with certainty
that the given state was ∣ψ1 ⟩. With outcome 3, we have no information about the state, i.e., we must proclaim ignorance. But in
two of the three outcomes we have been able to obtain an answer with certainty. So, in order to make the probability of outcome
3 as small as possible (since it yields no information), we have to increase α as much as possible while keeping E3 positive. If
we write out E3 as a matrix and√place the constraint of the eigenvalues of this matrix being positive, it easy to show that the
maximal allowed value of α is 1+√2 2 .
We will motivate the study of density operators by considering ensembles of pure quantum states. Suppose, instead of having
a single state vector, we only know that our system is in state ∣ψ1 ⟩ with probability q1 , or in state ∣ψ2 ⟩ with probability q2 , and
so on. In other words, we have an pure state ensemble {qi , ∣ψi ⟩}N
i=1 describing our system.
Now, we would like to understand what happens when we make measurements on this quantum system. Suppose the state
were ∣ψi ⟩ and we measure with a set of measurement operators {Mk }. The measurement transformation would be:
Mk ∣ψi ⟩
∣ψi ⟩ ↦ √ = ∣ψik ⟩ (49)
pk∣i
14
with probability pk∣i = ⟨ψi ∣Mk† Mk ∣ψi ⟩, which is the probability of outcome k, given a state ∣ψi ⟩.
Now, consider that we did not know what the state was but only that it came from the ensemble {qi , ∣ψi ⟩}N
i=1 . Then the
probability of obtaining the outcome k as a result of the measurement on the ensemble is:
pk = ∑ pk∣i qi (50a)
i
This is called the density matrix or density operator and is a central object in quantum mechanics. The density matrix is
completely equivalent to the pure state ensemble {qi , ∣ψi ⟩}N
i=1 , but it has the advantage of being directly useful for calculations.
Indeed, using the density matrix, Eq. (50c) becomes:
pk = Tr(Ek ρ), (52)
where we have defined Ek ≡ Mk† Mk as the element of a POVM.
What about the state that results after measurement result k has been observed? Suppose that outcome k is observed for a
known initial state ρ = ∑i qi ∣ψi ⟩ ⟨ψi ∣. If we let ∣ψik ⟩ ∶= M√kp∣ψi ⟩ [as in Eq. (49)], then {pi∣k , ∣ψik ⟩}i is the resulting ensemble,
k∣i
where pk∣i is the probability of outcome k given state ∣ψi ⟩. On the other hand, if outcome k was observed, and we don’t know
the initial state, then we should sum over all possible states compatible with outcome k (the states ∣ψik ⟩) with their respective
conditional probabilities pi∣k . Thus, the density operator for result k becomes
ρk = ∑ pi∣k ∣ψik ⟩⟨ψik ∣ (53a)
i
• Unit trace: The trace operation is reviewed in Appendix A 6. The density operator ρ has Tr [ρ] = 1. This property can
easily be seen by the following calculation:
Tr [ρ] = ∑ qi Tr [∣ψi ⟩⟨ψi ∣] = ∑ qi = 1. (55)
i i
• Hermiticity: The density operator ρ is Hermitian. The following line demonstrates this
ρ† = ∑ qi∗ (∣ψi ⟩⟨ψi ∣) = ∑ qi ∣ψi ⟩⟨ψi ∣ = ρ
†
(56)
i i
where we’ve used that probabilities qi are real and projectors formed from outer-products are Hermitian.
15
• Positive definite: For all vectors ∣ν⟩ ∈ H, the density operator ρ has ⟨ν∣ρ∣ν⟩ ≥ 0:
2
⟨ν∣ρ∣ν⟩ = ∑ qi ∣⟨ψi ∣v⟩∣ ≥ 0, (57)
i
since the qi are all non-negative by virtue of being probabilities. But since Trρ = 1 it clearly must have at least one
eigenvalue that is non-zero. Therefore ρ must be positive, not just positive semi-definite (positive operators are defined in
Appendix A 7).
Note that positivity implies Hermiticity, since an operator is Hermitian iff it has only real eigenvalues. Therefore we don’t
actually need to separately stipulate Hermiticity. Also note that the density operator deserves to be called an operator: it acts as
a transformation between two copies of the Hilbert space, i.e., ρ ∶ H ↦ H.
We define the space of positive, trace-one linear operators acting on H as D(H). Thus
ρ ∈ D(H). (58)
Recall the two equivalent descriptions of dynamics of the pure quantum state
where U (t) and H are related by U (t) = e−iHt . Consider one of the pure states forming the ensemble {qi , ∣ψi ⟩}i . This state will
evolve as
We can now summarize the four postulates in terms of the density operator.
16
1. The state space is the Hilbert-Schmidt space of linear operators ρ such that Tr [ρ] = 1 and ρ > 0. The inner product in the
Hilbert-Schmidt space is defined as Tr[A† B] for any√two operators
√ A and B acting on the same Hilbert space. This inner
product defines a length in the usual way, i.e., ∥ρ∥ = ⟨ρ, ρ⟩ = P . The quantity
P ≡ Tr[ρ2 ] (64)
is called the “purity” of the state ρ. Thus a density matrix can have “length” ≤ 1. A state is called “pure” if P = 1 and
“mixed” if P < 1.
2. State spaces are composed via the tensor product ⊗.
3. Density operators evolve as ρ̇ = −i [H, ρ] under a Hamiltonian H, or equivalently as ρ(t) = U (t)ρ(0)U † (t) where the
unitary U (t) = e−itH .
4. A general measurement operation defined by elements {Mk } results with probability pk = Tr [ρMk† Mk ] in the state
Mk ρMk†
transformation ρ ↦ pk
.
Expectation values are now computable in terms of the ρ as well. Consider an observable A measured for a system in the pure
state ensemble {qi , ∣ψi ⟩}. Previously, in Eq. (33), we showed that the expectation value was ⟨A⟩ψ = ⟨ψ∣A∣ψ⟩ = Tr(A ∣ψ⟩ ⟨ψ∣).
We need to modify this by assigning each pure state ∣ψi ⟩ in the ensemble its weight qi . Thus the new expression for the
expectation value is:
⟨A⟩ρ = ∑ qi ⟨ψi ∣A∣ψi ⟩ = ∑ qi Tr(A ∣ψi ⟩ ⟨ψi ∣) = Tr(A ∑ qi ∣ψi ⟩ ⟨ψi ∣) = Tr(Aρ) = Tr(ρA). (65)
i i i
We defined “pure” and “mixed” states above according to the value of the purity P = Tr[ρ2 ] being 1 or < 1. Prior to
introducing the density operator formalism, we had considered quantum states as vectors in the Hilbert space. This formalism
is equivalent to pure state ensembles of the type {1, ∣ψ⟩}, i.e., having only a single element. It is not hard to see that such
special ensembles are “pure” quantum states. The associated density operator is ρ = ∣ψ⟩⟨ψ∣. It is useful to think of pure states as
ensembles with only one member and probability 1.
Any state that is not not pure is by definition mixed. This means that they are described by ensembles of the form {pi , ∣ψ⟩i }
where for all i, 0 < pi < 1. The density operator associated with a mixed ensemble is the mixture of the pure states with their
associated weights [as seen in Eq. (51)].
17
Note that a pure state is a projector: (∣ψ⟩⟨ψ∣)(∣ψ⟩⟨ψ∣) = ∣ψ⟩⟨ψ∣. Therefore, if a state ρ is pure then ρ2 = ρ. The converse is also
true: ρ2 = ρ implies that ρ is pure. It is easy to check that these conditions are equivalent to the definition in terms of purity P .
We can also define a mixed state ensemble, i.e., a collection of mixed states ρk with associated probabilities pk , as
{pk , ρk } ⇔ ρ = ∑ pk ρk . (68)
k
When are two pure state ensembles equivalent? Consider for example the two ensembles
{(3/4, 1/4), (∣0⟩ , ∣1⟩)} (69a)
{(1/2, 1/2), (∣a⟩ , ∣b⟩)}, (69b)
where
√ √
∣a⟩ = 3/4 ∣0⟩ + 1/4 ∣1⟩ (70a)
√ √
∣b⟩ = 3/4 ∣0⟩ − 1/4 ∣1⟩ . (70b)
On the face of it, the first of these ensembles represents a biased classical coin (“heads”, or 0, with probability 3/4, tails, or 1
with probability 1/4), whereas the second is quantum in the sense that each state is a superposition state. But are they really
different? It is straightforward to check that in fact the two density matrices corresponding to these two ensembles are equal.
This being the case, there is no measurement that can distinguish them, and that means we must consider them to be the same.
Theorem 1. Two pure state ensembles with the same number of elements2 {qi , ∣ψi ⟩}i and {rj , ∣φj ⟩}j correspond to the same
density operator if and only if there exists a unitary U with entries [U ]ij such that
√ √
qi ∣ψi ⟩ = ∑[U ]ij rj ∣φj ⟩ (71)
j
Proof. We show explicitly the “if” direction of the proof. The complete proof is found in [3][p.104]. Consider the following
mixture,
√ √
∑ qi ∣ψi ⟩⟨ψi ∣ = ∑ ( qi ∣ψi ⟩) (⟨ψi ∣ qi ) (72a)
i i
⎛ √ ⎞ √
= ∑ ∑[U ]ij rj ∣φj ⟩ (∑ ⟨φk ∣ rk [U † ]ik ) (72b)
i ⎝ j ⎠ k
√
=∑ rj rk (∑[U ]ij [U † ]ik ) ∣φj ⟩⟨φk ∣ (72c)
j,k i
√
=∑ rj rk (δjk ) ∣φj ⟩⟨φk ∣ (72d)
j,k
where in Eq. (72d) we used the unitarity of U . Thus the two ensembles represent the same density operator.
A qubit is a quantum state ∣ψi ⟩ in a two-dimensional Hilbert space H = C2 = span{∣0⟩ , ∣1⟩} where ∣0⟩ and ∣1⟩ form an
orthonormal basis for H. The density operator for any state in this space is thus of the form ∑i qi ∣ψi ⟩⟨ψi ∣ and can hence be
represented by a 2 × 2 complex matrix of the form
a b
ρ=[ ]. (73)
c d
2 If necessary pad the smaller set with zeroes to make it equal in length to the larger set.
18
However, applying the properties of density operators can reduce this to an expression of only two variables. First, the unit trace
reduces to the condition d = 1 − a and Hermiticity reduces to the condition that c = b∗ and that a be real. Thus, the density matrix
is completely parametrized by the complex number b and the real number a and takes the form
a b
ρ=[ ]. (74)
b∗ 1 − a
Positivity is the statement that the eigenvalues λ± are non-negative:
∣ρ − λI∣ = 0 ⇒ λ2 − (Trρ)λ + ∣ρ∣ = 0, (75)
i.e., using Trρ = 1:
1 √
λ± = (1 ± 1 − 4∣ρ∣) ≥ 0. (76)
2
This parametrization requires only three parameters and we can thus embed it naturally in three dimensions. Before we proceed
to do this we will decompose the density operator one more time but in a more useful basis.
Recall the Pauli matrices σx , σy , σz , σ0 . Any qubit density matrix can represented by
1 1
ρ= (I + ∑ vi σi ) = (I + v⃗ ⋅ σ
⃗) , (77)
2 i 2
where v⃗ = (vx , vy , vz ) and σ
⃗ = (σx , σy , σz ). In terms of the elements of v⃗, ρ appears as
1 1 + vz vx − ivy
ρ= [ ]. (78)
2 vx + ivy 1 − vz
To relate this to our previous analysis simply let b = 1/2(vx − ivy ) and a = 1/2(1 + vz ). We call v⃗ the Bloch vector. The 2 × 2
matrix we have constructed using the Bloch vector is not, however, necessarily a valid quantum state. Unit trace is guaranteed
by the construction, and positivity can now be made explicit by noting that
1 1
∣ρ∣ = v ∥2 ) ,
(1 − vz2 − (vx2 + vy2 )) = (1 − ∥⃗ (79)
4 4
so that
1
λ± = (1 ± ∥⃗
v ∥) . (80)
2
The two solutions are ∥⃗
v ∥ ≤ 1 and ∥⃗
v ∥ ≥ −1, which is trivially satisfied. Thus if we require positivity, the relevant constraint is
∥⃗
v∥ ≤ 1 . (81)
Let us also relate the magnitude of the Bloch vector to the purity of the quantum state. Recall that a pure quantum state is a
projector and thus ρ2 = ρ for pure states. If we calculate the density operator ρ2 we find
1 1 2
ρ2 = (I + v⃗ ⋅ σ
⃗ ) (I + v⃗ ⋅ σ
⃗ ) = (I + 2⃗ ⃗ + (⃗
v⋅σ ⃗) ) .
v⋅σ (82)
4 4
The term (⃗ ⃗ )2 becomes
v⋅σ
∑ vk vl σk σl . (83)
k,l∈{x,y,z}
⃗ )2 =
v⋅σ
Recall Eq. (A21). Taking the trace and noting that the Pauli matrices are traceless only the δkl term remains. Thus Tr(⃗
2
∥⃗
v ∥ TrI, with TrI = 2, and Eq. (82) yields:
1
(1 + ∥⃗
Trρ2 = v ∥2 ) . (84)
2
From this form it is clear that any unit Bloch vector will make Trρ2 = 1, i.e., a pure state, and Bloch vectors of length less than
1 yield mixed states.
Having gathered the requisite intuition for the geometry at hand, we call the set of all valid Bloch vectors v⃗ the Bloch sphere,
also known as the Poincaré sphere in optics.
Since the Bloch sphere can describe all qubit states and can be embedded in three dimensions it is a useful tool for illustrating
various common qubit states.
19
FIG. 1. The Bloch sphere is a geometric representation of the collection of all Bloch vectors v⃗ which describe valid qubit density operators.
Thus, the sphere is of radius 1, its surface represents all pure states, and its interior represents all mixed states. In this diagram the blue vector
lies on the surface of the sphere indicating a pure state, whereas the red vector lies in its interior indicating a mixed state.
Now that we have discussed in detail the dynamics of a single system, let us consider more complex systems. Consider a two
component system, where we have a subsystem of our interest, A (often we’ll just call it “system”), and the other subsystem is
20
the bath, B. Together, the system and the bath comprise the lab, or even the entire universe. We can think of A as a quantum
computer, or a molecule, or any other system we’re interested in studying. We shall assume that the total system evolves
according to the Schrödinger equation and that it is described by a density matrix ρ(t). Further, let the subsystem Hilbert spaces
be
HA = span{∣i⟩A } (85)
HB = span{∣µ⟩B } (86)
Here, i goes from 0 to dA − 1, the dimension of the Hilbert space of A, and µ goes from 0 to dB − 1, the dimension of the Hilbert
space of B. Usually, the dimension of the bath, dB → ∞, while dA is finite. By the second postulate, the Hilbert space of the
two system combined is the tensor product of the individual spaces:
H = HA ⊗ HB (87a)
= span{∣i⟩A ⊗ ∣µ⟩B } (87b)
Let us figure out the structure of a density matrix in this combined Hilbert space. We can define a pure state ensemble {∣Ψa ⟩ , qa }
for a set of pure states ∣Ψa ⟩ ∈ H. Each of these states can be expanded in the basis above, i.e.,
ρ = ∑ qa ∣Ψa ⟩⟨Ψa ∣ = ∑ qa (∑ ca;iµ ∣i⟩A ⊗ ∣µ⟩B )(∑ c∗a;jν ⟨j∣A ⊗ ⟨ν∣B ). (89)
a a i,µ j,ν
Therefore any density matrix in the combined Hilbert space can be written down as
“factorized” state. Such states exhibit no correlations at all between the A and B subsystems. Clearly, however, this is a special
case and in general, ρ cannot be factored in this manner. When it cannot, the subsystems are correlated. These correlations can
be quantum (due to entanglement), classical, or both.
We are primarily interested in the system A. We thus need to find a way to remove the bath B from our description. To do,
we now define a new operation called partial trace, which effectively averages out of the components of B from the combined
density matrix. The resultant density matrix then describes only A.
B. Partial Trace
1. Definition
The partial trace is a linear operator that maps from the total Hilbert space to the Hilbert space of A, i.e., H ↦ HA , defined as
follows. Consider an operator O = MA ⊗ NB such that O acts on H = HA ⊗ HB . Then
It is understood in the last line that the basis vectors {∣µ⟩}, which span the space HB , act only on the second Hilbert space.
In other words, the expression ⟨µ∣[MA ⊗ NB ]∣µ⟩ is a partial matrix element, where the matrix element is taken only over the
21
second factor, and the result is an operator acting on HA . Thus, if O = ∑ij `ij MAi ⊗ NBj , then by linearity:
= ∑ ⟨µ∣O∣µ⟩ (92d)
µ
By linearity,
TrB [ ∑ λijµν ∣i⟩A ⟨j∣ ⊗ ∣µ⟩B ⟨ν∣] = ∑ λijµν ∣i⟩A ⟨j∣ ⟨ν∣µ⟩B (94a)
ijµν ijµν
where in the second line we assumed that {∣µ⟩B } forms an orthonormal basis, and we defined λ̄ij = ∑µ λijµµ . This shows that
taking the partial trace leads to a form that looks like a density matrix for the A subsystem. Of course, we’ll have to verify that
it satisfies the properties of a density matrix (unit trace and positivity). Positivity is more challenging, but unit trace is obvious if
we assume (as we should) that Trρ = 1. For, it is then easy to check that this implies ∑iµ λiiµµ = 1. On the other hand, if we are
to interpret TrB [ρ] as a valid density matrix then Tr(TrB [ρ]) = ∑i λ̄ii should be 1, which it is, since it equals ∑iµ λ̄iiµµ .
Crucially, we now claim that the density matrix of the subsystem A is given by taking the partial trace of the combined density
matrix with respect to B.
1. Case 1: Consider a case where the states of the bath and system are completely separate, and hence form a tensor product.
In such a case, we expect that the density operator of A obtained by partial trace should be the same as the component
of A contributed in the tensor product. And indeed, clearly, if ρ = ρA ⊗ ρB where both terms in the product are properly
normalized states, then
1
∣ψ⟩AB = √ (∣0⟩A ∣0⟩B + ∣1⟩A ∣1⟩B ). (97)
2
This means that the state state ∣ψ⟩AB contains no separate information about A or B’s state. The reason is that if we
measure, say, B using the measurement operators {M0 = ∣0⟩⟨0∣, M1 = ∣1⟩⟨1∣}, then we find the outcomes 0 and 1 with
equal probability 1/2, and at the same time the state of A becomes either ∣0⟩ or ∣1⟩, respectively. It is easy to check that
this random outcome remains true for any other choice of measurement operators. This means we gain no knowledge at
22
all about A or B since the measurement outcome is perfectly random. In terms of the partial trace we find:
ρA = TrB [ρ] = TrB [∣ψ⟩AB ⟨ψ∣] (98a)
1
= TrB [∣0⟩A ⟨0∣ ⊗ ∣0⟩B ⟨0∣ + ∣0⟩A ⟨1∣ ⊗ ∣0⟩B ⟨1∣ + ∣1⟩A ⟨0∣ ⊗ ∣1⟩B ⟨0∣ + ∣1⟩A ⟨1∣ ⊗ ∣1⟩B ⟨1∣]
2
(98b)
1
= [∣0⟩⟨0∣ × 1 + ∣0⟩⟨1∣ × 0 + ∣1⟩⟨0∣ × 0 + ∣1⟩⟨1∣ × 1] (98c)
2
1
= [∣0⟩A ⟨0∣ + ∣1⟩A ⟨1∣] = IA /2 (98d)
2
Therefore, the state of A is an equal probabilistic mixture of the ∣0⟩ and ∣1⟩ states, as expected.
Consider a composite system with the Hilbert space H = HA ⊗ HB . If we had an observable MA on subsystem A, then, the
expectation value of that operator would be given by
⟨MA ⟩ρA = Tr[ρA MA ], (99)
where we used Eq. (65).
̃ = MA ⊗ IB on the entire
However, in the case of this composite system, this measurement is actually of the observable M
system ρ in HA ⊗ HB , where we do nothing (the identity operation) to B. Thus,
̃⟩ = Tr[ρ M
⟨M ̃] (100)
The key idea is that these two operations should correspond to the same physical observation and they should produce the
same number. For the theory to be consistent, we demand that
̃⟩,
⟨MA ⟩ ≡ ⟨M (101)
i.e.,
̃ρ]
Tr[MA ρA ] = Tr[M (102)
It can be shown that this condition is satisfied iff we define ρA ≡ TrB (ρ). We shall prove the theorem in one direction, that is,
if ρA = TrB (ρ), then ⟨M ⟩ = ⟨M̃⟩.
In going from Eq. (103b) to (103c), we used the expression for the partial trace over operators as given in Eq. (92d). But note
that ρ is an operator acting on the composite system, not just on A, since B ⟨µ∣ ρ∣µ⟩B is a partial matrix element. If we wish to
likewise consider MA as an operator acting on the composite system, then we should extend it to MA ⊗ IB . Also, the correct
order for the product ∣µ⟩B ∣i⟩A , including the tensor product symbol explicitly, is: ∣i⟩A ⊗ ∣µ⟩B . Thus:
⟨MA ⟩ = ∑ ⟨i∣ ⊗ ⟨µ∣ [ρ(MA ⊗ IB )] ∣i⟩ ⊗ ∣µ⟩ (104a)
i,µ
= Tr[ρ(MA ⊗ IB )] (104b)
̃⟩ ,
= ⟨M (104c)
̃⟩.
which shows the desired equality ⟨MA ⟩ = ⟨M
23
In this section we shall find the dynamical evolution of an open quantum system.
Consider a system S and bath B, such that they have a joint unitary evolution given by U (t) = e−iHt . The initial joint state is
ρ(0). Then, by Schrödinger’s equation,
ρ(t) = U (t)ρ(0)U † (t) (105)
As the density operator of the bath is positive and normalized, it has a spectral decomposition in an orthonormal basis with
non-negative eigenvalues. Hence
ρB (0) = ∑ λν ∣ν⟩⟨ν∣ (106)
ν
where λν are the eigenvalues (probabilities) and {∣ν⟩} are the corresponding orthonormal eigenvectors.
The state of the system is then found by performing a partial trace over the bath, i.e.,
ρS (t) = TrB [ρ(t)]. (107)
We can perform the partial trace in the orthonormal basis of bath eigenstates, i.e.,
ρS (t) = TrB [U (t)ρ(0)U † (t)] (108a)
= ∑ ⟨µ∣U (t)ρ(0)U (t)∣µ⟩
†
(108b)
µ
Let us now assume that the initial state is completely decoupled, that is
ρ(0) = ρS (0) ⊗ ρB (0). (109)
Then
ρS (t) = ∑ ⟨µ∣[U (t)ρS (0) ⊗ ∑ λν ∣ν⟩⟨ν∣U † (t)]∣µ⟩ (110a)
µ ν
√ √
=∑ λν ⟨µ∣U (t)∣ν⟩B ρS (0) λν ⟨ν∣U † (t)∣µ⟩B (110b)
µν
The system-only operators {Kµν } are called the Kraus operators and are given by
√
Kµν (t) = λν ⟨µ∣U (t)∣ν⟩ (111)
(note the partial matrix element, leaving us with an operator acting on the system), and the equation defining the evolution of
the system in terms of Kraus operator is called the Kraus Operator Sum Representation (OSR)
This is a pivotal result; as we shall see it includes the Schrödinger equation as a special case.
This criterion can be verified for our definition of Kraus operators, given by Eq. (111).
∑ Kµν Kµν = ∑ λν ⟨µ∣U (t)∣ν⟩ ⟨ν∣U (t)∣µ⟩
† †
(115a)
µν µν
⎛ ⎞
= ∑ λν ⟨ν∣U † (t) ∑ ∣µ⟩⟨µ∣ U (t)∣ν⟩ (115b)
ν ⎝ µ ⎠
= ∑ λν ⟨ν∣ν⟩ (115c)
ν
= ∑ λν = 1 (115d)
Thus, such a set of Kraus operators preserves normalization.
Note that when there is just a single Kraus operator, the normalization condition (114) forces it to be unitary, which is just the
case of closed system evolution! We can see more explicitly how this comes about, as follows.
ρS (t) = ∑ cµν US (t)ρS (0)c∗µν US† (t) = US (t)ρS (0)US† (t) , (116)
µν
which is unitary, Schrödinger-like dynamics. Hence, the Kraus operator sum representation is more general than the Schrödinger
equation, because it contains the latter as a special case.
We have seen [Eq. (112)] that the evolution of the state ρS of an open quantum system can be expressed as unitary evolution
of the composite system+bath, followed by a partial trace, which leads to the Kraus operator sum representation (Kraus OSR):
ρS (t) = TrB [U (t)(ρS ⊗ ρB )U † (t)] = ∑ Kα (t)ρS (0)Kα† (t) , (117)
α
where we have collected the earlier µν indices into a single index: α = (µν). From now on let us drop the S subscript since
we’ll be focusing on the system alone. We’ll reintroduce it as necessary.
A. Non-selective measurements
Let us observe that the OSR represents more than dynamics. It can also capture measurements. Specifically, consider mea-
surement operators {Mk } with ∑k Mk† Mk = I. Recall that a state subjected to this measurement maps to
Mk ρMk†
ρ ↦ ρk = (118)
Tr [Mk ρMk† ]
with probability pk = Tr [Mk ρMk† ]. Consider the case where we perform this measurement but do not learn the outcome k.
What happens to ρ after this measurement? In this case
ρ z→ ⟨ρ⟩ = ∑ pk ρk = ∑ Mk ρMk† (119)
k k
which we recognize as a non-selective measurement. This last form is in the Kraus operator-sum representation with the Kraus
operators Mk . Thus, we can encapsulate the non-selective measurement postulate in the operator-sum formalism.
Since both dynamics and measurements are captured by the OSR, and there are no other quantum processes according to our
postulates, this suggests that the OSR is truly fundamental. It thus deserves further scrutiny.
25
It is useful to think of the OSR as a map (or synonymously a process or channel) Φ from the initial to the final system state,
i.e.,
where Φ[X] ≡ ∑α Kα XKα† . Note that Φ is an operator acting on operators, sometimes called a superoperator. While we
started with vectors ∣v⟩ in a Hilbert space HS , and moved the density operators ρ ∶ HS ↦ HS belonging to the space of positive
trace-class operators D(HS ), the map Φ ∶ D ↦ D belongs to D[D(HS )], as we shall see shortly. In terms of dimensions, if
dim(HS ) = d, then dim(D(HS )) = d2 , and dim(D[D(HS )]) = d4 , reflecting the fact that vectors are of dimension d × 1,
density matrices of dimension d × d, and quantum maps of dimension d2 × d2 .
It will prove to be profitable to adopt an even more abstract point of view, and seek to determine the key properties that any
such map possesses. We can easily identify three properties by inspection:
1. Trace Preserving:
where we used the fact that ∑α Kα† Kα = I. Thus the map Φ is trace-preserving.
2. Linear:
By direct substitution we find:
Φ(aρ1 + bρ2 ) = ∑ Tr(Kα aρ1 Kα† ) + ∑ Tr(Kα bρ2 Kα† ) = a ∑ Tr(Kα ρ1 Kα† ) + b ∑ Tr(Kα ρ2 Kα† ) = aΦ(ρ1 ) + bΦ(ρ2 )
α α α α
(122)
for any scalars a and b. Thus the map Φ is linear.
3a. Positivity:
This property means that Φ maps positive operators to positive operators. Assume the operator A > 0, i.e., it has only
non-negative eigenvalues, not all zero. Note that any density matrix ρ must be positive, and we can write A = ∑i λi ∣i⟩⟨i∣
where all λi ≥ 0 (the spectral decomposition of A).
In order to demonstrate that Φ(A) > 0 it is sufficient show that ⟨ν∣Φ(A)∣ν⟩ ≥ 0 for all ∣ν⟩ ∈ HS , since this means in
particular that the eigenvalues of Φ(A) are all non-negative. Let ∣wa ⟩ = Kα† ∣ν⟩. Then:
On the right hand side it is clear that each term in the sum is positive. Therefore Φ(A) > 0, and Φ itself is a positive map.
The Kraus OSR satisfies these three properties, but does every map that satisfy the same properties have a Kraus OSR? The
answer is negative. It turns out that we must modify and strengthen the positivity property into “complete positivity”.
C. Complete Positivity
The map Φ is a completely positive (CP) map. It maps positive operators to positive operators (is “positivity preserving”), and
(k)
moreover, it can be shown that even Φ ⊗ IR is positive for all k, where k is the dimension of an ancillary Hilbert space HR ,
and IR denotes the identity (super-)operator on HR . Conversely, every CP map can be represented as a Kraus OSR.
More formally, let B(H) denote the space of linear operators acting on the Hilbert space H, i.e., X ∶ H ↦ H is equivalent
to X ∈ B(H). Let A ∈ B(HS ⊗ HR ), where HS denotes the system space and HR is some auxiliary space with dimension k.
Assume that A > 0. Denote by IR the identity map on B(HR ) [i.e., IR (V ) = IV I for all V ∈ B(HR )]. Also, let Φ ∈ B[B(HS )],
i.e., Φ ∶ B(HS ) ↦ B(HS ).
To prove that Φ is CP we need to show that Φ ⊗ IR is positive for all d = dim(HR ). Indeed:
⟨ν∣ (Φ ⊗ IR )(A) ∣ν⟩ = ∑ ⟨ν∣ (Kα ⊗ IR )A(Kα† ⊗ IR ) ∣ν⟩ = ∑⟨wα ∣A∣wα ⟩ > 0 , (125)
α α
where we defined ∣wα ⟩ = (Kα† ⊗ IR ) ∣ν⟩, where now ∣ν⟩ ∈ HS ⊗ HR , and we drew upon the fact that A > 0, as in Eq. (123).
The key feature of the Kraus OSR that makes it a completely positive map is having the same operator (Kα ) on both sides.
For example, something like ∑αβ Kα XKβ† is not a CP map, and the proof of positivity as in Eq. (125) would clearly not have
worked.
To prove the reverse direction, that all maps that satisfy conditions 1, 2, 3b have a Kraus OSR, is more challenging and requires
a tool known as the Choi decomposition [4].
From now on we define a quantum map (or quantum channel) as a map that is (1) trace preserving, (2) linear, (3) completely
positive. This definition is motivated by the fact that we know that such maps have a Kraus OSR, and that the Kraus OSR arises
both from the physical prescription of unitary evolution followed by partial trace, and from (non-selective) measurements.
Do maps that are positive but not completely positive exist? The answer is affirmative. The canonical example is the elemen-
tary transpose map T .
Given a real basis {∣i⟩} for HS , the action of the transpose on the basis elements is: T (∣i⟩⟨j∣) = ∣j⟩⟨i∣ (for a real basis this is
the same as Hermitian conjugation). For example, for a 2 × 2 matrix:
a b a c
T ∶( ) = a∣0⟩⟨0∣ + b∣0⟩⟨1∣ + c∣1⟩⟨0∣ + d∣1⟩⟨1∣ z→ a∣0⟩⟨0∣ + b∣1⟩⟨0∣ + c∣0⟩⟨1∣ + d∣1⟩⟨1∣ = ( ). (126)
c d b d
Claim 1. T is a positive map.
Proof. To prove the claim it suffices to show that the eigenvalues of X and T (X) are the same for any X ∈ HS [since then
in particular their sign is preserved, so if X > 0 then also T (X) > 0]. The eigenvalues of X are found by solving for the
roots of its characteristic polynomial: p(X) = det(X − λI). Now, since the determinant is invariant under elementary row and
column operations, it is invariant under transposition. Therefore det(X − λI) = det[T (X − λI)] = det[T (X) − λT (I)] =
det[T (X) − λI], i.e., p(X) = p[T (X)].
Is T also completely positive? To test this we need to check if any extension T p ≡ T ⊗ IR of T is also positive. This extension
is called the partial transpose, and its action on any basis element of B(HS ⊗ HR ) is as follows:
T p (∣i⟩⟨j∣ ⊗ ∣µ⟩⟨ν∣) = ∣j⟩⟨i∣ ⊗ ∣µ⟩⟨ν∣ . (127)
To prove that T is not a CP map, it suffices to find a counterexample. Indeed, consider the pure state ρ = ∣ψ⟩⟨ψ∣, where
∣ψ⟩ = √12 (∣0⟩S ∣0⟩R + ∣1⟩S ∣1⟩R ). Then:
1
T p (ρ) = (T ⊗ I)[∣0S 0R ⟩⟨0S 0R ∣ + ∣00⟩⟨11∣ + ∣11⟩⟨00∣ + ∣11⟩⟨11∣] (128a)
2
⎛1 0 0 0 ⎞
1 1⎜0 0 1 0 ⎟
= (∣00⟩⟨00∣ + ∣10⟩⟨01∣ + ∣01⟩⟨10∣ + ∣11⟩⟨11∣) = ⎜ ⎟.
2⎜ ⎟ (128b)
2 ⎜0 1 0 0 ⎟
⎝0 0 0 1 ⎠
The eigenvalues of this matrix are ( 12 , 12 , 12 , − 12 ), and the existence of a negative eigenvalue shows that T is not a CP map, since
T p >/ 0. Therefore T does not have a Kraus OSR, and is not a quantum map. Note furthermore that this means that a maximally
entangled two-qubit state has a negative partial transpose. This observation motivates the study of the partial transpose as a tool
for entanglement testing.
27
Consider a separable (thus by definition unentangled) state ρ = ∑i pi ρAi ⊗ ρi , where the pi are probabilities and the ρi and
B A
ρB
i are quantum states (positive, normalized). The state ρ obviously arises from the mixed state ensemble {ρi ⊗ ρi , pi }, in
A B
which every element is a tensor product state. Mixing such states classically does not generate any entanglement between A and
B, hence the definition.
Applying the partial transpose yields:
Since the transpose does not change the eigenvalues, σiA ≡ T (ρA i ) is also a valid quantum state, and hence T (ρ) is another
p
separable quantum state. In particular, this shows that every separable state has a positive partial transpose (PPT). In other
words, separability implies PPT. Conversely, a negative partial transpose (NPT) implies entanglement. This means that PPT is a
necessary condition for separability.
Is PPT also sufficient for separability? It turns out that this is the case only for the 2 × 2 (two qubits) or 2 × 3 (qubit and qutrit)
cases. I.e., only in these cases a state is separable iff it has a positive partial transpose (PPT) (conversely, is entangled iff it has a
NPT) [5, 6]. Indeed, we saw in the previous subsection that a (maximally) entangled state has NPT.
In higher dimensions the PPT criterion it is still necessary but no longer sufficient. In such higher dimensions there are
examples of so-called “bound-entangled” states that have PPT but are not separable [7].
As an example of the use of the PPT criterion consider the Werner states:
I
ρ = p∣Ψ− ⟩⟨Ψ− ∣ + (1 − p) (130)
4
√
where ∣Ψ− ⟩ is a maximally entangled singlet state: ∣Ψ− ⟩ = (∣01⟩ − ∣10⟩)/ 2. This represents a family of quantum states
parametrized by the probability p of being in the singlet state as opposed to the maximally mixed state.
Its density matrix in the standard basis is
⎛1 − p 0 0 0 ⎞
1⎜ 0 p + 1 −2p 0 ⎟
ρ= ⎜ ⎟,
4⎜ ⎟ (131)
⎜ 0 −2p p + 1 0 ⎟
⎝ 0 0 0 1 − p⎠
⎛1 − p 0 0 −2p ⎞
1 ⎜ 0 p+1 0 0 ⎟
T p (ρ) = ⎜⎜
⎟.
⎟ (132)
4⎜ 0 0 p+1 0 ⎟
⎝ −2p 0 0 1 − p⎠
The eigenvalues of this matrix are (1−3p)/4 and (threefold) (1+p)/4. Therefore, the state is entangled for p > 1/3 and separable
for p ≤ 1/3 (for p = 1/3 all eigenvalues are non-negative so PPT).
Φ = TrB ○ U ○ A, (133)
where (i) A is the “assignment map” which associates to every initial system state ρS (0) a fixed bath state ρB (0), i.e.,
A[ρS (0)] = ρS (0) ⊗ ρB (0); (ii) U is the unitary evolution superoperator, i.e., U[X] = U XU † ; (iii) TrB is the usual par-
tial trace operator. This is depicted in Fig. 2.
Let us show that each of these three maps is, in turn, CP.
28
A TrB
Φ
ρS (0) ρS (t)
FIG. 2. A commutative diagram showing that the quantum map Φ can be viewed as a composition of three maps.
The map A is from D(HS ) to D(HS ⊗ HB ). To prove that it is CP we need to consider positive operators A ∈ D(HS ⊗ HR ).
Thus, writing A = ∑ir λir ∣i⟩S ⟨i∣ ⊗ ∣r⟩R ⟨r∣ and ρB = ∑µ λµ ∣µ⟩B ⟨µ∣, with λir , λµ ≥ 0:
⟨v∣ (A ⊗ IR )(A) ∣v⟩ = ∑ λir ⟨v∣ [∣i⟩S ⟨i∣ ⊗ ρB ⊗ ∣r⟩R ⟨r∣] ∣v⟩ (134a)
ir
= ∑ λir λµ ⟨v∣ [∣i⟩S ⟨i∣ ⊗ ∣µ⟩B ⟨µ∣ ⊗ ∣r⟩R ⟨r∣] ∣v⟩ (134b)
irµ
This is obvious since the unitary map is a special case of a Kraus OSR having such a single Kraus operator U .
To demonstrate that the partial trace, TrB ∶ ρSB ↦ ρS , is CP, we can perform a direct calculation like we did for the assignment
map. However, instead we can also directly demonstrate that it has a Kraus OSR (since this is a sufficient condition for CPness).
Consider the following explicit Kraus operators for the partial trace map:
Kα = IS ⊗ ⟨α∣ , (135)
where {∣α⟩} denotes the elements of some chosen basis for the bath Hilbert space. This choice is motivated by the fact that the
partial trace leaves the system alone but sandwiches the bath between basis states.
Applying the map Φ = {Kα } to an arbitrary system-bath state ρSB = ∑ijµν λijµν ∣i⟩⟨j∣ ⊗ ∣µ⟩⟨ν∣ written in the same basis for
the bath, and noting that TrB (ρSB ) = ∑ijµνα λijµν ∣i⟩⟨j∣⟨α∣µ⟩⟨ν∣α⟩ = ∑ijα λijαα ∣i⟩⟨j∣, we find the following:
⎛ ⎞
Φ(ρSB ) = ∑ Kα ρSB Kα† = ∑ IS ⊗ ⟨α∣ ∑ λijµν ∣i⟩⟨j∣ ⊗ ∣µ⟩⟨ν∣ IS ⊗ ∣α⟩ = ∑ λijαα ∣i⟩⟨j∣ = TrB (ρSB ) , (136)
α α ⎝ijµν ⎠ ijα
as desired. Thus, the partial trace has Kraus elements as given in Eq. (135), and is CP.
What would happen if we were to relax the initial condition? Will we still get a CP map?
Using a general orthonormal basis for the joint Hilbert space we can always write
If we go through the same steps as in the derivation of the Kraus OSR, we have, with {∣µ⟩} now representing the same bath
basis as {∣α⟩}:
ρS (t) = ∑ ⟨µ∣U (t)ρ(0)U † (t)∣µ⟩ (139a)
µ
⎛ ⎞ ⎛ ⎞
= ∑ ⟨µ∣ U (t) ∣α⟩ ∑ λijαα ∣i⟩⟨j∣ ⟨α∣U † (t)∣µ⟩ + ∑ ⟨µ∣ U (t) ∣α⟩ ∑ λijαβ ∣i⟩⟨j∣ ⟨β∣U † (t)∣µ⟩ . (139d)
µ,α ⎝ ij ⎠ µij,α≠β ⎝ ij ⎠
The first summand in Eq. (139d) has Kraus operators ⟨µ∣ U (t) ∣α⟩ and may look fine. However, because of the sum over α we
cannot factor out ρS (0). Moreove, the second summand in Eq. (139d) in addition involves off-diagonal terms λijαβ that do not
appear in ρS (0) [Eq. (138)]. Clearly, we cannot factor out ρS (0), so we do not even get a map from ρS (0) to ρS (t).
2. Separable states
where ρiS and ρiB are themselves states of the system and bath? For such a state the initial system state is ρS (0) = TrB [ρ(0)] =
∑i pi ρiS . Let’s decompose each bath state as
ρiB = ∑ λνi ∣νi ⟩⟨νi ∣ (141)
νi
We can move the sum over i inside if we first assume that all ρiB commute, i.e., are diagonal in the same basis so that νi = ν ∀i,
for then ρiB (0) = ∑ν λiν ∣ν⟩⟨ν∣ and hence
ρS (t) = ∑ ⟨µ∣U (t) ∣ν⟩ ∑ λiν pi ρiS ⟨ν∣ U † (t)∣µ⟩ , (143)
µ,ν i
but this still doesn’t allow us to extract the initial system state ∑i pi ρiS . To accomplish this we may moreover assume that
eigenvalues are the same, i.e., λiν = λν ∀i. If we do so we find ρS (t) = ∑µ,ν λν ⟨µ∣U (t) ∣ν⟩ ∑i pi ρiS ⟨ν∣ U † (t)∣µ⟩, and this
involves a map acting on ρS (0) = ∑i pi ρiS as desired, but we haven’t gained anything: this is the case if ρiB = ρB ∀i, i.e., we’re
back to Eq. (109) again.
1. Quantum Discord
In classical information theory there are two equivalent ways to define the mutual information between two random variables
X and Y :
I(Y ∶ X) = H(Y ) + H(X) − H(X, Y ) (144a)
J(Y ∶ X) = H(Y ) − H(Y ∣X) , (144b)
30
where H(X) = − ∑i pi log(pi ) is the Shannon entropy associated with X, with pi = Pr(xi ) being the probability of X assuming
the value xi . The quantity H(X, Y ) is the entropy of the joint distribution, and H(Y ∣X) is the entropy of Y conditioned on
X. The equivalence follows directly from Bayes’ rule [the joint probability satisfies p(y, x) = p(y∣x)p(x), where p(y∣x) is the
conditional probability], which implies that H(X, Y ) = H(Y ∣X) + H(X), and hence that I(Y ∶ X) = J(Y ∶ X).
In the quantum case, measuring system X generally affects system Y if the joint state ρXY is correlated, so the asymmetry
inherent in the second expression J(Y ∣X) means that there is the potential for a different outcome from the symmetric first
expression I(Y ∣X). This observation forms the basis for the definition of the quantum discord, IQ (Y ∶ X) − JQ (Y ∶ X). Let
us thus define the quantum mutual information expressions IQ (Y ∶ X) and JQ (Y ∶ X).
First, we need the quantum von Neumann entropy associated with a state ρ:
Then
where ρXY is the total state of systems X and Y , ρY = TrX ρXY , and ρX = TrY ρXY . The second mutual information JQ
arises from first measuring X. Assume that this is done using a projective measurement with projectors {Πi }, acting only on
X. Then the post-measurement state obtained in case i is ρY ∣Πi ≡ Πi ρXY Πi /pi , where pi = Tr[Πi ρXY ] is the probability of
case i. Let us associate an entropy to this state: S(ρY ∣Πi ). The entropy conditioned non-selectively on the entire measurement is
S(Y ∣{Πi }) = ∑i pi S(ρY ∣Πi ), and the conditional entropy is the minimum over all possible measurements, since we’re interested
in maximizing the mutual information: S(Y ∣X) = min{Πi } S(Y ∣{Πi }). Explicitly:
With this, we are ready to define the second quantum mutual information:
D(ρXY ) = IQ (Y ∶ X) − JQ (Y ∶ X) . (149)
D(ρXY ) = 0 only for zero-discord states (by definition), which are states that have no quantum correlations at all. Note
that separable states can have non-zero discord [i.e., JQ (Y ∶ X) ≠ IQ (Y ∶ X)], which means that they have some quantum
correlations despite being a convex combination of product states. However, it is not hard to show that a special class of
separable states does have zero discord. Such states are known as zero-discord states, and they are of the form
where the Πi are projectors, i.e., Πi Πj = δij Πi . This initial state would be the result of a non-selective projective measurement
of the system with measurement operators {Πi } (you can easily check that the state is invariant under a non-selective projective
measurement with the same set of measurement operators, which is the property we expect from the state after a first projective
measurement; see subsection VI A for non-selective measurements).
It turns out that zero-discord states do allow us to generalize the assumption of a factorizable initial state [Eq. (109)] in the
derivation of the Kraus OSR [9]. Let’s assume that the initial state is of the form given in Eq. (150). Thus the system state
becomes
where we used the fact that ρiB is a positive operator to take its square root, and inserted a bath identity operator ∑ν ∣ν⟩⟨ν∣ in the
last line. Let’s define
√
Diµν ≡ ⟨µ∣ U (t) ρiB ∣ν⟩ , (152)
and note that this is a system-only operator. Now, we can always write Diµν = ∑m Dmµν δim . Inserting this into the last equation
we have
where we used Π2i = Πi . Next, note that δim Πi = Πm Πi and δin Πi = Πi Πn , which allows us to replace the δ’s by Π’s:
ρS (t) = ∑
†
pi Dmµν Πm Πi Πn Dnµν . (154)
µν,i,m,n
We can now move the sum over i inside so it is performed first. Thus, we have
and using Eq. (150) we recognize the middle term as the initial system state: ρS (0) = ∑i pi Πi . We can also define new Kraus
operators as Kµν = ∑m Dmµν Πm . This then gives us a proper Kraus OSR:
It turns out that there are also discordant states that give rise to CP maps, and even entangled states. Read about generalizations
in Refs. [10–12].
Given two quantum maps, a natural question is under which conditions they are equivalent. As an example, consider the
two single-qubit quantum maps defined by the following two sets of Kraus operators: Φ = {K0 = √12 I, K1 = √12 σ z } and
Ψ = {L0 = ∣0⟩⟨0∣, L1 = ∣1⟩⟨1∣}. Note that Φ can be interpreted as the map the flips the phase or leaves the state alone with equal
probability, while Ψ can be interpreted as a non-selective measurement in the σ z basis. Thus, a priori it seems that the two maps
describe very different physical processes. Nevertheless, it’s easy to show that the two maps are identical,3 i.e., ∀ρ
I I σz σz
√ ρ √ + √ ρ √ = ∣0⟩⟨0∣ρ∣0⟩⟨0∣ + ∣1⟩⟨1∣ρ∣1⟩⟨1∣. (157)
2 2 2 2
What is the general condition such that two maps are equivalent? The following theorem provides the answer:
Theorem 2. Consider the maps produced by the following two sets of Kraus operators Φ = {Kα }, Ψ = {Lβ }: ρ′ = ∑ Kα ρKα†
and ρ′′ = ∑ Lβ ρL†β . Then
3 Simply write ρ as a general 2 × 2 matrix and note that ZρZ flips the sign of the off-diagonal elements, so that both Φ and Ψ erase ρ’s off-diagonal elements.
32
Proof. Here we prove the “if” direction, i.e., assume that such a unitary exists; then
1 1
K0 = √ (L0 + L1 ) , K1 = √ (L0 − L1 ), (160)
2 2
1 1
so the unitary is u = √1 ( ).
2 1 −1
Where does this unitary equivalence between Kraus operators come from? To see this intuitively, note that in deriving the
Kraus operators, after evolving with a unitary operator acting on both the system and the bath, we trace out the bath, so the Kraus
operators should remain equivalent under the change of basis of the bath. Let us show that this “gauge freedom” gives rise to the
unitary equivalence between different sets of Kraus operators. As we shall see, we need to be a bit careful in accounting for the
presence of the square-root of the eigenvalue of the bath density matrix in the definition of the Kraus operators.
Let us write Eq. (158) as
where we have let α = (µν) and β = (ηξ). In terms of the explicit form of the Kraus operators this becomes
√ √
λν ⟨µ∣ U ∣ν⟩ = ∑ uµνηξ λξ ⟨η∣ U ∣ξ⟩ . (162)
ηξ
where v and w are both unitary. We can then show that the matrix u is unitary:
[u† u]αβ = ∑[u† ]αγ [u]γβ = ∑ u∗γα uγβ = ∑ u∗µ′ ν ′ µν uµ′ ν ′ ηξ = ∑ (vµ′ µ wνν ′ )∗ vµ′ η wξν ′ (164a)
γ γ µ′ ν ′ µ′ ν ′
i.e., the gauge freedom giving rise to the unitary equivalence between to sets of Kraus operators is:
Eq. (166a) simply expresses the freedom to apply a unitary transformation on the bath basis vectors before taking the partial
trace (which we did by sandwiching inside ⟨µ∣ ⋯ ∣µ⟩). Eq. (166b) tells us that we can also apply a second unitary
√ transformation
√
on the eigenstates of ρB (i.e., ∣ν⟩ ↦ w ∣ν⟩), but that in general we should also replace the eigenvalue term λν by ρB . To
understand the latter, note that in deriving the Kraus OSR we can also proceed as follows:
√ √
ρ′S = TrB [U ρS ⊗ ρB U † ] = ∑ ⟨µ∣ [U ρS ⊗ ( ρB ∑ ∣ν⟩⟨ν∣ ρB ) U † ] ∣µ⟩ (167a)
µ ν
√ √
= ∑ (⟨µ∣ U ρB ∣ν⟩) ρS (⟨ν∣ ρB U † ∣µ⟩) , (167b)
µν
√
which means that the Kraus operators we derived originally by using ρB ’s spectral decomposition, ⟨µ∣ U λν ∣ν⟩, are equivalent
√
to Kraus operators of the form ⟨µ∣ U ρB ∣ν⟩. In other words, the spectral decomposition was just one of infinitely many
equivalent ways to decompose ρB . We recover the spectral decomposition if we choose the basis {∣ν⟩} in Eq. (167) as the
eigenbasis of ρB .
In this section, by focusing on the case of one qubit, we will develop a geometric picture of the action of quantum maps. The
main tool that will allow us to do this is the Bloch sphere representation.
Recall that the density matrix of a qubit may be written as ρ = 12 (I + v⃗ ⋅ σ
⃗ ) where σ
⃗ = (σx , σy , σz ) and v⃗ = (vx , vy , vz ) ∈ R3 is
the Bloch vector. In this way, a single-qubit state may be thought of as a point in or on the unit sphere in R3 —the Bloch sphere.
States with ∥⃗v ∥ = 1 lie on the surface of the sphere and correspond to pure states of the form ρ = ∣ψ⟩⟨ψ∣. Points on the interior of
the sphere correspond to mixed states with purity P = Tr[ρ2 ] < 1.
What happens when a quantum map acts on a single qubit? As a map of the density matrix, Φ ∶ ρ ↦ ρ′ . At the same time
ρ must be expressible in terms of a new Bloch vector v⃗′ , where ρ′ = 12 (I + v⃗′ ⋅ σ
′
⃗ ). We shall show that ρ ↦ ρ′ is equivalent to
mapping the Bloch vector
v⃗ ↦ v⃗′ = M v⃗ + c⃗ (168)
for some real 3 × 3 matrix M and a vector c⃗ ∈ R3 . This is an affine transformation. Before proving Eq. (168), let us decompose
M in a way that will reveal more of the geometric aspects of the transformation.
√ the polar decomposition, which allows us to write any square matrix A as A = U ∣A∣, where U is a unitary matrix and
Recall
∣A∣ ≡ A† A is Hermitian (since clearly its eigenvalues are real), a generalization of the polar representation of a complex number
z = eiθ ∣z∣. If A is a real matrix, U becomes real-unitary, i.e., orthogonal, and ∣A∣ becomes real-Hermitian,
√ i.e., symmetric. So,
for our 3 × 3 real matrix M we can write M = OS, for orthogonal O and symmetric S = M † M . S causes deformation by
scaling along the directions of the eigenvectors by a factor of the corresponding eigenvalues. O is a rotation matrix. Now we
may interpret the action of a quantum map on a qubit state as mapping the Bloch vector according to
v⃗ ↦ v⃗′ = OS⃗
v + c⃗, (169)
as a shift by c⃗, a deformation by S and a rotation by O. Because the Bloch sphere represents the set of possible Bloch vectors,
we may view the Kraus map acting on a qubit as a transformation of the Bloch sphere that displaces its center by c⃗ and turns the
sphere into an angled ellipsoid.
To prove Eq. (168), we plug the Bloch vector representation of ρ into the quantum map:
1 1
ρ′ = ∑ Kα ρKα† = ∑ Kα (I + v⃗ ⋅ σ
⃗ )Kα† = (∑ Kα Kα† + ∑ vj Kα σj Kα† ) . (170)
α 2 α 2 α αj
34
To isolate the components of v⃗′ we multiply both sides by σi and take the trace, while remembering that the Pauli matrices are
all traceless and satisfy Eq. (A22). Thus, Eq. (170) becomes
1
Tr(ρ′ σi ) = [∑ Tr(Kα Kα† σi ) + ∑ vj Tr(Kα σj Kα† σi )] , (171)
2 α αj
1
Tr(ρ′ σi ) = [Tr(σi ) + ∑ vj′ Tr(σj σi )] = 0 + vi′ . (172)
2 j
where
1
Mij = ∑ Tr(σi Kα σj Kα )
†
(174a)
2 α
1
ci = ∑ Tr(σi Kα Kα† ) . (174b)
2 α
Moreover, using the Hermiticity of the Pauli matrices and properties of the trace [Eq. (A19)]:
∗ 1 1 1
Mij = ∑ Tr(σi Kα σj Kα ) = ∑ Tr(Kα σj Kα σi ) = ∑ Tr(σi Kα σj Kα ) = Mij ,
† † † †
(175)
2 α 2 α 2 α
1 1 1
c∗i = ∑ Tr(σi Kα Kα ) = ∑ Tr(Kα Kα σi ) = ∑ Tr(σi Kα Kα ) = ci ,
† † † †
(176)
2 α 2 α 2 α
Returning temporarily to the general (beyond a single qubit) case, a quantum map is said to be unital if it maps the identity
operator to itself, i.e.:
Since a quantum map always has a Kraus OSR, we find that unital quantum maps satisfy
∑ Kα Kα = I ,
†
(177)
α
1 1
ci = Tr(σi ∑ Kα Kα† ) = Tr(σi ) = 0 (unital case) . (178)
2 α 2
⃗
Conversely, if Φ is non-unital, then c⃗ ≠ 0.
Note that, as is clear from Eq. (172), M is associated purely with the transformation of v⃗⋅⃗
σ under the map, while c⃗ is associated
purely with the transformation of I under the map. This observation will help us read off M and c⃗ in the examples we study
below.
35
[in agreement with the fact that the phase damping map is unital; recall Eq. (178)], and:
1 1 1
Mij = ∑ Tr(σi Kα σj Kα ) = [pTr(σi σj ) + (1 − p)Tr(σi Zσj Z)] = pδij + (1 − p)Jij ,
†
(182)
2 α 2 2
where Jij ≡ Tr(σi Zσj Z). Written explicitly the matrix J is:
⎛2p − 1 0 0⎞
⎜
M = diag[p − (1 − p), p − (1 − p), p + (1 − p)] = ⎜ 0 2p − 1 0⎟
⎟, (184)
⎝ 0 0 1⎠
and
The corresponding transformation of the Bloch sphere is shown in Fig. 3. There is no shift of the Bloch sphere, while there is a
rescaling along the vx and vy directions by a factor of (2p − 1), and all points on the vz axis are fixed. The map has two fixed
pure states, the north and south poles of the Bloch sphere, ∣0⟩⟨0∣ and ∣1⟩⟨1∣. For p = 1, the Bloch sphere remains unchanged.
Because p is a probability, −1 ≤ 2p − 1 ≤ 1. Hence the scaling factor can take negative values,
√ corresponding to a rotation by π
about the vz axis. To see why, let us use the polar decomposition to write M = OS, where S = M † M = diag(∣2p−1∣, ∣2p−1∣, 1).
Therefore the rotation matrix must be O = diag(sign(2p − 1), sign(2p − 1), 1) = (±1, ±1, 1). When 2p − 1 < 0, O is a rotation by
π about the vz axis.
The purity [Eq. (64)] of the transformed state is
1 1
P ′ = Tr[(ρ′ )2 ] = (1 + ∥⃗
v ′ ∥2 ) = [1 + (2p − 1)2 (vx2 + vy2 ) + vz2 ] ≤ P . (186)
2 2
Thus the purity always decreases under the phase damping channel, except for the states on the vz axis (with vx = vy = 0), whose
purity is invariant.
0.5
pauli Z 0
−0.5
−1
1
0.5 1
0 0.5
0
−0.5
−0.5
pauli Y −1 −1
pauli X
FIG. 3. The Bloch sphere become an ellipsoid after transformation by the phase damping channel. The invariant states are those on the σz
axis. The major axis has length 2, the minor axis has length 2(2p − 1).
In the computational basis, the bit flip map acts like a classical error channel, flipping bits at random. The phase damping map is
purely quantum in the same basis, since of course the notion of a phase is not classical. However, mathematically the two maps
are essentially identical. We can guess that since the phase flip map leaves the vz axis alone and shrinks the Bloch sphere in the
(vx , vy ) plane, the bit flip map will leave vx axis alone and shrinks the Bloch sphere in the (vy , vz ) plane. To confirm this, let
us use a more direct approach than the one we used for the phase flip map.
Using ρ = 21 (I + v⃗ ⋅ σ
⃗ ), we have:
1
ρ ↦ ρ′ = pρ + (1 − p)XρX = (I + p⃗ ⃗ + (1 − p)X v⃗ ⋅ σ
v⋅σ ⃗ X) . (188)
2
The key point is now that
X(⃗ ⃗ )X = X(vx X + vy Y + vz Z)X = vx X − vy Y − vz Z .
v⋅σ (189)
This shows that vx is unchanged, but the sign of both vy and vz is flipped. Had we studied the phase damping map instead, we
would have seen that vz is unchanged, but the sign of both vx and vy is flipped. We now have:
1 1
ρ′ = (I + vx X + (2p − 1)vy Y + (2p − 1)vz Z) = (I + v⃗′ ⋅ σ
⃗) . (190)
2 2
Thus, we find that the bit flip channel transforms v⃗ as:
v⃗ ↦ v⃗′ = [vx , (2p − 1)vy , (2p − 1)vz ] = M v⃗ + c⃗ , (191)
where
⎛1 0 0 ⎞
⎜
M = ⎜0 2p − 1 0 ⎟ ⎟ (192a)
⎝0 0 2p − 1⎠
c⃗ = 0̄ . (192b)
37
Geometrically, this corresponds to the exact same deformation of the Bloch sphere as depicted in Fig. 3, but with the vx and vz
axes interchanged. If we replace X with Y in Eq. (187) we have the “bit-phase flip channel,” where the roles of the vx and vy
axes is interchanged.
The depolarizing map acting on a qubit either takes the state to the maximally mixed state with probability p, or leaves the
state unchanged with probability 1 − p:
⎧
⎪
⎪1I w/ prob. p
ρ ↦ ρ′ = ⎨ 2 . (193)
⎪
⎪ w/ prob. 1 − p
⎩ρ
Thus, with probability p, all the information held in the state is erased. Equivalently,
I
ρ′ = p + (1 − p)ρ . (194)
2
Clearly, this is also a unital map. However, note that it is not in Kraus OSR form. To put it in Kraus OSR form, note that
which we can prove easily using the same idea as in Eq. (189):
Y (⃗ ⃗ )Y = −vx X + vy Y − vz Z
v⋅σ (196a)
Z(⃗ ⃗ )Z = −vx X − vy Y + vz Z ,
v⋅σ (196b)
so that
∑ σi (⃗ ⃗ )σi = 0 .
v⋅σ (197)
i
M = (1 − p)I (201a)
c⃗ = 0̄ . (201b)
This corresponds to the Bloch sphere shrinking uniformly to a radius of 1 − p, as illustrated in Fig. 4. The only invariant state is
the fully mixed state (the origin, v⃗ = 0̄). Every other state loses purity as it becomes more mixed.
Spontaneous emission (SE) is the process by which an atom, nucleus, etc., undergoes a transition from a higher state of energy
to a lower state of energy, thus releasing energy to the bath (relaxation). This could through the release of a photon, a phonon,
or some other elementary excitation. If the bath is at temperature T = 0, as we assume in this subsection, then the system cannot
absorb energy, so the reverse process of excitation does not occur. We shall deal with it in the next subsection.
38
0.5
pauli Z 0
−0.5
−1
1
0.5 1
0 0.5
0
−0.5
−0.5
pauli Y −1 −1
pauli X
FIG. 4. The Bloch sphere transformed by the depolarizing channel. As p → 1, all states converge to the fully mixed state at the origin.
We consider a single qubits, with a ground state ∣0⟩ and an excited state ∣1⟩. Thus the map Φ is:
∣0⟩ ↦ ∣0⟩ with probability 1 (202a)
∣1⟩ ↦ ∣0⟩ with probability p (202b)
Let us find the Kraus operators for this process. One Kraus operator is obvious: the transition from the excited state to the ground
state is given by
√
K1 = p∣0⟩⟨1∣ . (203)
The second Kraus operator should keep the ground state in place, i.e., contains ∣0⟩⟨0∣. But this isn’t enough, since the normal-
ization condition must be satisfied, and it’s easy to check that it isn’t if these are our Kraus operators. Instead, let us add an
unspecified matrix A and find out its form from the normalization condition. Thus:
1 a
K0 = ∣0⟩⟨0∣ + A = ( ), (204)
b c
1 + ∣b∣2 a + b∗ c 1 0
( ∗ ) + p∣1⟩⟨0∣0⟩⟨1∣ = ( ). (205)
a + bc∗ ∣a∣2 + ∣c∣2 0 1
On equating the upper √left entries we get b = 0, which implies from the off-diagonal entries that a = 0. Equating the bottom right
entries then yields c = 1 − p. Thus:
1 0
K0 = ( √ ). (206)
0 1−p
√
The (perhaps curious) 1 − p component expresses the fact that not observing an emission event (imagine a detector for the
emitted photons) increases the likelihood that the system is in its ground state, but we cannot know this with certainty since
39
the emission event might yet arrive in the future. We will see this more clearly later when we discuss quantum trajectories in
Sec. XII.
We can now directly derive M and c⃗. Since ρ′ = ∑1α=0 Kα [ 21 (I + v⃗ ⋅ σ
⃗ )] Kα† , the most direct way to do this is to map I and
v⃗ ⋅ σ
⃗ via the Kraus OSR and read off M and c⃗. Starting with I, we have:
1 0 1+p 0
I ↦ K0 K0† + K1 K1† = ( ) + p∣0⟩⟨1∣1⟩⟨0∣ = ( ) = I + pZ. (207)
0 1−p 0 1−p
Thus SE is not a unital map. Since c⃗ captures the mapping of I, we see that
c⃗ = (0, 0, p) . (208)
Next, v⃗ ⋅ σ
⃗ ↦ K0 (⃗ ⃗ )K0† + K1 (⃗
v⋅σ ⃗ )K1† . It is simple to check by explicit matrix multiplication that
v⋅σ
√
K0 XK0† + K1 XK1† = 1 − pX (209a)
√
K0 Y K0† + K1 Y K1† = 1 − pY (209b)
K0 ZK0† + K1 ZK1† = (1 − p)Z (209c)
The geometric meaning of the spontaneous emission map is now clear. The center (0, 0, 0) ↦ (0, 0, p), and the Bloch sphere
is compressed more along the vz -axis than along the vx and vy -axes. In other words, all points on the Bloch sphere move closer
to its north pole, which is the ground state. If p = 1 then the entire Bloch sphere is compressed to a single point, the north pole.
The latter is a fixed point of the map. To see this, note that
If the qubit is able to absorb energy from the bath (since the latter is at a temperature T > 0), then the reverse process, of
excitation from the ground state to the excited state, is also possible. To account for this let us assume that the spontaneous
emission process of the previous subsection occurs with probability q, while the reverse process occurs with probability 1 − q.
Then the Kraus operators for the SE event become
√ 1 0
K0 = q( √ ) (212a)
0 1−p
√
K1 = qp∣0⟩⟨1∣ . (212b)
Thus:
1+p 0 1−p 0
I ↦ K0 K0† + K1 K1† + K2 K2† + K3 K3† = q ( ) + (1 − q) [( ) + p∣1⟩⟨0∣0⟩⟨1∣] (214a)
0 1−p 0 1
= q(I + pZ) + (1 − q)(I − pZ) = I + (2q − 1)pZ , (214b)
As for the M matrix it is again simple to check by explicit matrix multiplication that
3 √
∑ Ki XKi = 1 − pX
†
(216a)
i=0
3 √
∑ Ki Y Ki = 1 − pY
†
(216b)
i=0
3
∑ Ki ZKi = (1 − p)Z ,
†
(216c)
i=0
3
Φ(ρeq ) = ∑ Ki ρeq Ki† = ρeq . (217)
i=0
Note that the case q = 1/2 is unital (it corresponds to c⃗ = ⃗0) and has a fixed point the fully mixed state. Also note that when
q < 1/2 the new center is at (0, 0, −∣2q − 1∣p), which corresponds to a preference for the excited state rather than the ground state.
So far we postulated the form of certain quantum maps. Let us now consider examples where we can analytically derive the
Kraus operators from first principles.
Consider a system of two qubits, such that the first qubit is the system (HS ) and the second is the bath (HB ). Consider also the
interaction Hamiltonian HSB = λσSα ⊗ σB β
where α, β ∈ {x, y, z}. The system qubit is initially in the pure state ρS (0) = ∣ψ⟩⟨ψ∣,
∣ψ⟩ = a ∣0⟩ + b ∣1⟩, written in the computational basis (eigenbasis of σ z ). The initial state of the bath is mixed:
λ0 0
ρB (0) = λ0 ∣0⟩⟨0∣ + λ1 ∣1⟩⟨1∣ = [ ], (218)
0 λ1
Let θ ≡ λt. Recall now that if A2 = I then eiθA = cos θI + i sin θA (which can be easily checked by Taylor expansion). Therefore
β
eiθσS ⊗σB = cos θ ⋅ IS ⊗ IB + i sin θ ⋅ σSα ⊗ σB
α
β
, (220)
1. Z ⊗ X coupling
Consider first HSB = λZS ⊗ XB . In this case, we can use Eq. (221) to find
√
K00 = λ0 cos θ ⋅ I (223a)
√
K11 = λ1 cos θ ⋅ I (223b)
√
K01 = −i λ1 sin θ ⋅ σ z (223c)
√
K10 = −i λ0 sin θ ⋅ σ z (223d)
The density matrix for this map evolves under the action of these 4 Kraus operators:
which we would like to equate with Eq. (224d). Clearly, this requires 2p − 1 = f (θ), so that
1 + f (θ)
p= , (226)
2
and the phase damping map has as a physical origin the model given by HSB = λZS ⊗ XB .
Why did λ0 and λ1 drop out? The intuitive reason is that by having the bath qubit subject to σ x , its ∣0⟩ and ∣1⟩ state are
constantly flipped, which also interchanges λ0 and λ1 , so it is as if they are averaged to 1/2.
There is much more to say about this result, but first let us consider another case, which will turn out to subsume this one.
2. Z ⊗ Z coupling
Altogether, the pure state ∣ψ⟩ under each of these operators becomes
√ e−iθ 0 a √ ae−iθ
K00 ∣ψ⟩ = λ0 [ iθ
] [ ] = λ0 [ iθ ] (228a)
0 e b be
√ aeiθ
K11 ∣ψ⟩ = λ1 [ ] (228b)
be−iθ
42
Therefore:
ae−iθ aeiθ
] [ a∗ eiθ b∗ e−iθ ] + λ1 [ −iθ ] [ a∗ e−iθ b∗ eiθ ]
ρS (t) = λ0 [
iθ
be be
⎡ ∣a∣2 ab∗ (λ0 e−2iθ + λ1 e2iθ ) ⎤⎥
⎢
=⎢ ∗ ⎥ (229)
⎢ a b (λ0 e + λ1 e
2iθ −2iθ
) ∣b∣2 ⎥
⎣ ⎦
where the diagonal elements have again been simplified with the use of the fact that λ0 + λ1 = 1. Defining
f (θ) = λ0 e−2iθ + λ1 e2iθ , (230)
yields
∣a∣2 ab∗ f (θ)
ρS (t) = [ ∗ ∗
]. (231)
a bf (θ) ∣b∣2
The previous example, HSB = Z ⊗X, is now seen to be a special case of this one, where λ0 = λ1 = 1/2 [for then f (θ) = cos(2θ)],
so everything we discuss next applies to it as well.
Note that the diagonal elements (“population”) haven’t changed under time evolution and yet the off-diagonal elements (“co-
herence”) are modulated by the periodic function f . This is like elastic scattering where no energy is exchanged and only relative
phases are impacted. More precisely, this is a dephasing process, although in our case, the phase coherence recurs periodically.
The period of f is τ = π/λ.
Consider the purity P = Tr(ρ2 ):4
⎡ 2⎤
⎢ ∣a∣2 ab∗ f (θ) ⎥⎥
P = Tr ⎢⎢( ∗ ∗ ) ⎥ (232a)
⎢ a bf (θ) ∣b∣2 ⎥
⎣ ⎦
∣a∣4 + ∣a∣2 ∣b∣2 ∣f ∣2 ...
= Tr ( ) (232b)
... ∣b∣ + ∣a∣2 ∣b∣2 ∣f ∣2
4
The purity achieves a minimum of 1/2 when the bath qubit is in √ a maximally mixed state, ρB = IB /2 (so that λ0 = λ1 = 1/2),
and when the system qubit is an equal superposition, ∣a∣ = ∣b∣ = 1/ 2.5
For short times t ≪ τP the purity decays quadratically. This is typical of non-Markovian decay, as we will see later (in
contrast, Markovian decay is always exponential, i.e., it starts out linearly). One might also write this inequality as a weak
coupling limit λ ≪ π/t, which suggests that in this limit the purity appears to be only decaying (i.e., there is no time for a
recurrence). However, if the coupling between the system and the bath is strong, that is λ ≫ 1, then we may not necessarily
resolve the oscillations in purity and instead measure an average purity significantly lower than 1. In both these limits an observer
would conclude that the state of the system is mixed, even though it started out pure.
Can we relate this model to the phase damping map discussed in Sec. VII C? Clearly, this requires 2p − 1 = f (θ) = f ∗ (θ).
Thus, equality only holds subject to the additional constraint that f (θ) is real. The constraint that f (θ) = f ∗ (θ) requires that
λ1 = λ2 = 1/2, i.e., the initial bath state is I/2. Therefore this Hamiltonian model is more general than the phase damping map.
This is because the bath operator ZB in the former does not affect the bath state ρB (0) = diag(λ0 , λ1 ), with which it commutes.
This keeps λ0 and λ1 in play, unlike the previous case where they were averaged out.
4 We can obtain the same result using the formula P = 21 (1+∥⃗v ∥2 ), as follows: 12 (vx −ivy ) = ab∗ f and 21 (vx +ivy ) = a∗ bf ∗ , so that vx2 +vy2 = 4∣a∣2 ∣b∣2 ∣f ∣2 .
Also, vz = ∣a∣ − ∣b∣ , and 1 = (∣a∣ + ∣b∣ ) . Adding all this up gives Eq. (232c).
2 2 2 2 2
√ note that ∣a∣ + ∣b∣ = ∣a∣ + (1 − ∣a∣ ) = 2∣a∣ − 2∣a∣ + 1 = 2x − 2x + 1 with x = ∣a∣ ; this is minimized at 4x − 2 = 0, i.e., x = 1/2, or
5 To see this 4 4 4 2 2 4 2 2 2
∣a∣ = 1/ 2.
43
B. Irreversible open system dynamics: infinite dimensional bath with a continuous density of states
Our previous example involved a finite-dimensional bath, and we saw that the purity in this case is periodic. This reflect
reversibility, which is a general characteristic of the finite dimensional case. To exhibit irreversibility we shall investigate an
infinite-dimensional bath, but as we shall see, one additional ingredient (a continuum) will be needed as well.
Assume the system is either a qubit or a quantum harmonic oscillator (QHO). We will work out both cases. The system-bath
interaction Hamiltonian has one of the following forms:
where n̂S is the number operator satisfying n̂S ∣n⟩ = n ∣n⟩ for n = 0, 1, ..., ∞. The total Hamiltonian is H = HSB + HB , where
we have set HS = 0 for simplicity. We assume that the bath is itself a QHO with Hamiltonian
∞
HB = ∑ Eν ∣ν⟩⟨ν∣ , (236)
ν=0
̵ ≡ 1).
where n̂B ∣ν⟩ = ν ∣ν⟩ and n̂B is the number operator, and Eν are QHO energies: Eν = ω(ν + 12 ) (where as before we set h
We assume that the initial state of the bath is a Gibbs state:
1 −βHB 1 ∞ −βEν ∞
ρB (0) = e = ∑e ∣ν⟩⟨ν∣ ≡ ∑ λν ∣ν⟩⟨ν∣ , (237)
Z Z ν=0 ν=0
where β = kB1T and λν = Z1 e−βEν are the eigenvalues of the bath density matrix. The denominator is the partition function:
Z = Tr[e−βHB ] = ∑∞ν=0 e
−βEν
.
Using the Hamiltonians in Eq. (235) and again defining θ ≡ λt, the joint unitary evolution operator becomes
where we expanded the density matrix in the eigenvectors of the σ z or n̂ operator, with the upper limits being 1 or ∞, respectively.
In the case where the system is a qubit, we have, using Eq. (112):
1
ρS (t) = ∑ rmn ∑ λν e−iθνσ ∣m⟩⟨n∣eiθνσ
z z
(243a)
m,n=0 ν
1
= ∑ rmn ∑ λν e−iθν(−1) ∣m⟩⟨n∣eiθν(−1) .
m n
(243b)
m,n=0 ν
where
∞ n
−(−1)m ]νθ
gn,m (θ) ≡ ∑ λν ei[(−1) . (245)
ν=0
The diagonal terms gm,m = ∑∞ ν=0 λν = 1 are constant, and therefore they do not evolve in this case. Let us focus next on the case
in which both system and bath are QHO’s. We then have, using Eq. (112):
where
Note that f0 (θ) = ∑ν λν = 1, so the state of the system at time t can be split into diagonal (population) and off-diagonal
(coherence) terms:
ρS (t) = ∑ rnn ∣n⟩⟨n∣ + ∑ rmn ∣m⟩⟨n∣fn−m (λt) (249)
n m≠n
The population term is time-independent, i.e., is the same as in ρS (0). The coherence term is time-dependent and is affected by
the coupling to the bath. Its behavior is completely determined by the modulation function f , which can be computed explicitly
by performing the geometric sum:
1 ∞ e− 2 βω ∞ −βων ixνθ
1
= , (250c)
Z 1−q
where convergence of the infinite series is guaranteed since ∣q∣ = e−βω < 1 due to βω > 0.
Note that fx (θ) = fx (θ + 2π/x), i.e., f is periodic, with period T (x) = 2π/(λx). Each off-diagonal element ∣m⟩⟨n∣ in
Eq. (249) thus has a different period τmn = 2π/(λ∣m − n∣). This suggests that we might have an example of irreversible
45
decoherence [decay of the off-diagonal elements of ρS (t)], if ρS (t) isn’t periodic. But is it? Periodicity requires there to be a
time τ̃ that is simultaneously divisible by all periods τmn (i.e., all such periods fit an integer number of times into τ̃ ). Clearly,
τ̃ = 2π/λ is just such a time: τ̃ /τmn = ∣m − n∣. Thus ρS (t) is periodic after all, with a period of 2π/λ, and we do not have
irreversibility.
Note that the qubit-system case is just a special case of the QHO-system. To see this observe that g00 = g11 = ∑ν λν = 1, and
∗
g01 = g10 = ∑ν λν e2iθν = f2 (θ).
To better understand the emergence of irreversibility, we thus consider a modified model, where we introduce a mode density
Ω(ν) (a standard trick in condensed matter physics; consider, e.g., the Debye model). We thus replace the sum by an integral
over ν, and write
1 ∞
dνe−βω(ν+ 2 ) eixθν Ω(ν).
1
fx (θ) = ∫ (251)
Z 0
If Ω(ν) = ∑∞ ′
ν ′ =0 δ(ν − ν ) then we recover Eq. (250a). The modified model has the following mode density:
Ω0 if νc ≥ ν ≥ 0;
Ω(ν) = { , (252)
0 otherwise.
i.e., it has a continuous set of modes with a high-mode cutoff of νc . The cutoff is physically well-motivated: it reflects the fact
that any physical model must have a highest but finite accessible energy. Then:
Ω0 νc
e−βω(ν+ 2 ) eixθν dν
1
fx (θ) = ∫ (253a)
Z 0
Ω0 e− 2 βω
1
νc
= ∫ e−(βω−ixθ)ν dν (253b)
Z 0
Ω0 e− 2 βω e−(βω−ixθ)νc − 1
1
= (253c)
Z −βω + ixθ
The numerator is periodic just like in the previous case, so the same comments apply. However, the denominator contains a
(n − m)λt dependence (the xθ term), which shows that the coherences decay irreversibly as 1/t, with the decay being faster for
off-diagonal elements that are farther apart.
We have thus seen how an infinite-dimensional bath with a continuous mode density can result in a decay which is truly
irreversible. The decay of the off-diagonal elements is often called decoherence, since it refers to the gradual disappearance of
coherence, the name given to the off-diagonal elements. This is not an entirely satisfactory definition of decoherence, since it is
obviously basis dependent. We shall give a more careful definition later.
IX. DERIVATION OF THE LINDBLAD EQUATION FROM A SHORT TIME EXPANSION OF QUANTUM MAPS
Just as the Hamiltonian is the generator of unitary evolution, we may ask if there is a generator for open system dynamics. By
this we mean that the solution of the differential equation ρ̇ = Lρ is a quantum map, and L plays the role of a generator. In this
section we will see how to find such a generator for very short evolution times using just a short time expansion of the Kraus
OSR. We will then postulate that the same generator applies for all times (a type of Markovian approximation), and thus arrive
at a “master equation” of the form ρ̇ = Lρ that generates a quantum map. The generator L is called the Lindbladian, and the
master equation is the Lindblad equation, whose special form guarantees complete positivity (i.e., that the evolution it generates
is a quantum map).
A. Derivation
Let’s try to find the Kraus operators that make these two equations agree up to O(dt). Clearly, to get the ρ(0) term in Eq. (254)
one of the Kraus operators must contain the identity operator. Thus, let us write
K0 = I + L0 dt , (256)
so that
This contributes one term of order dt, but there must be more (since as we know a Kraus OSR with a single Kraus operator is
equivalent to unitary evolution). Thus, we can pick all other Kraus operators as
√
Kα = dtLα , α ≥ 1 , (258)
so that
Without loss of generality we can decompose the general operator L0 into a Hermitian and anti-Hermitian part: L0 = A − iH,
with A = A† and H = H † . Thus, Eq. (260) tells us that to O(dt):
1
A=− ∑ L Lα .
†
(261)
2 α≥1 α
Plugging all this back into the Kraus OSR, Eq. (255), we find:
Therefore:
ρ(dt) − ρ(0) 1
ρ̇(t)∣0 = lim = −i[H, ρ(0)] + ∑ (Lα ρ(0)L†α − {L†α Lα , ρ(0)}) . (263)
dt→0 dt α≥1 2
This is almost
√ the form of the master equation we are after. Note that the √ operators Lα are not dimensionless, but must have
units of 1/ time. To make them dimensionless, let us replace them by γα′ Lα , where γα′ has units of 1/time, so that the new
√ √
Lα are dimensionless. Substituting this into Eq. (263) only generates the combinations γα′ γα′∗ = ∣γα′ ∣ ≡ γα ≥ 0. Thus:
1 †
ρ̇(t)∣0 = −i[H, ρ(0)] + ∑ γα (Lα ρ(0)L†α − {L Lα , ρ(0)}) . (264)
α≥1 2 α
This result is valid as a short time expansion near t = 0. We now make an extra, very significant assumption:
Assumption 1. Eq. (264) is valid for all times t > 0.
This is essentially the Markovian limit, which states (informally) that there is no memory in the evolution, as manifested by
the fact that the evolution “resets” every dt. It is motivated in part by the observation that if we limit our attention just to ρ̇(t)∣0 =
−i[H, ρ(0)], then we already know that this replacement is valid, i.e., that we can indeed replace this with ρ̇(t) = −i[H, ρ(t)]
for all t, since this is just the Schrödinger equation. With this we finally arrive at the Lindblad equation:
dρ 1
= −i[H, ρ(t)] + ∑ γα (Lα ρ(t)L†α − {L†α Lα , ρ(t)}) ≡ Lρ . (265)
dt α 2
47
The generator of the evolution, L, is called the Lindbladian. The Lα are called the Lindblad operators. The operator H is
Hermitian and will be interpreted later as the Hamiltonian of the system (plus a correction called the Lamb shift). The form of
the dissipative part of the Lindbladian, also known as the dissipator, is:
1
LD [⋅] = ∑ γα (Lα ⋅ L†α − {L†α Lα , ⋅}) , γα ≥ 0 . (266)
α 2
We can now define decoherence in a basis-independent manner. Decoherence is what happens when LD ≠ 0. In this case the
evolution of the density matrix is governed not only by the Schrödinger component −i[H, ⋅] (responsible for unitary evolution),
but also by the dissipator, which gives rise to non-unitary evolution.
The positivity of the Lindblad rates (they have units of 1/time) is a direct consequence of complete positivity. Conversely, it
guarantees that the map generated by the Lindblad equation (265) is CP, as we will show in Sec. IX C. As our derivation shows,
this map has Kraus operators given by
where Λt is called the Markovian evolution operator (it is also a quantum map). The set {Λt }t≥0 forms a one-parameter
semigroup. The one-parameter part is clear: the set depends only on the time t, once the Lindblad generator L is fixed. The
reason this is a semi-group is that the superoperators Λt only satisfy three of the four properties of a group:
1. Identity operator: Λ0 = I.
2. Closed under multiplication: Λt Λs = eLt eLs = eL(t+s) = Λt+s .
3. Associative: (Λt Λs )Λr = Λt (Λs Λr ).
However, not every element has an inverse: as we shall see, complete positivity forces all the eigenvalues of L to be non-positive,
so that the map Λt is contractive, corresponding to exponential decay. This means that Λ∞ has at least one zero eigenvalue, so
it does not possess an inverse. We shall shortly see this in examples.
The argument we use to prove that the solution of the Lindblad Equation is a CP map is essentially the reverse of that presented
in Sec. IX A, plus a proof that the concatenation of CP maps (and in particular of a CP map with itself) is still a CP map.
Let us start from the Lindblad equation and let A ≡ − 21 ∑α≥1 L†α Lα :
1 †
ρ̇(t) = Lρ(t) = −i[H, ρ(t)] + ∑ (Lα ρ(t)L†α − {L Lα , ρ(t)}) (269a)
α≥1 2 α
1 †
= −i[H, ρ(t)] + ∑ (Lα ρ(t)L†α − {L Lα , ρ(t)}) (269b)
α≥1 2 α
= −i[H, ρ(t)] + {A, ρ(t)} + ∑ Lα ρ(t)L†α (269c)
α≥1
Now define K0 ≡ [I + (A − iH)dt] + O[(dt)2 ], where in the end we will take the dt → 0 limit to remove any residual O[(dt)2 ]
terms. Then:
which is in Kraus OSR form. However, to prove that this is a valid quantum map we still need to show that the set {Kα }α≥0
satisfies the normalization condition. Indeed, we have:
where in the first equality we used the Hermiticity of A and H, and in the second equality we used the definition of A.
Thus, we have shown that in the dt → 0 limit the map Φ [Eq. (271d)] is a quantum map from ρ(t) to ρ(t + dt). Let
dt = limn→∞ t/n, and consider the concatenated sequence of maps limn→∞ Φ○n [ρ(0)] = Φ[Φ[⋯Φ[ρ(0)]]] = Λt ρ(0), which
is clearly equivalent to the solution of the Lindblad equation [since it maps ρ(0) → ρ(dt) → ρ(2dt) → ⋯ → ρ(t)], i.e., if
ρ̇ = Lρ(t), with L the Lindbladian of Eq. (269a), then Λt = eLt . Since we have shown that Φ is a CP map, it remains to be
shown that a concatenation of quantum maps is still a quantum map. This is true, since if Φ1 and Φ2 are quantum maps then
Φ2 ○ Φ1 (ρ) = Φ2 [Φ1 (ρ)] = ∑ Kβ′ Φ1 (ρ)Kβ′ = ∑ Kβ′ Kα (ρ)Kα† Kβ′ = ∑ Kγ′′ ρKγ′′ ,
† † †
(273)
β αβ γ
where Kγ′′ = Kβ′ Kα , and ∑γ Kγ′′ Kγ = ∑α Kα† (∑β Kβ′ Kβ )Kα = ∑α Kα† Kα = I as required.
† †
D. Examples
Consider the Lindblad equation with all γα = 0, i.e., ρ̇ = −i[H, ρ]. This is just the Schrödinger equation written for density
matrices (also known as the Liouville-von Neumann equation). Let us solve it for the case of a single qubit. We can always write
⃗ = (hx , hy , hz ) ∈ R3 (since the Pauli matrices with identity form a basis over R4 for all 2 × 2
H = h0 I + ∑i∈{x,y,z} hi σ i , with h
matrices). Thus, using ρ = 2 (I + v⃗ ⋅ σ
1
⃗ ):
i i ⃗ × v⃗) ⋅ σ
− i[H, ρ] = − ∑ hi [σ , v⃗ ⋅ σ
⃗] = − hi vj [σ i , σ j ] = εijk hi vj σ k = (h ⃗.
i
∑ ∑ (274)
2 i∈{x,y,z} 2 i,j∈{x,y,z} i,j,k∈{x,y,z}
Since ρ̇ = 12 (v⃗˙ ⋅ σ
⃗ ), we find
v̇ ⋅ σ ⃗ × v⃗) ⋅ σ
⃗ = 2(h ⃗, (275)
which are three coupled first order differential equations for the components of v⃗. These are known as the Bloch equations,
and their solution has the Bloch vector v⃗ rotating around the vector h ⃗ with a frequency equal to 2∥h∥,
⃗ as is easily checked.
For example, consider a rotation about the vx axis, i.e., let h ⃗ = (h, 0, 0). Then Eq. (275) becomes: v̇x = 0, v̇y = −2hvz , and
v̇z = 2hvy . Differentiating again gives v̈y = −4h2 vy . The solution of these equations is
The general case follows from this one by a reorientation of the axes to align with what we called the vx axis in the solution
above.
49
We already encountered the phase damping model in the Kraus OSR setting in Sec. VII C. Let us now study a Lindblad
equation model that generates the same map.
Let L1 = σ z = Z, γ1 = γ, γα≥2 = 0, and H = 0. Thus,
1
ρ̇(t) = γ(ZρZ † − {Z † Z, ρ}) = γ(ZρZ − ρ) . (277)
2
Using ρ = 12 (I + v⃗ ⋅ σ
⃗ ), the left-hand side evaluates to ρ̇ = 21 v⃗˙ ⋅ σ
⃗ . For the right hand side ZρZ = 12 (I − vx X − vy Y + vz Z), and
we thus arrive at:
1
(v˙x X + v˙y Y + v˙z Z) = −γ (vx X + vy Y ) . (278)
2
Equating the two sides componentwise (multiply both sides by X, Y , or Z, and take the trace) gives:
We can see that, since γ ≥ 0, the map is contractive, and the Bloch sphere collapses to the vz -axis exponentially fast with time.
In the limit t → ∞, this simply projects every state directly to the vz -axis, which is manifestly uninvertible.
We can now match the Lindblad√ equation solution to the Kraus OSR result from Sec. VII C, where we found the Kraus
√
operators K0 = pI and K1 = 1 − pZ, and found that the Bloch vector is mapped to
v⃗′ = ((2p − 1)vx (0), (2p − 1)vy (0), vz (0)) ≡ (vx (t), vy (t), vz (t)) . (280)
The Lindblad phase damping result and the Kraus OSR thus have exactly the same effect provided we identify
1
2p − 1 = e−2γt Ô⇒ p(t) = (1 + e−2γt ) . (281)
2
The probability in this model approaches 1/2 in the limit t → ∞.
If we now allow H ≠ 0, i.e., solve the full Lindblad equation ρ̇ = −i[H, ρ] + γ(ZρZ † − ρ), then the result in Sec. IX D 1 shows
that this gives rise to a rotating Bloch ellipsoid that is simultaneously shrinking exponentially along its principal axis.
Likewise, we can construct a Lindblad equation for amplitude damping, which we encountered as a quantum map in
Sec. VII D.
Let L1 = σ − = ∣0⟩⟨1∣ = (σ + )† , γ1 = γ, γα≥2 = 0, and H = 0. Plugging these into the Lindblad equation we get:
1 1
ρ̇(t) = γ (σ − ρσ + − {σ + σ − , ρ}) = γ (∣0⟩⟨1∣ρ∣1⟩⟨0∣ − {∣1⟩ ⟨0∣0⟩ ⟨1∣ , ρ}) (282)
2 2
Using ρ = 21 (I + v⃗ ⋅ σ
⃗ ) we find, for the right-hand side:
1 1
∣0⟩⟨1∣ [ (I + vx X + vy Y + vz Z)] ∣1⟩⟨0∣ = (∣0⟩⟨0∣ − vz ∣0⟩⟨0∣) (283a)
2 2
1 1
− ∣1⟩⟨1∣(I + vx X + vy Y + vz Z) = − (∣1⟩⟨1∣ + (vx + ivy )∣1⟩⟨0∣ − vz ∣1⟩⟨1∣) (283b)
4 4
1 1
− (I + vx X + vy Y + vz Z)∣1⟩⟨1∣ = − (∣1⟩⟨1∣ + (vx − ivy )∣1⟩⟨0∣ − vz ∣1⟩⟨1∣) . (283c)
4 4
Adding up all these terms gives:
1 1 1
(Z − vx X − vy Y − vz Z) , (284)
2 2 2
50
In this section we provide an alternative analysis leading to the Lindblad equation. The derivation is longer than the one
we saw in Sec. IX, but provides additional insight and generalizability. Our analysis follows Ref. [13], with some changes of
notation as well as clarifications and minor corrections.
A. Derivation
Let us start again with the Kraus OSR, and recall that the Kraus operators act on HS , i.e., Kα ∈ B(HS ). Let us introduce a
d2 −1
fixed (time-independent) operator basis {Fi }i=0
S
for B(HS ), where dS = dim(HS ), such that F0 = I. We can then expand the
Kraus operators in this basis:
d2S −1
Kα (t) = ∑ biα (t)Fi , (287)
i=0
where biα are the time-dependent elements of a (rectangular) d2S × d2B -dimensional matrix b, and dB = dim(HB ). Then the
Kraus OSR becomes:
ρ(t) = ∑ Kα (t)ρ(0)Kα† (t) = ∑ χij (t)Fi ρ(0)Fj† (288a)
α ij
= χ00 (t)ρ(0) + ∑[χ0i (t)ρ(0)Fi† + χi0 (t)Fi ρ(0)] + ∑ χij (t)Fi ρ(0)Fj† , (288b)
i>0 i,j>0
where
χij (t) = ∑ biα (t)b∗jα (t) , (289)
α
i.e., χ = bb† . It follows immediately that χ is positive semidefinite: ⟨v∣ χ ∣v⟩ = ∥b† ∣v⟩ ∥2 ≥ 0. Note that χ is a d2S × d2S matrix.
Now consider the normalization condition:
I = ∑ Kα† (t)Kα (t) = ∑ χij (t)Fj† Fi (290a)
α ij
We can use this to eliminate the χ00 ρ(0) term from Eq. (288b). Multiply Eq. (290b) first from the right by 12 ρ(0), then from the
left, and add the resulting two equations:
1 1
ρ(0) = χ00 (t)ρ(0) + ∑ [χ0i (t) (Fi ρ(0) + ρ(0)Fi ) + χi0 (t) (Fi ρ(0) + ρ(0)Fi )] + ∑ χij (t) {Fj Fi , ρ(0)} .
† † †
(291)
2 i>0 2 i,j>0
and note that Q = Q† , i.e., Q is Hermitian. Then we can rewrite Eq. (292) as:
1
ρ(t) − ρ(0) = −i[Q(t), ρ(0)] + ∑ χij (t) (Fi ρ(0)Fj† − {F † Fi , ρ(0)}) . (294)
i,j>0 2 j
This obviously resembles the Lindblad equation, but it relates the state at t = 0 to the state at some arbitrary later time t, i.e., it
still represents a quantum map. Indeed, everything we have done so far is exact and we have simply rewritten the Kraus OSR
in a fixed operator basis. As a first step towards getting this closer to standard Lindblad form, let us diagonalize the χ matrix,
which will allow us to rewrite the double sum in Eq. (294) as a single sum. We have already noted that χ ≥ 0, so that it can be
diagonalized via some unitary matrix u: γ̃ = uχu† , where γ̃ is diagonal and positive semidefinite. Define Lk = ∑j>0 u∗kj Fj , so
that, using the unitarity of u:
Fi = ∑ uki Lk , (295)
k>0
where the sum over k > 0 excludes L0 = I. Thus, again using the unitarity of u:
where γk ≥ 0 are the eigenvalues of χ. We can now rewrite Eq. (294) as:
1 †
ρ(t) − ρ(0) = −i[Q(t), ρ(0)] + ∑ γk (t) (Lk ρ(0)L†k − {L Lk , ρ(0)}) . (297)
k>0 2 k
This is as far as we can go towards the Lindblad equation without introducing an approximation.
Let us now take a step back and introduce a generator for the exact quantum map. I.e., let us write ρ(t) = Λ(t, 0)[ρ(0)],
where
t
L(s)ds
Λ(t, 0) = T+ e∫0 . (298)
Let τ denote a short time interval, where the meaning of short will become clear momentarily. We define a “coarse-grained”
generator Lj as follows:
1 (j+1)τ
Lj = ∫ L(s)ds . (299)
τ jτ
t
∫0 L(s)ds = ∑j=0 Lj provided t = nτ , so that
1 n−1
Then τ
n−1
Lj
Λ(t, 0) = T+ eτ ∑j=0 . (300)
This assumption amounts to there being no memory of the evolution from one interval to the next.6 Under this assumption,
which we can also understand as a Markovian approximation, the time-ordered exponential becomes a product of exponentials:
n−1 n−1
Λ(t, 0) = ∏ eτ ∑ Lj ≡ ∏ Λj . (302)
j=0 j=0
τ ∥Lj ∥ ≪ 1 ∀j . (304)
Note that Eq. (485) sets an upper bound on τ in terms of the largest eigenvalue of the coarse-grained Lindblad generator.
This eigenvalue determines the fastest timescale for the system evolution (we’ll see later that these eigenvalues are all possible
differences of energies, i.e., they correspond to transition frequencies). Thus, Eq. (485) can also be interpreted as stating that the
coarse-graining timescale should be small compared to the timescale over which ρj changes.
Eq. (303) implies that, in particular, for j = 0:
ρ(τ ) − ρ(0)
= L0 [ρ(0)] . (305)
τ
Lemma 1.
Therefore
It follows immediately that χ(0) is already diagonal, and its eigenvalues are γ0 (0) = 1 and γk>0 (0) = 0. It also follows
immediately from Eq. (293) that Q(0) = 0.
Now define
1 (j+1)τ
⟨X⟩j ≡ ∫ X(s)ds . (310)
τ jτ
6 It is an interesting open problem to derive rigorous conditions for this to hold from first principles.
53
Then
ρ(τ ) − ρ(0)
⟨ρ̇⟩0 = (311a)
τ
Q(τ ) − Q(0) Q(τ )
⟨Q̇⟩0 = = (311b)
τ τ
γk (τ ) − γk (0) γk (τ ) − δk0
⟨γ̇k ⟩0 = = . (311c)
τ τ
We can therefore rewrite Eq. (297) as:
ρ(τ ) − ρ(0) Q(τ ) γk (τ ) − δk0 1
= −i[ , ρ(0)] + ∑ (Lk ρ(0)L†k − {L†k Lk , ρ(0)}) , (312)
τ τ k>0 τ 2
which must equal L0 [ρ(0)] by Eq. (305). Hence, we can read off L0 :
1 †
L0 [X] = −i[⟨Q̇⟩0 , X] + ∑ ⟨γ̇k ⟩0 (Lk XL†k − {L Lk , X}) , (313)
k>0 2 k
This generator is precisely in Lindblad form. However, it only connects ρ(0) to ρ(τ ). In order to connect ρ(jτ ) to ρ((j + 1)τ )
we may now postulate that the same generator form remains valid, i.e., that
1 †
Lj [X] = −i[⟨Q̇⟩j , X] + ∑ ⟨γ̇k ⟩j (Lk XL†k − {L Lk , X}) ∀j , (314)
k>0 2 k
which we can do as long as Eq. (301) is satisfied. The simplest way to ensure this is to demand that in fact
Lj = L0 ∀j . (315)
This is again the Markovian limit, where there is no memory of the previous evolution segment. If, instead, we keep the more
general form of Eq. (314) [again, subject to Eq. (301)], then we have a time-dependent Markovian process, where the generator
is allowed to change over time, as long as these changes are uncorrelated between different time-segments.
Retaining the time-independent Markovian form of Eq. (315), and further replacing ⟨ρ̇⟩j by ρ̇ (another approximation, that
becomes exact in the limit τ → 0), we finally have the following result for the coarse-grained Lindblad equation, representing a
time-independent Markovian limit:
1 †
ρ̇(t) = −i[⟨Q̇⟩0 , ρ(t)] + ∑ ⟨γ̇k ⟩0 (Lk ρ(t)L†k − {L Lk , ρ(t)}) (316)
k>0 2 k
One point remains, which is to show that the coefficients ⟨γ̇k ⟩0 are non-negative, which is a requirement for complete positivity
of the map generated by the Lindblad equation. To show this, note that
1 τ 1
⟨γ̇k ⟩0 = ∫ γ̇k (t)dt = (γk (τ ) − γk (0)) . (317)
τ 0 τ
We already know that γk (t) ≥ 0 ∀t (recall that these are the eigenvalues of χ), so we need to show that nothing is spoiled by
subtracting γk (0). But, this is true since we already showed above that γk>0 (0) = 0. Thus, Eq. (317) shows that the coefficients
are all non-negative, as required for the Lindblad equation.
B. Interaction picture
As a brief digression, let us review the interaction picture, in preparation for the example we shall study in the next subsection.
Consider a (time-dependent) Hamiltonian H of the form:
H(t) = H0 (t) + V (t) . (318)
The unitary evolution operators satisfy:
dU (t)
= −iH(t)U (t) (319a)
dt
dU0 (t)
= −iH0 (t)U0 (t) . (319b)
dt
Define the interaction picture propagator with respect to H0 via:
Ũ (t) = U0† (t)U (t, 0) . (320)
54
dŨ (t)
= −iH̃(t)Ũ (t), (321)
dt
with the interaction picture Hamiltonian
Proof. Differentiate both sides of Eq. (320), while making use of Eqs. (318), (321) and (322):
To make contact with open quantum systems, let V = HSB and H0 = HS + HB . Then U0 = e−itHS ⊗ e−itHB . We can now
transform the Schrödinger picture density matrix to the interaction picture via ρ̃SB (t) = U0† (t)ρSB (t)U0 (t), and if we write
HSB = ∑a λa Sa ⊗ Ba (Sa and Ba are system-only and bath-only operators, respectively), then H̃SB (t) = ∑a λa Sa (t) ⊗ Ba (t),
where Sa (t) = eitHS Sa e−itHS and Ba (t) = eitHB Ba e−itHB . This interaction picture density matrix satisfies
(note that the Schrödinger picture and the interaction picture coincide at t = 0).
At this point everything we’ve shown for quantum maps and the Lindblad equation carries through with appropriate modifi-
cations. The Kraus OSR in the interaction picture becomes
where Q̃ = Q − HS and γ̃k are the eigenvalues of the interaction picture χ-matrix χ̃ = b̃b̃† , with b̃ the expansion matrix of the
interaction picture Kraus operators: K̃α (t) = ∑i b̃iα (t)Fi .
To illustrate the predictions of the coarse-grained Lindblad equation, we consider the spin-boson model for phase damping of
a single qubit, described by the Hamiltonian
H = HS + HB + HSB (328a)
1
HS = − gZ , HB = ∑ ωk (nk + 1/2) , HSB = Z ⊗ (∑ λk bk + λ∗k b†k ) , (328b)
2 k k
where nk = b†k bk and bk are the bosonic number and annihilation operator for mode k, respectively ([bk , b†l ] = δkl I). Here
HSB describes
√ coupling of the qubit phase to the position x of each oscillator; recall that quantization means replacing x by
(b + b† ) / 2mω (where m is the oscillator mass), so that
√
λk ∝ 1/ ωk , (329)
55
GHtL
Exact
2
Τ=2Ωc
Τ=8Ωc
1.5
Τ=32Ωc
0.5
Ωc t
2 4 6 8 10 12 14
FIG. 5. Comparison of the exact solution of the spin-boson model for single-qubit phase damping to the result obtained from the coarse-grained
Markovian master equation. Plotted are the arguments Γ(t) of the exponentials in Eq. (342). Straight lines correspond to the Markovian
solution, which intersects the exact solution (thick line) at t = τ , as seen from Eqs. (342a) and (342b). The bosonic bath density of states is
represented by the Debye model [Eq. (341)]. The results shown correspond to C = 0.05 and ωc = 1. Reproduced from Ref. [13].
a relation we will need later. In the interaction picture, it is easy to show that:7
Assume that the bath is initially in a thermal Gibbs state at inverse temperature β = 1/T : ρB (0) = e−βHB /Z [Eq. (237)], and
let ⟨X⟩B ≡ Tr(XρB ). It is then a standard exercise to show that
1
⟨b†k bl ⟩B = δkl , ⟨b†k ⟩B = ⟨bk ⟩B = ⟨bk bl ⟩B = ⟨b†k b†l ⟩B = 0 . (331)
eβωk −1
Using this, it can be shown that the coarse-grained, interaction picture Lindblad equation Eq. (327) becomes [13]:
˙
ρ̃(t) = γ(τ ) (Z ρ̃(t)Z − ρ̃(t)) , (332)
˙ = 0, L = Z, and there are no other Lindblad operators (as should be obvious from the form of H above), and where
i.e., ⟨Q̃⟩ 0 1 SB
As we shall see in Sec. XI, the spin-boson model we are considering here has an exact analytical solution. The exact solution
for the coherence is:
(e)
ρ̃01 (t) = e−2γ(t)t ρ01 (0) . (336)
This allows us to compare the Markovian result to the exact one, and better understand the condition the coarse-graining
timescale τ must satisfy. The only difference between the two is the argument of γ: τ versus t. However, this is a very
7 Some basic quantum mechanics would make this process very simple. Note that [b, n] = b gives bn = (n + 1)b. And therefore we would have ben = en+1 b.
56
significant difference, since while the Markovian solution represents irreversible exponential decay, the exact solution is oscil-
latory: γ(t)t ∼ ∑k sin2 (ωk t). In order to observe closer agreement, we must once again invoke a continuous density of states
γ(ω), as we did in Sec. VIII B [recall Eq. (252)], which results in irreversible decay also in the case of the exact solution. Doing
so replaces Eq. (333) by
ωc
γ(τ ) = π ∫ Ω(ω)∣λ(ω)∣2 coth(βω/2)δ̄(ω, τ )dω , (337)
0
where we assumed that Ω(ω) has a high-frequency cutoff at ωc . Now note that δ̄ behaves similarly to the Dirac-δ function:
∞
∫ δ̄(ω, τ )dω = 1 , lim δ̄(ω, τ ) = δ(ω) , (338)
0 τ →∞
i.e., it is sharply peaked at ω = 0, and the peak becomes sharper as τ grows. The peak width is ∼ 1/τ . This suggests under
ω
what condition γ(t) ≈ γ(τ ), such that the exact and Markovian solutions agree: τ ≫ 1/ωc . The reason is that then ∫0 c captures
nearly all the area under δ̄(ω, τ ), whereas in the opposite case (τ ≲ 1/ωc ), most of the area under δ̄(ω, τ ) is not captured by the
same integral. Thus, assuming τ ≫ 1/ωc , δ̄(ω, τ ) effectively behaves as a Dirac-delta function, and if we assume in addition
that t > τ , then certainly also δ̄(ω, t) behaves as a Dirac-δ function. Thus, assuming
t > τ ≫ 1/ωc , (339)
we have
ωc
γ(τ ) ≈ γ(t) ≈ π ∫ Ω(ω)∣λ(ω)∣2 coth(βω/2)δ(ω)dω , (340)
0
so that the exact and Markovian cases agree. This is borne out numerically as well. Assume a Debye model, so that
ω 2 for ω < ωc
Ω(ω) ∝ { , (341)
0 for ω ≥ ωc
and that ∣λ(ω)∣2 ∝ ω −1 , in accordance with Eq. (329). In the high-temperature limit coth(βω/2) ∝ ω −1 , so that in all we have
ωc
ρ̃01 (t) ∝ exp (−Ctτ ∫ dωsinc2 (ωτ /2)) (342a)
0
ωc
(e)
ρ̃01 (t) ∝ exp (−Ct2 ∫ dωsinc2 (ωt/2)) , (342b)
0
where C is the temperature-dependent coupling-strength, with dimensions of frequency. Figure 5 shows the argument of the
exponentials in Eq. (342), Γ(t), for the exact solution and for the coarse-grained Lindblad equation, corresponding to different
values of the course-graining time-scale, τ . The curves corresponding to the Markovian solutions are just straight lines, as they
all describe simple exponential decays. It is clear that the Markovian solutions cannot account for the initial transition period,
but for sufficiently large τ (in units of the bath cutoff time 1/ωc ) the Lindblad result approximates the exact solution very well
at large times.
To summarize, the Markovian approximation gives reliable results for times greater than the coarse-graining time-scale, which
in turn must be greater than the inverse of the bath high-frequency cut-off. It does not account for the initial (Zeno-like) time
evolution.
We present the analytical solution of the spin-boson model for pure dephasing. The derivation is based on [13, 14].
The model is the same as the one we considered in Sec. X C, except that we will consider a system of multiple qubits (indexed
by i). Starting from the interaction picture system-bath Hamiltonian [generalizing Eq. (330)]:
∗
H̃SB (t) = ∑ Zi ⊗ [λik e−iωk t ak + (λik ) eiωk t a†k ] , (343)
i,k
Note that H̃(t) does not commute with itself at different times, which is why we need the time-ordered product:
where we used the canonical bosonic commutation relations [ak , a†k′ ] = − [a†k , ak′ ] = Iδkk′ , [ak , al ] = [a†k , a†l ] = 0. Note that
further,
[[H̃(t), H̃(t′ )] , H̃(t′′ )] = 0. (347)
This means that we can use the Baker-Hausdorf formula exp(A + B) = exp(−[A, B]/2) exp(A) exp(B) (valid if [[A, B], A] =
[[A, B], B] = 0) to calculate U (t). To do so note the generalization
1
exp (∑ An ) = ( ∏ exp (− [An , An′ ])) (∏ exp(An )) , (348)
n n<n′ 2 n
which is valid if every second-order commutator vanishes. To apply this result for our case let us formally discretize the integrals
and denote Hn ≡ −iH̃(n∆t). We let ∆t = t/N and take the limit N → ∞. Then:
t N
U (t) = T+ exp [−i ∫ H̃(τ )dτ ] = T+ lim exp [ ∑ Hn ∆t] (349a)
0 ∆t→0 n=0
1 2
= lim ∏ exp (− [Hn , Hn′ ] (∆t) ) ∏ exp(Hn ∆t) (349b)
∆t→0 n<n′ 2 n
1 2
= lim ∏ (1 − [Hn , Hn′ ] (∆t) ) ∏ (1 − Hn ∆t) (349c)
∆t→0 n<n′ 2 n
1 2
= lim [1 − ∑ [Hn , Hn′ ] (∆t) ] [1 − ∑ Hn ∆t] (349d)
∆t→0 2 n<n′ n
1 2
= lim exp (− ∑ [Hn , Hn′ ] (∆t) ) exp (∑ Hn ∆t) (349e)
∆t→0 2 n<n′ n
1 t t1 t
= exp [ ∫ dt1 ∫ dt2 [H̃(t2 ), H̃(t1 )]] exp [−i ∫ H̃(τ )dτ ] . (349f)
2 0 0 0
Note that in the second line we enforced time-ordering by keeping n < n′ . To go from the third to the fourth line we kept the
lowest relevant order in each term, inherited from the second line. Note how in the last line the time-ordering is implemented
via t2 ≤ t1 . We find:
t
− i∫ H̃(τ )dτ = ∑ Zi ⊗ ∑ ((αki )∗ ak − αki a†k ) , (350)
0 i k
where
∗
(λik ) (eiωk t − 1)
αki (t) = . (351)
ωk
Now, since
t t1 t e−iωk t1 − 1 1 − eiωk t + iωk t
∫ dt1 ∫ dt2 e−iωk (t2 −t1 ) = ∫ dt1 eiωk t1 = , (352)
0 0 0 −iωk ωk2
we have, using Eq. (346b):
i t t1 ∗ t t1
j′ −iωk (t2 −t1 )
− ∫ dt1 ∫ dt2 [H̃(t2 ), H̃(t1 )] = ∑ Zj Zj ′ ∑ I [λk (λk ) ∫ dt1 ∫ dt2 e ] ⊗ IB
j
(353a)
2 0 0 jj ′ k 0 0
′ ∗ 1 − eiωk t + iωk t
= ∑ Zj Zj ′ ∑ I [λjk (λjk ) ] ⊗ IB . (353b)
jj ′ k ωk2
58
Therefore, defining
′ ∗ eiωk t − iωk t − 1
fjj ′ (t) ≡ ∑ I [λjk (λjk ) ], (354)
k ωk2
we can write the first term in Eq. (349f) as follows:
1 t t1
exp [ ∫ dt1 ∫ dt2 [H̃(t2 ), H̃(t1 )]] = ei ∑jj′ fjj′ (t)Zj Zj′ ⊗ IB . (355)
2 0 0
Note that this is an operator acting non-trivially just on the system, and is a global phase for the case of a single qubit. Its action
is, however, non-trivial for multiple qubits (it represents a Lamb shift).
Since the ak operators commute for different modes we have as our final simplified result for the evolution operator:
Ũ (t) = eif (t) ∏ exp [Zi ⊗ (αki (t)ak − αki (t)∗ a†k )] . (356)
i,k
Now recall the definition of the coherent states. These are eigenstates of the annihilation operator:
a∣α⟩ = α∣α⟩ . (357)
They are minimum-uncertainty states in a harmonic potential, and can be expanded as
∞
αn
∣α⟩ = e−∣α∣ /2
2
∑ √ ∣n⟩ (358)
n=0 n!
where ∣n⟩ are number (Fock) states. The completeness relation for coherent states is:
1
∫ d α ∣α⟩⟨α∣ = 1
2
(359)
π
where the integration is over the entire complex plane. They are useful in our context since they are created by the displacement
operator
D (α) ≡ exp (αa† − α∗ a) = D(−α)† (360)
acting on the vacuum state:
D (α) ∣0⟩ = ∣α⟩, (361)
which is clearly related to U (t). We will need the result:
αβ ∗ − α∗ β
D (α) D (β) = exp D(α + β), (362)
2
which is easily derived from D (α) = exp (αa† − α∗ a), [a, a† ] = 1, and the Baker-Hausdorf formula exp(A + B) =
exp(−[A, B]/2) exp(A) exp(B) (again, valid if [[A, B], A] = [[A, B], B] = 0).
Now let Rik (t) ≡ αki (t)a†k − αki (t)∗ ak and consider exp [Zi ⊗ Rik (t)]:
∞ ∞
R2n R2n+1
exp [Z ⊗ R] = IS ⊗ ∑ +Z ⊗ ∑ = IS ⊗ cosh R + Z ⊗ sinh R (363a)
n=0 (2n)! n=0 (2n+!)!
1 1
= IS ⊗ [D (α) + D (−α)] + Z ⊗ [D (α) − D (−α)] = ∣0⟩⟨0∣ ⊗ D (α) + ∣1⟩⟨1∣ ⊗ D (−α) . (363b)
2 2
This shows that depending on whether the field is coupled to the qubit ∣0⟩ or ∣1⟩ state, the field acquires a different displacement.8
The evolution operator can thus be written as:
U (t) = ei ∑jj′ fjj′ (t)Zj Zj′ ∏ [∣0⟩i ⟨0∣ ⊗ D (αki ) + ∣1⟩i ⟨1∣ ⊗ D (−αki )] . (364)
i,k
8 Note that this is the source of the dephasing the qubits undergo, since when acting on a superposition state of a qubit, the qubit and field become entangled:
exp [σz ⊗ R] (a∣0⟩ + b∣1⟩)∣β⟩ = a∣0⟩ ⊗ D (α) ∣β⟩ + b∣1⟩ ⊗ D (−α) ∣β⟩ = e(αβ a∣0⟩ ⊗ ∣α + β⟩ + e−(αβ b∣1⟩ ⊗ ∣β − α⟩ .
∗
−α∗ β)/2 ∗
−α∗ β)/2
59
−1
1 −βHB e−βωk /2 1
ρB = e = [∏ ] exp (−β ∑ ωk (Nk + )) = ∏ ρB,k , (365a)
k 1−e
Z −βωk 2
k k
where
1
ρB,k = exp (−βωk Nk ) , (366)
⟨Nk ⟩
and the mean boson occupation number is given by the Bose-Einstein distribution:
1
⟨Nk ⟩ = . (367)
eβωk −1
As shown in [15], p.122-3, this can be transformed into the coherent-state representation, with the result:
1 ∣αk ∣2
ρB,k = ∫ d αk exp (−
2
) ∣αk ⟩⟨αk ∣ . (368)
π⟨Nk ⟩ ⟨Nk ⟩
For simplicity let us from now on consider the case of a single qubit. It suffices to calculate the evolution of each of the four
pure states ∣x⟩⟨y∣, where x, y ∈ {0, 1}, separately. Thus
= TrB [∏ [∣0⟩⟨0∣ ⊗ D (αk ) + ∣1⟩⟨1∣ ⊗ D (−αk )] ∣x⟩⟨y∣ ⊗ ∏ ρB,m ∏ [∣0⟩⟨0∣ ⊗ D† (αl ) + ∣1⟩⟨1∣ ⊗ D† (−αl )]] .
k m l
The terms in the three products match one-to-one for equal indices, so we can write everything as a product over a single index
k. Using Tr(A ⊗ B) = TrA × TrB to rearrange the order of the trace and the products, and D† (−α) = D (α), we have:
Consider the Tr terms: for ∣0⟩⟨0∣ and ∣1⟩⟨1∣ by cycling in the trace the displacement operators cancel and Tr [ρB,k ] = 1. Thus, as
expected the diagonal terms do not change:
1 ∣βk ∣2
Tr [D (±2αk ) ρB,k ] = ∫ d βk exp (−
2
) ⟨βk ∣D (±2αk ) ∣βk ⟩ . (371)
π⟨Nk ⟩ ⟨Nk ⟩
Now:
Thus:
1 ∣βk ∣2
Tr [D (±2αk ) ρB,k ] = exp (−2∣αk ∣2 ) ∫ d βk exp (−
2
± 2 (αk βk∗ − αk∗ βk )) (373a)
π⟨Nk ⟩ ⟨Nk ⟩
exp (−2∣αk ∣2 )
= [π⟨Nk ⟩ exp (−4∣αk ∣2 ⟨Nk ⟩)] (373b)
π⟨Nk ⟩
1
= exp [−4∣αk ∣2 (⟨Nk ⟩ + )] (373c)
2
⎡ 2 ⎤
⎢ ∗
λk (e iω k t
− 1) 1 1 ⎥⎥
⎢
= exp ⎢−4 ∣ ∣ ( βω + )
2 ⎥⎥
(373d)
⎢ ω e k −1
⎣ k
⎦
1 − cos(ωk t) βωk
= exp [−4∣λk ∣2 2
coth ] . (373e)
ωk 2
which coincides with the exact result quoted in Sec. X C, specifically Eq. (333) with τ replaced by t.
Solving the Lindblad equation numerically is demanding. For a d-dimensional system Hilbert space, the density matrix is
d × d, involving d2 − 1 real numbers that one must store and update at each time-step. Is there a more space-efficient alternative?
It turns out that instead one can propagate a wavefunction (only 2d − 1 real numbers, so a quadratic savings), at the expense
of introducing statistical averaging over many runs. A very interesting side-benefit of this so-called unravelling procedure is
that each wavefunction undergoes a “quantum trajectory”, that can be correlated to an individual sequence of quantum events,
whereas the density matrix instead corresponds to an ensemble of such events.
Let us write down the Lindblad equation [Eq. (265)] in the following form:
2
d
1
ρ̇ = −i[H, ρ] + ∑ γk (Lk ρL†k − {L†k Lk , ρ}) (375)
k=1 2
Here Lk are the Lindblad operators and γk are scalars. As is clear from the derivation presented in Sec. X, the number of
non-zero terms in the sum is at most d2 . If one sets ∥Lk ∥ = 1 then the scalars γk can be understood as rates of the corresponding
relaxation process.9
There are multiple ways we can proceed to study this equation:
2. Find equivalent dynamics of the wavefunction ∣ψ(t)⟩ (in the closed system case the wavefunction is a d-dimensional
vector over C such that ⟨ψ∣ψ⟩ = 1; this time we will let its norm be arbitrary);
3. Suppose that measurements are performed repeatedly on the system, and derive the equation for dynamics given a string
of measurement outcomes.
Here we will address points 2 and 3. In a very narrow sense we will address 1, if the closed system + measurement apparatus
are thought of as an open system.
A. Method summary
where
i
HC = H − ∑ γα Lα Lα
†
(377)
2 α
is called the “conditional Hamiltonian”. Note that it is non-Hermitian. Consider the evolution of a pure state ∣ψ(0)⟩ subject to
HC :
HC
∣ψ(0)⟩ z→ e−iHC t ∣ψ(0)⟩ = ∣ψ̃(t)⟩ . (378)
Since HC is non-Hermitian, the norm of ∣ψ̃(t)⟩ decreases over time (hence the tilde):
d
∥ ∣ψ̃(t)⟩ ∥2 = ⟨ψ(0)∣ eiHC t (iHC† )e−iHC t + eiHC t (−iHC )e−iHC t ∣ψ(0)⟩
† †
(379a)
dt
= i ⟨ψ(0)∣ eiHC t (HC† − HC )e−iHC t ∣ψ(0)⟩
†
(379b)
= − ∑ γα ⟨ψ(0)∣ e iHC† t
L†α Lα e−iHC t ∣ψ(0)⟩ (379c)
α
The action of the other term in Eq. (376) can be viewed as inducing a “quantum jump”:
In this way, the probability of a second jump at τ = 0 is zero, but the probability increases exponentially as τ grows.
Putting these steps together one arrives at the following algorithm for evolution from t = 0 to tf :
1. Initialize the state as ∣ψ(0)⟩, set j = 1
4. If a jump took place, advance j to j + 1: call the new (normalized state) ∣ψ(tj+1 )⟩ and set tj+1 = tj + τ
5. Return to step 2, unless tj+1 ≥ tf
6. Repeat K times from step 1, calling the output from the k th round ψk (tf ), and construct ρ(tf ) = 1
K ∑k=1 ∣ψk (tf )⟩⟨ψk (tf )∣,
K
Each sequence {ψk (0), ψ̃k (t1 ), ψk (t1 ), ψ̃k (t2 ), ψk (t2 ), . . . , ψk (tj ), ψ̃k (tj ), . . . } is a “quantum trajectory”. It describes a
series of norm-decreasing evolutions interrupted by quantum jumps. This provides an interesting and insightful interpretation of
what actually takes place during open quantum system evolution. Consider, e.g., generalized amplitude damping (Sec. VII G).
An atom undergoes spontaneous emission to its ground state, but due to thermal excitation it can repopulate its excited state. As
we saw in Sec. IX D 3, the probability of a transition from the excited state to the ground state increases exponentially with time,
which is in accordance with Eq. (381). But now we see that the actual emission event is a “jump”, whereby the atom suddenly
and discontinuously finds itself in the ground state. The process can also work in the opposite direction, and by absorbing energy
from the bath, the atom can find itself in an excited state, etc. The downward transition event is accompanied by the emission
of a photon (by energy conservation), or phonon, or some other elementary excitation, which can be detected. And indeed, such
quantum trajectories have been measured in quantum optics experiments (see, e.g., Ref. [16] and references therein).
We now proceed to give a more careful and detailed description and analysis.
1. Naive form
Starting over, we note that we can rewrite the Lindblad equation as follows, in the limit dt → 0:
1
ρ(t + dt) = ρ(t) − i[H, ρ(t)]dt − ∑ γk {L†k Lk , ρ(t)}dt + ∑ γk Lk ρ(t)L†k dt (382a)
k=1 2 k=1
We note that instead of using a differential equation solver to obtain ρ(t + dt), using the non-selective measurement formalism
of Sec. VI A we can instead mathematically “simulate” the above formula in the following way:
1. choose k ≥ 0 with probability pk = Tr[Mk ρ(t)Mk† ];
Now we note that the whole process was linear with respect to ρ(t) = ∑i pi ∣ψi (t)⟩⟨ψi (t)∣. So we can work with the states
instead! Generate a random number i with probability pi given by initial conditions, so as to choose ∣ψi (0)⟩ as the initial state
(a pure state). Then follow these instructions with normalized ∣ψ(t)⟩ at each step to produce ∣ψ(t + dt)⟩:
1. choose k ≥ 0 with probability pk = ⟨ψ(t)∣Mk† Mk ∣ψ(t)⟩
Here we average over random numbers {k}, i to obtain the same density matrix as the solution of the master equation. For a
small range of k and simple operators Lk this method already leads to substantial savings, as one never needs to store d × d
63
matrices during the simulation, only d-dimensional vectors. However, note that to obtain the average in practice one needs to
sample from {k}, i some number of times K, repeating the whole simulation. In principle K can be as large as d, thus defeating
the purpose of the method, but in practice one can observe convergence of the average with increasing K, e.g. by studying the
dispersion of some observable
We consider a slightly different perspective that is essentially the same as above, but we note that k = 0 corresponding to
†
M0 = eiHC dt dominates the probability distribution for k in the limit dt → 0:
this means that one does not need to calculate ⟨ψ(t)∣Mk† Mk ∣ψ(t)⟩ every dt. One only calculates
and generates an auxilliary random variable JUMP= 0, 1 with probability pno-jump , 1 − pno-jump respectively. Only if JUMP= 1 we
ask which k actually happened.
Looking at Eq. (387) we see that at first the probability of a jump happening within an interval [t, t + τ ] increases from 0
linearly with τ , and at large τ it approaches 1 exponentially. The coefficient in front of the linear dependence is ∣ψ⟩-dependent,
but weakly so. There is a well-known stochastic process given by
In other words, independent jumps occur with rate r per unit of time. This process is called telegraph noise. The simula-
tion method described above is a quantum evolution interrupted by essentially independent jumps following a telegraph noise
distribution. Below we will study a different type of noise.
What we did above was produce a map from a wavefunction ∣ψ(t)⟩ plus a random variable ξ to the wavefunction at the next
time step ∣ψ(t + dt)⟩. The way we proved that this map is equivalent to the original master equation is by observing that
Avξ ∣ψ(t + ∆t, ξ)⟩⟨ψ(t + ∆t, ξ)∣ = ∣ψ(t)⟩⟨ψ(t)∣ + L(∣ψ(t)⟩⟨ψ(t)∣)∆t + O(∆t) . (390)
1 1
L = AρA − A2 ρ − ρA2 . (391)
2 2
Let the random variable ξ actually be a stochastic function of time ξ(t) on the interval [t, t + ∆t]. Define the time-step for our
trajectory as:
t+∆t
ξ(τ )dτ
∣ψ(t + ∆t)⟩ = eiA ∫t ∣ψ(t)⟩ . (392)
d
∣ψ(t)⟩ = iAξ(t)∣ψ(t)⟩ . (393)
dt
64
We can do a second order Taylor series expansion of Eq. (408). The average of the density matrix after our time-step is then
given by:
to get rid of the first order in A. We also define the correlation function
to be translation-invariant in time (i.e., to depend only the difference τ − τ ′ ). Together these two conditions define the first two
moments of Gaussian stochastic random variable. We then have:
We note that this will be the case if C(t) is peaked at 0 with width w ≪ ∆t and height C0 :
t+∆t t+∆t
∫ ∫ C(τ − τ ′ )dτ dτ ′ ≈ wC0 ∆t (399)
t t
Setting wC0 = 1 will recover the desired Lindblad generator L given in Eq. (391). Since w is the smallest timescale in the
problem we can just choose
Now the smallness of the remaining terms can be guaranteed as O(∆t2 ), and we have indeed recovered the Lindblad generator
L given in Eq. (391).
Using the same idea for the derivation, we can prove the equivalence between the original Lindblad equation (375) and the
following differential equation on ∣ψ(t)⟩:
d √
∣ψ(t)⟩ = −i(H − ∑ Lk ξk (t))∣ψ(t)⟩, Avξ ξk (t)ξm (t′ ) = δkm δ(t − t′ ) γk . (402)
dt k
Here δkm is the Kronecker delta function. The equivalence states that
the limit is included in δ-function and the definition of the differential equation, so no additional limit needs to be taken here. In
t+∆t
practice, though, some discretization scheme needs to be applied and the numerical simulation uses ∫t ξ(τ )dτ instead of the
raw ξ(t).
65
4. Comparison between the telegraph noise and Stochastic Schrödinger equation approaches
If we compare the Stochastic Schrödinger equation approach to the telegraph noise method, we find that ξ tel (t) is a sequence
of randomly spaced peaks with 0 in between. It is possible to arrange for the correlation function of that signal to be C(t) = δ(t),
however the higher order correlation functions will be vastly different from the Gaussian noise that is usually used for stochastic
differential equations. The defining characteristic of the Gaussian noise is that higher order correlations (or moments) are
T
expressed via C(t) according to Wick’s theorem. Another way to think about it is that the Fourier transforms ∫0 ξ(t)eikt dt are
i.i.d. random variables for each k for Gaussian noise, but not for telegraph noise.
Let us discuss the properties of individual ∣ψ(t, ξ)⟩ or ∣ψ(t, {k})⟩ for a given realization of random variables, under the two
approaches. One way to look at this is to take an observable O (s.t. ∥O∥ = 1) and follow its average:
If the closed system evolution of the observable O(t) has a characteristic frequency ω ∼ ∥[O, H]∥ and the relaxation has the
characteristic rates r ∼maxk γk , then there are two possible regimes: ω ≪ r and ω ≫ r. The qualitative picture that we will see
is as follows:
ω≪r ω≫r
telegraph smooth curves interrupted rapid sine wave interrupted
by discontinuities by discontinuities
stochastic noisy diffusive behaviur noisy almost periodic behavior
Even though the two methods are both equivalent to the same master equation, other characteristics such as the dispersion D(O)
or the diffusion coefficient of individual trajectories vary between the two methods. Thus, we find very different visual behavior
of individual trajectories. It is possible to interpolate between the two by chossing a non-Gaussian ξ(t). We note that the results
for a single trajectory are reminiscent of experimental measurements. We next make this analogy more precise.
C. Weak measurements
One way is to choose the distribution of the random process ξ(t) in such a way that an individual trajectory ⟨ψ(t, ξ)∣O∣ψ(t, ξ)⟩
matches the measurement output M (t) of some repeated measurement. However this is an unphysical approach. What we should
be doing is to come up with a mapping M (ξ) since ξ contain the information about random choices made outside of the system,
while ⟨ψ(t, ξ)∣O∣ψ(t, ξ)⟩ contains information “private” to the system, something that has not been measured yet.
We note that the first method with the decomposition given in Eq. (383) can be directly interpreted as a measurement where
k is an answer. The stochastic one requires some transformations before this can be done, as the width of the δ-function is
the smallest time-scale that is faster than the supposed data collection timescale. We do not know of any research that makes
this connection. There is a lot of research connecting weak measurements with trajectories, which could be seen as such an
interpretation of stochastic equations. The difference with telegraph noise is that every Mk is close to identity I with a small
probability in front.
In this section we discuss the analytical solution of the Lindblad equation in arbitrary dimensional (but finite) Hilbert spaces.
Let us first introduce a “nice” operator basis for B(HS ), where d = dim(HS ). Let F0 = IS and choose M other traceless,
Hermitian operators {Fj }Mj=1 , where M = d − 1, such that
2
A common choice is the generators of su(d) (just as in the single-qubit case we chose the Pauli matrices), but for our purposes
the explicit form of the operator basis won’t matter. Note that this is similar to what we did in Sec. X A, except that for later
convenience we make our basis choice somewhat more explicit here.
66
We can now expand any operator in this basis, including the density matrix:
M
1 1
ρ= F0 + ∑ vj Fj = I + F⃗ ⋅ v⃗ , (406)
d j=1 d
where v⃗ = (v1 , . . . , vM )T ∈ RM is called the “coherence vector” (a generalized Bloch vector), and F⃗ = (F1 , . . . , FM ) collects
the operator basis into a vector. Thus the components of the coherence vector are
vj = Tr(ρFj ) . (407)
In analogy to Eq. (168) for the single qubit case, we shall see that as a consequence of the Lindblad equation ρ̇ = Lρ, the
coherence vector satisfies the first order, inhomogeneous differential equation
v⃗˙ = G⃗
v + c⃗ . (408)
L = LH + LD , (409)
with
i.e.,
1 1/2
0 ≤ ∥⃗
v ∥ ≤ (1 − ) . (412)
d
1/2
The upper bound is saturated for pure states, which thus live on the surface of an d2 −1-dimensional sphere with radius (1 − d1 ) .
Let us assume that LD = 0. In this case we have, starting from Eq. (407):
= (Q⃗
v )k , (413c)
i.e.,
v⃗˙ = Q⃗
v, (414)
67
where
Note that the appearance of the commutator [Fj , Fk ] is a good reason to use as an operator basis the generators of a Lie algebra,
for which the commutator can be expressed in terms of the algebra’s structure constants.
The matrix M × M dimensional Q is clearly skew symmetric: Qjk = −Qkj , i.e.,
Q = −QT . (416)
where we used the skew-symmetry of Q. This immediately implies that the norm of the coherence vector is preserved: ∥⃗ v (t)∥2 =
v⃗ (0)Ω Ω⃗
T T
v (0) = ∥⃗
v (0)∥ .
2
Thus, the evolution of the coherence vector in the absence of the dissipative part LD = 0 is a rotation in RM , generated by Q.
Let us now assume that both LH , LD ≠ 0. Starting again from Eq. (407), and using Eq. (410b), we have:
1 1 1
v )k + ∑ aij Tr [Fi ( I + F⃗ ⋅ v⃗) Fj Fk − {Fj Fi , I + F⃗ ⋅ v⃗}Fk ]
v̇k = (Q⃗ (419a)
ij d 2 d
1 1
v )k + ∑ ∑ aij Tr [(Fj Fk Fi − (Fk Fj Fi + Fj Fi Fk )) Fl ] vl + ∑ aij Tr([Fi , Fj ] Fk )
= (Q⃗ (419b)
l ij 2 d ij
= [(Q + R)⃗
v ]k + ck , (419c)
where
1
Rkl ≡ ∑ aij Tr [(Fi Fl Fj − {Fj Fi , Fl }) Fk ] (420a)
ij 2
1
ck ≡ ∑ aij Tr([Fi , Fj ] Fk ) . (420b)
d ij
Thus, we have established that Eq. (408) holds, with G = Q + R, and with Q, R, and c⃗ as given in Eqs. (415) and (420),
respectively.
Equation (408) is a linear, first order, inhomogeneous differential equation. Solving it is a standard exercise in linear algebra.
For simplicity, let us assume that G is diagonalizable over RM and also invertible. Neither of these assumptions holds in general,
and we deal with the general case in the next subsection.
We look for a solution in the form
where v⃗(0) (t) is the homogeneous part and v⃗(∞) is the inhomogeneous part. Let x
⃗(k) and λk represent the eigenvectors and
(possibly degenerate and complex) eigenvalues of G, i.e.,
x(k) = λk x
G⃗ ⃗(k) , k = 1, . . . , M . (422)
68
i.e., X is the matrix whose columns are the eigenvectors of G. Also, v⃗(0) (0) = v⃗(0) − v⃗(∞) . Thus
s⃗ = X −1 (⃗
v (0) + G−1 c⃗) . (426)
Now, since the eigenvalues are in general complex numbers, they can be decomposed as λk = R(λk )+iI(λk ). The imaginary
part describes a rotation of the coherence vector (though we can be sure that since this vector lives in RM , such rotations
are ultimately described by an orthogonal (purely real) matrix). The real part is constrained by complete positivity and trace
preservation to be non-positive, or else the norm of the coherence vector would not be bounded [recall Eq. (412)]. Thus,
the overall behavior of the coherence vector is described by rotations at frequencies given by {I(λk )}, some of which are
exponentially damped on a timescale given by the set of non-zero {R(λk )}.
The general case is where G is not diagonalizable over RM , and may not be invertible. In this case we can still use a similarity
transformation S to transform G into Jordan canonical form:
⎛ J1 ⎞
GJ = SGS −1
=⎜
⎜ ⋱ ⎟,
⎟ (427)
⎝ Jq ⎠
⎛ µj 1 ⎞
⎜ µj ⋱ ⎟
⎜ ⎟
⎜ ⎟
Jj = ⎜ ⋱ ⋱ ⎟ = µj I + Kj . (428)
⎜ ⎟
⎜ ⎟
⎜ µj 1 ⎟
⎝ µj ⎠
d
The µj ’s are the (possibly degenerate, complex) eigenvalues and Kj are nilpotent matrices: Kj j = 0, where dj is the dimension
of Jj . When all dj = 1, G is diagonalizable and GJ reduces to the diagonalized form of G.
Applying S from the left to Eq. (408) yields
S v⃗˙ = SGS −1 S⃗
v + S⃗
c Ô⇒ w ⃗ + c⃗′ ,
⃗˙ = GJ w (429)
where w⃗ = S⃗v and we defined c⃗′ = S⃗c. This is still a linear, first order, inhomogeneous differential equation. The different Jordan
blocks don’t couple, so we can solve this as a set of q independent problems, and take the direct sum of all the sub-solutions.
Consider first the case of a 2 × 2 Jordan block, i.e., dj = 2. The homogeneous part becomes:
(0) µ 1 (0)
⃗˙ j = ( j
w ⃗j ,
)w (430a)
µj
69
i.e.,
(0) (0) (0)
ẇj,1 = µj wj,1 + wj,2 (431a)
(0) (0)
ẇj,2 = µj wj,2 . (431b)
dj
(0) (0) tn−k
w ⃗j,n (0)
⃗j,k (t) = eµj t ∑ w , k = 1, . . . , dj . (434)
n=k (n − k)!
where the direct sum notation means that the summands need to be joined into a single column vector. The new aspect of the
general G case is thus the appearance of the degree dj − 1 polynomials in t. These polynomials induce an additional non-trivial
time-dependence in addition to the rotations and exponential decay we found for the case of diagonalizable G. Note that we
can be certain that for all dj > 1 [when the degree of the polynomial in Eq. (434) is ≥ 1], the corresponding R(µj ) < 0, since a
positive or zero real part would violate the general norm upper bound (412).
As for the inhomogeneous part, we can write the solution of Eq. (429) as
⃗
w(t) ⃗ (0) (t) + w
=w ⃗ (∞) , (436)
⃗ (∞) = −⃗
GJ w c′ . (437)
Depending on the rank r(G) of G, this equation has either zero [r(G) = 1], one [r(G) = M ], or infinitely many [0 < r(G) < M ]
solutions. The first case is unphysical, the second is unproblematic, and for the third every initial condition still determines a
corresponding final state in a unique way.
As a simple example meant to illustrate how we construct and solve the differential equation for the coherence vector, assume
that a single qubit is subject to a magnetic field along the z direction along with dephasing:
As a fixed operator basis satisfying the conditions in Eq. (405), we choose the Pauli matrices:
√
Fj = σ j / 2 , (439)
with the normalization due to the requirement that Tr(Fi Fj ) = δij . The Q matrix elements [Eq. (415)] are then
1
Qjk = iω Tr (Z[σj , σk ]) , (440)
2
and are non-vanishing only when [σj , σk ] ∝ Z, i.e., [X, Y ] = 2iZ and [Y, X] = −2iZ. Therefore Q12 = 2ω = −Q12 , and all
other Q matrix elements are zero.
Next, we need to calculate the R matrix and the c⃗ vector, using Eq. (420). Note that, in this case, only a33 = 2γ is non-zero in
the a-matrix of the Lindblad equation (the factor of 2 is due to the normalization of the F ’s). Therefore ∑ij reduces to just the
term with i = j = 3:
1
ck = γTr([Z, Z] σk ) = 0 (441a)
23/2 d
1 1 1
Rkl = Tr (Zσl Zσk − {Z 2 , σl } σk ) = γTr (Zσl Zσk − σl σk ) . (441b)
4 2 4
Clearly, σk must equal σl in order for the trace to be non-zero. When σk = σl = X, or when σk = σl = Y , we get 14 γTr(−I − I) =
−γ, whereas when σk = σl = Z we get 0. Thus R = diag(−γ, −γ, 0). Combining with the result for Q, we have:
⎛ −γ 2ω 0 ⎞
G=⎜ ⎟
⎜ −2ω −γ 0 ⎟ . (442)
⎝ 0 0 0⎠
This G matrix is diagonalizable but not invertible (its rank is 2), so we are in a scenario that is in between that of Secs. XIII D
and XIII E. Non-invertibility only affects the existence of the limit of v⃗(t) as t → ∞. Since G is diagonalizable, all its Jordan
blocks have dimension dj = 1, i.e., they are simply the eigenvalues. The eigenvalues are −γ ± 2iω and 0. This corresponds to
a coherence vector rotating at angular frequency 2ω in the X − Y plane, while exponentially decaying towards the Z axis with
rate γ. This means that the entire Z axis is the limit as t → ∞, hence there is no unique final state. However, every initial state
decays to a unique final state (its projection onto the Z axis).
XIV. DERIVATION OF THE LINDBLAD EQUATION FROM THE CUMULANT EXPANSION AND COARSE GRAINING
We now present a derivation of the Lindblad equation (LE) from first principles, following Ref. [17]. This derivation avoids
the so-called rotating wave approximation (RWA), which is the most commonly used approach to deriving the LE. We shall
return to an RWA-based approach later.
A. Cumulant expansion
H = HS + HB + λHSB (443)
with
HSB = A ⊗ B (444)
where A is a Hermitian system operator and B is a Hermitian bath operator. We have restricted ourself to a single term to
simplify the notation, but the more general case with multiple terms follows in an analogous fashion.
Define:
H0 ≡ HS ⊗ IB + IS ⊗ HB , (445a)
−itHS −itHB
U0 (t) ≡ exp (−itH0 ) = US (t) ⊗ UB (t) = e ⊗e , (445b)
ρ̃SB (t) ≡ U0† (t)ρSB (0)U0 (t) , (445c)
71
where ρ̃SB (t) is the state in the interaction picture (recall Sec. X B). We have the interaction picture Hamiltonian
H̃(t) = U0† (t)HSB U0 (t) = US† (t)AUS (t) ⊗ UB† (t)BUB (t) ≡ A(t) ⊗ B(t) . (446)
The density matrix in the interaction picture satisfies
d
ρ̃SB (t) = −i [λH̃(t), ρ̃SB (t)] , (447)
dt
which we can solve formally by integration followed by substitution and iteration:
t
ρ̃SB (t) = ρSB (0) − i ∫ ds [λH̃(s), ρ̃SB (s)] (448a)
0
t t s
= ρSB (0) − iλ ∫ ds [H̃(s), ρSB (0)] + (−iλ)2 ∫ ds ∫ ds′ [H̃(s), [H̃(s′ ), ρSB (0)]] + ⋯ , (448b)
0 0 0
and it is clear how this continues. A simple norm estimate (see Sec. ??) shows that the norm of the nth order term is
O[(∥HSB ∥t)n ]. Therefore a sufficient convergence condition is λ∥HSB ∥t < 1. Terms of third order and above can be ne-
glected provided λ∥HSB ∥t ≪ 1. This is known as the Born approximation.
We are interested in the reduced density matrix:
ρ̃(t) = TrB [ρ̃SB (t)] ≡ Λλ (t)ρ(0) . (449)
The cumulant expansion is given by introducing unknown, to be determined operators K (n) in the exponent:
∞
Λλ (t) = exp ( ∑ λn K (n) (t)) (450a)
n=1
1 2
= I + λK (1) (t) + λ2 (K (2) (t) + (K (1) (t)) ) + O(λ3 ) , (450b)
2
where in the second line we used a Taylor expansion of the exponential. We solve for K (n) by matching powers of λ with
Eq. (448). We get:
t
K (1) (t)ρ(0) = −i ∫ ds TrB ([H̃(s), ρSB (0)]) . (451)
0
We will see later that, without loss of generality, this can always be made to vanish (for a stationary bath) by shifting the operator
B, i.e:
K (1) (t)ρ(0) = 0 . (452)
The next order in λ gives:
t s
K (2) (t)ρ̃(0) = − ∫ ds ∫ ds′ TrB ([H̃(s), [H̃(s′ ), ρSB (0)]]) . (453)
0 0
It turns out to be convenient to express the interaction picture system operator A(t) in the frequency domain. To do so, let us
first expand HS in its eigenbasis:
A(t) = US† (t)AUS (t) = ∑ e−i(εb −εa )t ∣εa ⟩⟨εa ∣A∣εb ⟩⟨εb ∣ = ∑ Aω e−iωt , (459)
a,b ω
To clarify, the sum over εb − εa = ω in Eq. (460) is over all pairs of eigenenergies {εb , εa } whose difference gives the same Bohr
frequency ω. The sum over ω in Eq. (459) is a sum over all Bohr frequencies (negative, zero, and positive). This then gives the
following map from time 0 to t:
K (2) (t)ρ(0) = ∑ Bωω′ (t) (Aω ρ(0)A†ω′ − A†ω′ Aω ρ(0)) + h.c., (461)
ω,ω ′
where
t s
Bωω′ (t) ≡ ∫ ds′ ei(ω s−ωs ) B(s, s′ ) .
′ ′
ds ∫ (462)
0 0
We will see that Eq. (461) can be rewritten in the form of a Lindblad generator:
1
K (2) (t)ρ(0) = −i [Q(t), ρ(0)] + ∑ bωω′ (t) [Aω ρ(0)A†ω′ − {A†ω′ Aω , ρ(0)}] , (463)
ω,ω ′ 2
where
i
Qωω′ (t) = − (Bωω′ − Bω∗ ′ ω ) (466a)
2
i t s
= − ∫ ds ∫ ds′ (ei(ω s−ωs ) B(s, s′ ) − e−i(ωs−ω s ) B(s′ , s)) .
′ ′ ′ ′
(466b)
2 0 0
∗
Note that (Qωω′ ) = Qω′ ω , so that Q† = Q, as required for the interpretation of Q as a Hamiltonian.
We argued [Eq. (452)] that we can shift the bath operator B such that K (1) (t)ρ(0) = 0. Here we show why.
Let ρB (0) = ∑µ λµ ∣µ⟩⟨µ∣ and
i.e., the diagonal part of B in the eigenbasis of ρB (0). Here Bµµ (t) = ⟨µ∣ B(t) ∣µ⟩. Let us define a new bath operator
Then
⟨B ′ (t)⟩ = ⟨B(t)⟩ − ⟨Bd (t)⟩ = ∑ λµ ⟨µ∣ B(t) ∣µ⟩ − ∑ λµ ⟨µ∣ [∑ Bνν (t)∣ν⟩⟨ν∣] ∣µ⟩ = 0 . (469)
µ µ ν
′
Let HSB = A ⊗ B ′ , so that H̃ ′ (t) = U0† (t)HSB
′
U0 (t). Then
TrB ([H̃ ′ (t), ρSB (0)]) = TrB ([A(t) ⊗ B ′ (t), ρS (0) ⊗ ρB (0)]) = ⟨B ′ (t)⟩[A(t), ρS (0)] = 0 . (470)
Therefore, K ′(1) (t)ρ(0) = 0, with K ′(1) defined with the modified system-bath interaction HSB ′
. The price we have to pay
for this is the shift of B to B . This shift manifests itself only through the bath correlation function B(s, s′ ) [Eq. (455)]. The
′
shifted correlation function becomes B ′ (s, s′ ) = ⟨B † (s)B ′ (s′ )⟩, and nothing else changes, since the bath operators only appear
′
We will now prove that Eq. (461) can be transformed into Eq. (463). It turns out that the unequal upper integration limits
in Bωω′ [Eq. (462)] are problematic, while the equal upper integration limits in bωω′ [Eq. (464)] are what allows us to prove
complete positivity, as we show in Sec. XIV E directly below. To replace the unequal upper limits by equal limits we note the
following relations for the integral, where for notational simplicity we suppress the t-dependence for now:
t s t t t t
Bωω′ ≡ ∫ ds′ ei(ω s−ωs ) B(s, s′ ) , = [∫ ds′ − ∫ ds′ ] ei(ω s−ωs ) B(s, s′ ) ,
′ ′ ′ ′
ds ∫ ds ∫ ds ∫ (471a)
0 0 0 0 0 s
t t t s′
= [∫ ds′ − ∫ ds′ ∫ ds] ei(ω s−ωs ) B(s, s′ ) ,
′ ′
ds ∫ (471b)
0 0 0 0
t t t s
=∫ ds′ ei(ω s−ωs ) B(s, s′ ) − ∫ ds′ ei(ω s −ωs) B(s′ , s) ,
′ ′ ′ ′
ds ∫ ds ∫ (471c)
0 0 0 0
= bωω′ − Bω∗ ′ ω , (471d)
where bωω′ [Eq. (464)] has the desired equal upper integration limits. It follows immediately that
∗
Bωω ′ = bω ′ ω − B ω ′ ω . (472)
where in the second term on the RHS we have switched ω ↔ ω ′ , which is permissible since we are summing over all ω and ω ′ .
Furthermore, this second term is now exactly in the form of the original term, so we have the result:
∗ 1
∑ [Bωω′ Aω ρAω′ + Bωω′ Aω′ ρAω ] = ∑ [bωω′ Aω ρAω′ + bω′ ω Aω′ ρAω ] = ∑ bωω′ Aω ρAω′ .
† † † † †
(474)
ω,ω ′ 2 ω,ω′ ω,ω ′
The second summand +h.c. in Eq. (461) is of the form −A†ω′ Aω ρ(0), which reminds us of the anti-commutator term in the
Lindblad equation, except that it doesn’t have the factor of 1/2. However, note that since b∗ω′ ω = bωω′ = Bωω
∗
′ + Bω ′ ω , where we
∗
used Eq. (472). Therefore by writing bωω′ = 2 (bωω′ + bω′ ω ), and again using Eq. (472), we have:
1
1 1
Bωω′ = bωω′ + (Bωω′ − Bω∗ ′ ω ) . (475)
2 2
74
This allows us to write the second summand +h.c. in Eq. (461) as:
∗
− ∑ [Bωω′ A†ω′ Aω ρ + Bωω ′ ρAω Aω ′ ]
†
(476a)
ω,ω ′
1 1 ∗ ∗
=− ∑ (bωω′ Aω′ Aω ρ + bω′ ω ρAω Aω′ ) − ∑ [(Bωω′ − Bω′ ω ) Aω′ Aω ρ + (Bω′ ω − Bω′ ω ) ρAω Aω′ ]
† † † †
(476b)
2 ω,ω′ 2 ω,ω′
1 1 ∗
=− ∑ bωω′ (Aω′ Aω ρ + ρAω′ Aω ) − ∑ (Bωω′ − Bω′ ω ) [Aω′ Aω ρ − ρAω′ Aω ]
† † † †
(476c)
2 ω,ω′ 2 ω,ω′
1 1 ∗
=− ∑ bωω′ {Aω′ Aω , ρ} − ∑ (Bωω′ − Bω′ ω ) [Aω′ Aω , ρ] .
† †
(476d)
2 ω,ω′ 2 ω,ω′
We can now write the RHS of Eq. (461) as:
∑ Bωω′ (t) (Aω ρ(0)Aω′ − Aω′ Aω ρ(0)) + h.c.
† †
(477a)
ω,ω ′
1 1 ∗
= ∑ bωω′ Aω ρ(0)A†ω′ − ∑ bωω′ {Aω′ Aω , ρ(0)} − ∑ (Bωω′ − Bω′ ω ) [Aω′ Aω , ρ(0)]
† †
(477b)
ω,ω ′ 2 ω,ω′ 2 ω,ω′
−i 1
= −i[ ∑ (Bωω′ − Bω∗ ′ ω ) [A†ω′ Aω , ρ(0)]] + ∑ bωω′ (t)[Aω ρ(0)A†ω′ − {A†ω′ Aω , ρ(0)} ] , (477c)
ω,ω ′ 2 ω,ω ′ 2
which is Eq. (463), together with the identification of the term in the commutator as the Lamb shift Q(t) as defined in Eq. (465).
E. Complete positivity
Clearly, the dissipative (second) term on the RHS of Eq. (463) appears to be in Lindblad form, but we must still prove the
positivity of the matrix b(t). To this end we again expand the bath density matrix in its eigenbasis, and use this to write the
correlation function B(s, s′ ) = ⟨B † (s)B(s′ )⟩B explicitly. Let v⃗ be some arbitrary vector; then positivity amounts to showing
that v⃗b(t)⃗
v † ≥ 0 for all v⃗. Indeed:
t t
v † = ∑ vω bωω′ (t)vω∗ ′ = ∫
v⃗b(t)⃗ ds′ ∑(vω e−iωs ) ∑(vω′ e−iω s )∗ Tr[∑ λµ ∣µ⟩ ⟨µ∣ B † (s)B(s′ )]
′ ′
ds ∫ (478a)
ωω ′ 0 0 ω ω′ µ
where F (t) ≡ ∫0 dsB(s) ∑ω vω∗ e−iωs . Note how it was crucial in this proof that the upper limits of the integrals are the same,
t
associated advantages (e.g., a quantum trajectories unravelling) we need to introduce an additional, Markovian approximation.
Let us show how to obtain the LE from the results above. Expanding the exponential in Eq. (479) to second order in λ, we
have:
1
ρ̃(t) − ρ̃(0) = −i [λ2 Q(t), ρ(0)] + ∑ λ2 bωω′ (t) [Aω ρ(0)A†ω′ − {A†ω′ Aω , ρ(0)}] . (480)
ω,ω ′ 2
It is straightforward to check that Q(0) = bωω′ (0) = 0 (due to the upper integration limit being 0). Therefore, dividing both sides
of Eq. (480) by τ , and setting t = τ , we have:
which implies that ρB (t) = UB (t)ρB (0)UB† (t) = ρB (0). This is the case, e.g., if ρB (0) = e−βHB /Z, i.e., is a Gibbs state. In
addition we assume that the bath correlation function decays over a timescale τB , i.e.,
while the coarse graining is done over a much longer timescale, so that the integrand in Eq. (464) has already decayed. The RHS
of Eq. (518) is then valid for all times, allowing us to also shift the time argument of ρ to arbitrary jτ . Let us now define the
Lamb-shift and the Lindblad rates as:
Moreover, we assume that τ is very small on the timescale τS over which ρ(t) changes, so that ⟨ρ̃⟩
˙ j = [ρ̃((j + 1)τ ) − ρ̃(jτ )]/τ
can be replaced by ρ̇(t). These assumptions can be summarized as
τB ≪ τ ≪ τS . (485)
We can thus write the interaction picture Lindblad equation in the final form:
1
˙
ρ̃(t) = −i [HLS , ρ(t)] + ∑ γωω′ [Aω ρ(t)A†ω′ − {A†ω′ Aω , ρ(t)}] . (486)
ω,ω ′ 2
The RHS contains the free parameter τ , which can be determined using Eq. (485). Everything else is determined in terms of the
given specification of the Hamiltonian H = HS + HB + HSB and the initial state of the bath ρB (0). In particular,
• The Bohr frequencies ω are determined by HS ;
• The Lindblad operators are determined by the system operator A in HSB and the Bohr frequencies (i.e., HS );
• The bath correlation function B(s, s′ ) is determined by the bath operator B in HSB , the bath Hamiltonian HB (which
determines the time-dependence of B(t)), and the initial bath state ρB ;
• The Lamb shift is determined by the bath correlation function and the Bohr frequencies.
Consider once more the spin-boson model defined in Sec. X C. Let us denote the eigenvalues of HS = −(g/2)Z by ε± = ±g/2
and their respective eigenvectors by ∣ε− ⟩ = ∣0⟩ (ground state) and ∣ε+ ⟩ = ∣1⟩ (excited state). Using Eq. (460), the Lindblad
operators are then given by:
A−g = ∣ε+ ⟩⟨ε+ ∣Z∣ε− ⟩⟨ε− ∣ = 0 (487)
A0 = ∣ε+ ⟩⟨ε+ ∣Z∣ε+ ⟩⟨ε+ ∣ + ∣ε− ⟩⟨ε− ∣Z∣ε− ⟩⟨ε− ∣ = Z (488)
Ag = ∣ε− ⟩⟨ε− ∣Z∣ε+ ⟩⟨ε+ ∣ = 0 . (489)
Thus, only the (elastic, or on-shell) ω = 0 term contributes to the sums over ω. This means that the Lamb shift is given by:
λ2 λ2
HLS = Q(τ ) = Q00 (t)A†0 A0 ∝ I , (490)
τ τ
so that [HLS , ρ̃(t)] = 0. The dissipative part of the LE [Eq. (486)] is given by:
1 1
∑ γω,ω′ [Aω ρAω′ −
†
{A†ω′ Aω , ρ}] = γ00 (Z ρ̃Z − {I, ρ̃}) = γ (Z ρ̃Z − ρ̃) , (491)
ω,ω ′ 2 2
76
where
λ2 λ2 τ τ
′ ′
γ ≡ γ00 = b00 (τ ) = ∫ ds ∫ ds B(s, s ) , (492)
τ τ 0 0
and where we used Eq. (464). We already computed this decay rate when we solved the spin-boson model analytically, and
found it in Eq. (374). The result after coarse graining is given in Eq. (333).
While we already saw the solution of the corresponding LE in Sec. IX D 2, let us solve it again using a nice and useful
“vectorization” trick. Let us define:
⎛ col1 (ρ) ⎞
⎜ ⋮ ⎟
vec(ρ) ≡ ⎜
⎜
⎟
⎟ (493)
⎜ colj (ρ) ⎟
⎝ ⋮ ⎠
i.e., vec(ρ) corresponds to stacking the columns of ρ (in some basis). We now use the identity [18]:
⎛1 ⎞ ⎛ ρ00 (0) ⎞
⎜ exp(−2γt) ⎟ ⎜ ρ10 (0) ⎟
vec(ρ̃(t)) = exp(L)vec(ρ̃(0)) = ⎜
⎜
⎟⎜
⎟⎜
⎟
⎟ (496)
⎜ exp(−2γt) ⎟ ⎜ ρ01 (0) ⎟
⎝ 1 ⎠ ⎝ ρ11 (0) ⎠
Therefore, we find as before:
XV. FIRST-PRINCIPLES DERIVATION OF THE LINDBLAD EQUATION FROM THE BORN, MARKOV, AND ROTATING
WAVE APPROXIMATIONS
We now present our last derivation of the Lindblad equation. This is the standard approach found in textbooks such as [1], but
we will add some clarifications concerning the limitations of the validity of this approach. We will also discuss the differences
between this and the cumulant-based approach.
A. Setting up
Our starting point is identical to the one we used in the cumulant expansion approach (Sec. XIV A). The only difference is
that we now consider the more general system-bath interaction
HSB = g ∑ Aα ⊗ Bα , (499)
α
and
Let us now substitute this solution back into Eq. (502) and take the partial trace:
d d t
ρ̃(t) = TrB { ρ̃SB (t)} = −iTrB {[H̃(t), ρSB (0)]} + (−i)2 TrB {[H̃(t), ∫ ds [H̃(s), ρ̃SB (s)]]} . (504)
dt dt 0
Just as we argued in Sec. XIV C, the first order term can again be made to vanish provided we shift the bath operators. We are
thus left with
d t
ρ̃(t) = −TrB {[H̃(t), ∫ ds [H̃(s), ρ̃SB (s)]]} . (505)
dt 0
t 0 t
Let us change variables to τ = t − s, so that ∫0 ds = − ∫t (−dτ ) = ∫0 dτ , and:
d t
ρ̃(t) = −TrB {[H̃(t), ∫ dτ [H̃(t − τ ), ρ̃SB (t − τ )]]} . (506)
dt 0
B. Born approximation
To proceed we now make our first approximation. For a sufficiently large bath that is in particular much larger than the system,
it is reasonable to assume that while the system undergoes non-trivial evolution, the bath remains unaffected, and hence that the
state of the composite system at time t is
where ρB is the time-independent, stationary bath state, and the correlations χ(t) can be neglected. This is (again) called the
Born approximation.
Using this and Eq. (500), we have:
dρ̃ t
= −g 2 ∑ TrB {[Aα (t) ⊗ Bα (t), ∫ dτ [Aβ (t − τ ) ⊗ Bβ (t − τ ), ρ̃(t − τ ) ⊗ ρB ]]} . (508)
dt α,β 0
We now assume again that the bath is stationary (i.e., [ρB , HB ] = 0). As in Eq. (456), let ⟨X⟩B ≡ Tr[ρB X]. Similarly to
Eq. (457), we define the bath two-point correlation function:
Bαβ (t, t − τ ) ≡ ⟨Bα (t)Bβ (t − τ )⟩B = Tr(eiHB t Bα e−iHB t eiHB (t−τ ) Bβ e−iHB (t−τ ) ρB ) (510a)
−iHB (t−τ ) iHB t −iHB t iHB (t−τ ) −iHB τ
= Tr(e e Bα e e Bβ ρB ) = Tr(e iHB τ
Bα e Bβ ρB ) = ⟨Bα (τ )Bβ ⟩B (510b)
= Bαβ (τ, 0) ≡ Bαβ (τ ) , (510c)
78
where we used the bath stationarity assumption to go the second line, and in the third line we denoted Bαβ (τ, 0) by Bαβ (τ )
for simplicity, since only the time shift τ matters, so we can measure everything from t = 0. Thus, Bαβ (τ ) measures the
autocorrelation of the bath after time τ . Note that in the Bαβ (τ ) notation we implicitly associate t = τ with the first index (in
this case α), whereas the second index is associated with t = 0. Also,
∗
Bβα (τ ) = Tr[(ρB Bβ (τ )Bα )† ] = Tr[Bα (UB† (τ )Bβ UB (τ ))† ρB ] = Tr[ρB Bα UB† (τ )Bβ UB (τ )] = Bαβ (0, τ ) (511a)
= Tr[ρB UB (τ )Bα UB† (τ )Bβ ] = Bαβ (−τ ) . (511b)
Then, noting that the terms in lines (509b) and (509e) are Hermitian conjugates, as are the terms in lines (509c) and (509d),
we have:
dρ̃ t
= −g 2 ∑ ∫ dτ {Bαβ (τ ) [Aα (t), Aβ (t − τ )ρ̃(t − τ )] + h.c.} . (512)
dt αβ 0
Note that the RHS of Eq. (512) depends on the entire history of the system state, since the argument of ρ̃(t − τ ) ranges from t
to 0 as τ increases from the lower to the upper limit of the integral. Thus, Eq. (512) is time-nonlocal. We would like to arrive at
a time-local differential equation for the system state, which depends only on t, but not on the state’s history.
To attain this, at this point we need to introduce our second approximation, the Markov approximation. Informally, it states that
the bath has a very short correlation time τB , i.e., that the correlation function Bαβ (τ ) decays rapidly with some characteristic
timescale τB , e.g., ∣Bαβ (τ )∣ ∼ e−τ /τB . We also assume that
g ≪ 1/τB , t ≫ τB . (513)
The first of these is a weak-coupling limit (g is small), and the second states that we do not expect our approximation to be
accurate for times t that are comparable to the bath correlation time (instead, we only consider times much larger than the latter).
Now, since the correlation function Bαβ (τ ) is essentially zero for τ ≫ τB , and since we assume that t ≫ τB , we can replace
ρ̃(t − τ ) by ρ̃(t), since the short “memory” of the bath correlation function causes it to keep track of events only within the short
period [0, τB ]. Under this approximations, Eq. (512) becomes:
dρ̃ t
= −g 2 ∑ ∫ dτ {Bαβ (τ )[Aα (t), Aβ (t − τ )ρ̃(t)] + h.c.} , (514)
dt α,β 0
which is known as the Redfield equation. It is notoriously non-CP, which means that the density matrix can be become non-
positive (though various fixes have been proposed [19, 20]).
Moreover, for the same reason (correlation function negligible for τ ≫ τB ) we can extend the upper limit of the integral to
infinity without changing the value of the integral.
dρ̃ ∞
= −g 2 ∑ ∫ dτ {Bαβ (τ )[Aα (t), Aβ (t − τ )ρ̃(t)] + h.c.} + O(g 4 τB3 ) , (515)
dt α,β 0
That Eq. (512) can be replaced by Eq. (515) can be proven rigorously under the following sufficient condition [21], as we will
show in Sec. XVIII D:
∞
∫ τ n ∣Bαβ (τ )∣dτ ∼ τBn+1 , n ∈ {0, 1, 2} . (516)
0
∞ 1 n + 1 n+1
∫ τ n ∣Bαβ (τ )∣dτ = Γ( ) τB , (518)
0 k k
where Γ(x) is the gamma function [recall that Γ(n + 1) = n! for n ∈ N]. Thus, in fact even a subexponential (k < 1) decay will
suffice.
Note that thanks to Eq. (516), the integral in Eq. (515) is of order τB . Thus the ratio between the leading order correction and
the integral is (g 4 τB3 )/(g 2 τB ) = (gτB )2 ≪ 1, by our assumption that gτB ≪ 1.
79
After dropping the correction term, Eq. (515) is now a differential equation for ρ̃(t), but is not yet in Lindblad form. To
convert it into this form we once again convert the system operators A(t) to the frequency domain. The procedure is essentially
the same as in Sec. XIV B, except that we need to keep track of the system operator index as well. Thus, after expanding HS in
its eigenbasis as HS = ∑a εa ∣εa ⟩⟨εa ∣, we have
Aα (t) = US† (t)Aα US (t) = ∑ e−i(εb −εa )t ∣εa ⟩⟨εa ∣Aα ∣εb ⟩⟨εb ∣ = ∑ Aα (ω)e−iωt , (519)
a,b ω
where the last equality follows since Hermitian conjugation interchanges εa and εb . Also, note that since Aα (t) is Hermitian,
−iωt
∑ Aα (ω)e = ∑ A†α (ω)eiωt . (521)
ω ω
Returning to Eq. (515), consider the two terms in the commutator [Aα (t), Aβ (t − τ )ρ̃(t)]:
Aα (t)Aβ (t − τ )ρ̃(t) = ∑ eiω t e−iω(t−τ ) A†α (ω ′ )Aβ (ω)ρ̃(t) = ∑ eiωτ ei(ω −ω)t A†α (ω ′ )Aβ (ω)ρ̃(t)
′ ′
(522a)
ωω ′ ωω ′
The entire τ -dependence is thus in the factor eiωτ , which motivates collecting everything that is τ -dependent in Eq. (515) into
one function:
∞
Γαβ (ω) ≡ ∫ dτ eiωτ Bαβ (τ ) , (523)
0
which is the one-sided Fourier transform of the bath correlation function. This allows us to rewrite Eq. (515) as
dρ̃
= −g 2 ∑ ∑ {Γαβ (ω)ei(ω −ω)t [A†α (ω ′ ), Aβ (ω)ρ̃(t)]} + h.c.
′
(524)
dt α,β ω,ω ′
Note that Γ as defined here has dimensions of time, and g 2 Γ has units of frequency.
Alas, Eq. (524) is still not in Lindblad form. The problem is the “non-secular” (off-diagonal) terms with ω ≠ ω ′ . While
these did not present a problem in the cumulant derivation (recall that we proved complete positivity in Sec. XIV E), they do
now. Therefore we next introduce the final approximation, known as the rotating wave approximation (RWA), sometimes also
called the secular approximation. This approximation is based on the idea that the terms with ω ≠ ω ′ in Eq. (524) are rapidly
oscillating if t ≫ ∣ω − ω ′ ∣−1 , which thus (roughly) average to zero. Since we already assumed that t ≫ τB , the former assumption
is consistent provided we also assume that the Bohr frequency differences satisfy
Note that this means that also the Bohr frequencies themselves (by setting ω ′ = 0) must be large compared to the inverse of the
bath correlation time, and this therefore excludes the treatment of systems with gaps that are small relative to 1/τB (this has
implications for the applicability to systems that are typically of interest in adiabatic quantum computing, for example). Also
note that, combining this with the previous assumption [Eq. (513)], we get:
This shows that the coupling also lower bounds the Bohr frequencies.
80
Let
∞
γαβ (ω) = ∫ eiωτ Bαβ (τ )dτ , (527)
−∞
i.e., the full Fourier transform of the bath correlation function. Using Eq. (511):
∞ ∞
∗
γαβ (ω) = ∫ e−iωτ Bβα (−τ )dτ = ∫ eiωτ Bβα (τ )dτ = γβα (ω) , (528)
−∞ −∞
1 f (x)
P ( ) [f ] = lim ∫ dx , (532)
x →0 − x
for smooth functions f with compact support on the real line R. Substituting Eq. (531) into Eq. (530), we can thus write
1
Γαβ (ω) = γαβ (ω) + iSαβ (ω) , (533)
2
where
1 ∞ 1
Sαβ (ω) = ∫ γαβ (ω ′ )P ( ) dω ′ = Sβα
∗
(ω) , (534)
2π −∞ ω − ω′
and we used the fact that γ is Hermitian in the last equality. Therefore:
1
γαβ (ω) = Γαβ (ω) + Γ∗βα (ω) , Sαβ (ω) = (Γαβ (ω) − Γ∗βα (ω)) . (535)
2i
Finally, we will show in Sec. XV F 1 that by introducing Eq. (533) and the RWA into Eq. (524), we arrive at the interaction
picture Lindblad equation:
dρ̃ 1
= −i[HLS , ρ̃(t)] + g 2 ∑ ∑ γαβ (ω)(Aβ (ω)ρ̃(t)A†α (ω) − {A†α (ω)Aβ (ω), ρ̃(t)}) , (536)
dt ω αβ 2
[HLS , HS ] = 0 . (539)
81
Note that γαβ (ω) and Sαβ (ω) as defined in Eqs. (527) and (534) have dimensions of time, while g 2 γαβ (ω) and g 2 Sαβ (ω) have
units of frequency. The factor g 2 can always be reabsorbed into the definition of γαβ (ω) and Sαβ (ω).10
We will show in Sec. XV F 3 that we can transform back to Schrödinger picture via ρ(t) = US (t)ρ̃(t)US† (t) and thus finally
obtain the RWA-LE:
dρ 1
= −i[HS + HLS , ρ] + g 2 ∑ ∑ γαβ (ω)(Aβ (ω)ρA†α (ω) − {A†α (ω)Aβ (ω), ρ}) . (540)
dt ω αβ 2
Let us start by expanding the commutator and the Hermitian conjugate term in Eq. (524), relabelling indices, and combining
terms. This gives us for the summands:
ei(ω −ω)t (Γαβ (ω)A†α (ω ′ )Aβ (ω)ρ̃(t) + Γ∗βα (ω ′ )ρ̃(t)A†α (ω ′ )Aβ (ω)) − ei(ω −ω)t (Γαβ (ω) + Γ∗βα (ω ′ )) Aβ (ω)ρ̃(t)A†α (ω ′ ) .
′ ′
(542)
Applying the RWA (i.e., setting ω = ω ′ ) and substituting Γαβ (ω) = 21 γαβ (ω) + iSαβ (ω), this becomes:
∗
1
γ (ω)A†α (ω)Aβ (ω)ρ̃(t) + 12 γβα
2 αβ
(ω)ρ̃(t)A†α (ω)Aβ (ω) (543a)
∗
+ iSαβ (ω)A†α (ω)Aβ (ω)ρ̃(t) − iSβα (ω)ρ̃(t)A†α (ω)Aβ (ω) − γαβ (ω)Aβ (ω)ρ̃(t)A†α (ω) . (543b)
Since γ(ω) and S(ω) are Hermitian this becomes:
1
γ (ω) (A†α (ω)Aβ (ω)ρ̃(t) + ρ̃(t)A†α (ω)Aβ (ω))
2 αβ
(544a)
+ iSαβ (ω) (A†α (ω)Aβ (ω)ρ̃(t) − ρ̃(t)A†α (ω)Aβ (ω)) − γαβ (ω)Aβ (ω)ρ̃(t)A†α (ω) (544b)
= 21 γαβ (ω) {A†α (ω)Aβ (ω), ρ̃(t)} + iSαβ (ω) [A†α (ω)Aβ (ω), ρ̃(t)] − γαβ (ω)Aβ (ω)ρ̃(t)A†α (ω) . (544c)
Putting this back into the original sum in Eq. (524) then gives us our desired result:
d ρ̃
= −ig 2 ∑ Sαβ (ω) [A†α (ω)Aβ (ω), ρ̃(t)] + g 2 ∑ γαβ (ω) (Aβ (ω)ρ̃(t)A†α (ω) − 12 {A†α (ω)Aβ (ω), ρ̃(t)}) (545a)
dt α,β,ω α,β,ω
= −i [HLS , ρ̃(t)] + g 2 ∑ γαβ (ω) (Aβ (ω)ρ̃(t)A†α (ω) − 21 {A†α (ω)Aβ (ω), ρ̃(t)}) . (545b)
α,β,ω
where the projectors Π(εa ) = ∣εa ⟩⟨εa ∣ are the eigenprojectors of HS , i.e.,
HS = ∑ εa Π(εa ) , (547)
a
10 Also note that in our derivation of the LE using coarse graining (Sec. XIV) we did not include the coupling strength g. Instead we used a dimensionless
parameter λ when we wrote down the system-bath interaction as λHSB , where HSB has dimensions of energy. As a result, γωω′ in the CG-LE has units of
̵ = 1).
frequency, while as noted above, in the RWA-LE γ(ω) has units of time, and g 2 γ(ω) has units of frequency (or energy, since we’re using units where h
82
HS A†α (ω)Aβ (ω) = ∑ εa Π(εa ) ∑ Π(εi )A†α Π(εj ) ∑ Π(εl )Aβ Π(εk ) (548a)
a εi −εj =ω εk −εl =ω
and similarly:
A†α (ω)Aβ (ω)HS = ∑ Π(εi )A†α Π(εj ) ∑ εk Π(εl )Aβ Π(εk ) . (549)
εi −εj =ω εk −εl =ω
It follows that
[HS , A†α (ω)Aβ (ω)] = ∑ (εi − εk ) Π(εi )A†α Π(εj )Π(εl )Aβ Π(εk ) (550a)
εi −εj =ω
εk −εl =ω
=0, (550c)
where the second line follows from the product of the two inner projection operators, and the third line from the summation
conditions, which set εi = εk . Consequently:
dρ d ρ̃
= −i [HS , ρ] + e−iHS t eiHS t . (552)
dt dt
Also, using Eq. (547) again:
Thus:
e−iHS t Aβ (ω)ρ̃A†α (ω)eiHS t = ∑ e−iHS t Π(εj )Aβ Π(εi )eiHS t ρ(t)e−iHS t Π(εk )A†α Π(εl )eiHS t (554a)
εi −εj =ω
εk −εl =ω
= ∑ ei(−εj +εi −εk +εl )t Π(εj )Aβ Π(εi )ρ(t)Π(εk )A†α Π(εl ) (554b)
εi −εj =ω
εk −εl =ω
and
e−iHS t A†α (ω)Aβ (ω)ρ̃(t)eiHS t = ∑ e−iHS t Π(εi )A†α Π(εj )Π(εl )Aβ Π(εk )eiHS t ρ(t)e−iHS t eiHS t (555a)
εi −εj =ω
εk −εl =ω
= ∑ ei(−εi +εk )t Π(εi )A†α Π(εj )Π(εl )Aβ Π(εk )ρ(t) (555b)
εi −εj =ω
εk −εl =ω
= ∑ ei(−εj +εl )t Π(εi )A†α Π(εj )Π(εl )Aβ Π(εk )ρ(t) = A†α (ω)Aβ (ω)ρ(t) , (555c)
εi −εj =ω
εk −εl =ω
83
and similarly for the second term in the anti-commutator. This shows that
1 1
e−iHS t (Aβ (ω)ρ̃(t)A†α (ω) − {A†α (ω), Aβ (ω), ρ̃(t)})eiHS t = (Aβ (ω)ρ(t)A†α (ω) − {A†α (ω), Aβ (ω), ρ(t)}) . (556)
2 2
Now, since we showed that HS and HLS commute:
e−iHS t [HLS , ρ̃(t)]eiHS t = e−iHS t HLS eiHS t e−iHS t ρ̃(t)eiHS t − e−iHS t ρ̃(t)eiHS t e−iHS t HLS eiHS t (557a)
= [HLS , ρ(t)] . (557b)
Hence, using Eqs. (536) and (552) we obtain Eq. (540) as required.
1 T T
I(ω, T ) ≡ ∫ dt ∫ e
iω(t−s)
Bαβ (t − s)ds . (559)
T 0 0
First, we change the variables from (t, s) to (u, s) with u = t − s. For every value of s, sweeping t from 0 to T will yield a
horizontal line of length T in the (u, s) plane. The new integration region is therefore a parallelogram in the variables (u, s),
as illustrated in Fig. 6. We can split this region into u ∈ [−T, 0] and u ∈ [0, T ], and perform the integration over s first. As is
clear from the figure, s varies from −u to T in the u ∈ [−T, 0] region, and from 0 to T − u in the u ∈ [0, T ] region. The area is
preserved so the Jacobian yields 1. Consequently,
T T 0 T T T −u
∫ ds ∫ dt = ∫ du ∫ ds + ∫ du ∫ ds . (560)
0 0 −T −u 0 0
1 T
I(ω, T ) = ∫ e Bαβ (u)(T − ∣u∣)du
iωu
(562a)
T −T
T 1 T
= ∫ eiωu Bαβ (u)du − ∫ eiωu Bαβ (u)∣u∣du . (562b)
−T T −T
84
∞
Now recall that in the Markov approximation we assumed [Eq. (516)] that ∫0 un ∣Bαβ (u)∣du ∼ τBn+1 , where τB < ∞ is the bath
correlation time. Therefore, using Eq. (510c):
T T 0
∫ eiωu ∣u∣Bαβ (u)du = ∫ eiωu uBαβ (u)du − ∫ eiωu uBαβ (u)du (563a)
−T 0 −T
T T
=∫ eiωu uBαβ (u)du + ∫ e−iωu uBβα
∗
(u)du (563b)
0 0
∞ ∞
∗
≤∫ u∣Bαβ (u)∣du + ∫ u∣Bβα (u)∣du ∼ 2τB2 . (563c)
0 0
T
∫−T e ∣u∣Bαβ (u)du = 0, and
1 iωu
Consequently limT →∞ T
as claimed.
t
Now, for any vector v = (v1 , v2 , . . . ) we have
+∞
v † γ(ω)v = ∑ vα∗ γαβ (ω)vβ = ∑ vα∗ vβ ∫ eiωu Bαβ (u)du (565a)
αβ αβ −∞
1 T T
∗
= lim ∑ vα vβ ∫ dt ∫ e
iω(t−s)
Bαβ (t − s)ds (565b)
T →∞ T αβ 0 0
1 T T
∗ iωt −iωs
= lim ∑ λµ ⟨µ∣ ∫ vα e Bα (t)dt ∫ vβ e Bβ (s)ds∣µ⟩ (565c)
T →∞ T αβ,µ 0 0
2
1 T
−iωs
= lim ∑ λµ ∣∑ ∫ vα Bα (s)e ds∣µ⟩∣ ≥ 0 . (565d)
T →∞ T µ α 0
Therefore γ(ω) ≥ 0.
b. Second proof The following proof uses Bochner’s theorem as suggested, e.g., in the textbook [1].
Since γ(ω) is Hermitian [Eq. (528)] we can diagonalize it using a unitary transformation:
∗
D ≡ U γU † ⇒ Dαβ = ∑ Uαi γij Uβj . (566)
i,j
∞
D is diagonal so we need only consider the diagonal elements (i.e., the eigenvalues of γ). Plugging in γij = ∫−∞ eiωs Bij (s)ds
gives
∞ ⎛ ∗ ⎞
Dα = ∫ eiωs ∑ Uαi Bij (s)Uαj ds . (567)
−∞ ⎝ i,j ⎠
We wish to show that Dα is non-negative for each α. To do this we must consider the function in parenthesis. Dα is the Fourier
transform of this function so if we can show that it is of positive type then Dα must be positive by Bochner’s theorem [22].
Define the following function with {ti } an arbitrary time partition:
∗
α
fmn ≡ ∑ Uαi Bij (tm − tn )Uαj . (568)
i,j
Now use the property ⟨Bα (s)Bβ (0)⟩ = ⟨Bα (t)Bβ (t − s)⟩ [Eq. (510c)] to write fmn
α
as
∗ ⎛ ∗ ⎞
α
fmn = ∑ Uαi Tr [ρB Bi (tm )Bj (tn )] Uαj = Tr ρB ∑ Uαi Bi (tm ) ∑ Uαj Bj (tn ) . (569)
i,j ⎝ i j ⎠
We need to show that f α is a positive matrix. For arbitrary ∣v⟩ we have
⎡ ⎤
⎢⎛ √ ⎞⎛ √ ⎞⎥⎥
∗ ⎢ ∗ ∗
⟨v∣ f ∣v⟩ = ∑ vm vn fmn = Tr ⎢ ∑ vm Uαi ρB Bi (tm ) ∑ vn Uαj Bj (tn ) ρB ⎥
α α
(570a)
⎢⎝i,m ⎠ ⎝j,n ⎠⎥
m,n
⎣ ⎦
⎡ ⎤
⎢⎛ √ ⎞⎛ √ ⎞ ⎥
= Tr ⎢⎢ ∑ vm ∗
Uαi ρB Bi (tm ) ∑ vm Uαi ∗
Bi (tm ) ρB ⎥⎥ (570b)
⎢⎝i,m ⎠ ⎝i,m ⎠⎥
⎣ ⎦
≡ Tr (Mα† Mα ) ≥ 0 , (570c)
85
where the final inequality follows from the fact that Mα† Mα is non-negative which follows immediately from right polar decom-
posing Mα (then Mα† Mα = RU † U R = R2 ≥ 0).
We have established that ⟨v∣ f α ∣v⟩ ≥ 0 for any time partition {ti }. Therefore Dα is positive by Bochner’s theorem. Conse-
quently, γ is a positive matrix since all its eigenvalues are non-negative.
XVI. THE KUBO-MARTIN-SCHWINGER (KMS) CONDITION AND THE GIBBS STATE AS A STATIONARY STATE
LINDBLAD EQUATION
In this section we formalize the folklore notion that “systems like to relax into lower energy states”, and that systems “tend to
equilibrate”.
Consider a general system-bath Hamiltonian of the form HSB = ∑a Aa ⊗ Ba (we’re using a and b since we’ll reserve β
for the inverse temperature in this subsection). Let us assume again that the bath state is stationary [Eq. (482)], which as we
saw implies that ρB (t) = UB (t)ρB (0)UB† (t) = ρB (0) ≡ ρB . We also saw that this means that the bath correlation function is
time-translation-invariant:
where for notational simplicity we dropped the B subscript we used before in ⟨X⟩B = Tr[ρB X].
If we assume not only that the bath state is stationary, but that it is also in thermal equilibrium at inverse temperature β, i.e.,
ρB = e−βHB /Z, then it follows that the correlation function satisfies the Kubo-Martin-Schwinger (KMS) condition [1]:
If in addition the correlation function is analytic in the strip between τ = −iβ and τ = 0, then it follows that the Fourier
transform of the bath correlation function satisfies the frequency domain KMS condition:
τ
0
−iβ
This is an extremely important condition, which is used in proving “detailed balance”, as we shall see when we discuss the Pauli
master equation, in Sec. XVII.
To prove this let us use the time-domain KMS condition, Eq. (574):
∞ ∞
γab (ω) = ∫ dτ eiωτ ⟨Ba (τ )Bb (0)⟩ = ∫ dτ eiωτ ⟨Bb (−τ − iβ)Ba (0)⟩ (576)
−∞ −∞
To perform this integral we replace it with a contour integral in the complex τ plane, ∮C dτ eiωτ ⟨Bb (−τ − iβ)Ba (0)⟩, with the
contour C as shown in Fig. 7. This contour integral vanishes by the Cauchy-Goursat theorem [23] since the closed contour
encloses no poles (by assumption, the correlation function ⟨Bb (τ )Ba (0)⟩ is analytic in the open strip (0, −iβ) and is continuous
at the boundary of the strip [24]), so that
∮ (. . . ) = 0 = ∫ (. . . ) + ∫ (. . . ) + ∫ (. . . ) + ∫ (. . . ) (577)
C ↑ ↓ → ←
where (. . . ) is the integrand of Eq. (576), and the integral ∫→ is the same as in Eq. (576). After making the variable transformation
τ = −x − iβ, where x is real, we have
∞
∫ (. . . ) = −e ∫
βω
dx e−iωx ⟨Bb (x)Ba ⟩ = −eβω γba (−ω) . (578)
← −∞
Assuming that ⟨Ba (±∞ − iβ)Bb (0)⟩ = 0 (i.e., the correlation function vanishes at infinite time), we further have ∫↑ (. . . ) =
∫↓ (. . . ) = 0, and hence we find the result:
Consider a bath at inverse temperature β. We would like to show that the system Gibbs state
⎛ e−βε0 ⎞
1 −βHS 1 ⎜ ⎟,
ρG = e = ⎜ e−βε1 ⎟ Z = Tr[e−βHS ] (580)
Z Z
⎝ ⋱ ⎠
is always a stationary state, in the sense that ρ̇G = 0. Here the energies are listed in increasing order, starting from ground state
energy 0 . We will show this here directly from the RWA-LE, and given an alternative derivation from the Pauli master equation
in Sec. XVII.
In the Schrödinger picture the RWA-LE has the form:
To show that ρ̇G = 0, consider first the Hamiltonian part. That [HS , ρG ] = 0 follow immediately from Eq. (580). Now recall
that [HS , HLS ] = 0 [Eq. (539)]. Thus HS and HLS are diagonalizable in the same basis, i.e., there exists a unitary V such that
V HS V † = D1 and V HLS V † = D2 , where D1 and D2 are both diagonal (and of course commute). Therefore
1
V [HLS , ρG ]V † = V HLS V † V ρG V † − V ρG V † V HLS V † = [D2 , e−βD1 ] = 0 , (583)
Z
11 Recall Eq. (460): ω = εb − εa < 0 corresponds to a transition from ∣εb ⟩ to ∣εa ⟩, i.e., from energy εb to a higher energy εa .
87
where the projectors Πa = ∣a⟩⟨a∣ are in the energy basis, i.e., HS Πa = Πa HS = aΠa , where HS = ∑a aΠa . Using the property
Πa Πb = δab Πa , note that:
[HS , Aα (ω)] = ∑ aΠa ∑ Πa′ Aα Πb − ∑ Πa Aα Πb ∑ a′ Πa′ (587a)
a b−a′ =ω b−a=ω a′
= ∑ aΠa Aα Πb − ∑ bΠa Aα Πb = ∑ (a − b)Πa Aα Πb (587b)
b−a=ω b−a=ω b−a=ω
= −ωAα (ω) (587c)
[HS , A†α (ω)] = ωA†α (ω) , (587d)
where Eq. (587d) follows by taking the Hermitian conjugate of Eq. (587c).
Therefore:
[HS , Aα (ω)]n = (−ω)n Aα (ω) (588a)
[HS , A†α (ω)]n = ω n Aα (ω) . (588b)
Hence, using Eq. (584):
∞
(−β)n (−ω)n
e−βHS Aα (ω)eβHS = ∑ Aα (ω) = eβω Aα (ω) , (589)
n=0 n!
which tells us that
Aα (ω)ρG = e−βω ρG Aα (ω) . (590)
It follows by Hermitian conjugation that:
ρG A†α (ω) = e−βω A†α (ω)ρG . (591)
We are now ready to consider the terms in the dissipator, Eq. (582). Commuting ρG to the right we find:
Aβ (ω)ρG A†α (ω) = e−βω Aβ (ω)A†α (ω)ρG (592a)
−βω
ρG A†α (ω)Aβ (ω) =e A†α (ω)ρG Aα (ω) = A†α (ω)Aβ (ω)ρG , (592b)
The action of the dissipator thus becomes:
D(ρG ) = g 2 ∑ ∑ γαβ (ω) (e−βω Aβ (ω)A†α (ω) − A†α (ω)Aβ (ω)) ρG . (593)
αβ ω
Let us now separate the sum over ω as ∑ω<0 +(ω = 0) + ∑ω>0 . Recall that KMS result [Eq. (575)]: γαβ (−ω) = e−βω γβα (ω).
We know from Eq. (586) that Aα (0) = A†α (0), so that the ω = 0 cancels since the remaining sum is over all α and β, and by
KMS, γαβ (0) = γβα (0). As for the sum over negative frequencies, using KMS and Eq. (586) again we have:
∑= γαβ (−ω ′ ) (eβω Aβ (−ω ′ )A†α (−ω ′ ) − A†α (−ω ′ )Aβ (−ω ′ )) ρG
′
∑ (594a)
ω<0 ω ′ =−ω>0
so that
−βω
∑ ∑ = γβα (ω) (Aβ (ω)Aα (ω) − e Aα (ω)A†β (ω)) ρG
†
(595a)
αβ ω<0
= −∑ ∑ , (595b)
αβ ω>0
ρ̇G = 0 . (596)
C. Return to equilibrium, quantum detailed balance, and ergodicity under the RWA-LE
A natural next question is under which conditions the Gibbs state is actually reached. To answer this we need to define the
concept of ergodicity. A system is ergodic if it holds that for any arbitrary system operator X
for (A, B) ≡ Tr[ρ̃A† B] and all A, B ∈ domain(L† ), then for any initial state ρ(0) the stationary state is the Gibbs state. I.e., the
t→∞
Gibbs state is an attractor for the dynamics: ρ(t) = eLt ρ(0) ÐÐ→ ρG . This is a fundamental result, as it tells us the conditions
under which a system is guaranteed to become thermally equilibrated. The proof is given in Sec. 1.3.4 of Ref. [2] (see also
Ref. [30]).
However, not all systems are ergodic [25]. For example, consider a system of N qubits coupled to a bath such that
N
Aα = ∑ σjα , α ∈ {x, y, z} . (599)
j=1
Clearly, all Aα are invariant under permutations, so that they commute with the elements of the permutation group. This means
that Eq. (597) is satisfied for operators X that are not proportional to the identity (e.g., the SWAP operator between any pair of
qubits), and hence such a system is not ergodic. Indeed, Eq. (599) describes “collective decoherence”, under which there exist
subspaces that are invariant under the action of the Aα operators, and undergo unitary dynamics [26, 27]. Initial states in such
subspaces do not converge to the Gibbs state, and do not equilibrate.
More generally, if the system-bath interaction possesses some symmetry (e.g., a permutational symmetry as above), then
ergodicity does not hold and the system need not equilibrate [28].
Sometimes we are particularly interested in finding out the evolution of just the populations (diagonal elements) in the energy
eigenbasis. For example, this is the case in adiabatic quantum computing and quantum annealing, where the answer to a
computation is encoded in the ground state [29]. In other applications we are interested in finding out the Gibbs distribution
ρG [Eq. (580)] in order to compute various thermodynamic averages ⟨X⟩ = Tr(XρG ), where X could be any observable
of interest; the Gibbs state is an example of a state that is diagonal in the energy eigenbasis, i.e., the eigenbasis {∣a ⟩} of
HS = ∑a a ∣a ⟩⟨a ∣ = ∑a α Πa .
Recall that the RWA-LE in the Schrödinger picture is
dρ 1
= −i [HS + HLS , ρ] + g 2 ∑ ∑ γαβ (ω) [Aβ (ω)ρA†α (ω) − {A†α (ω)Aβ (ω), ρ}] .
dt αβ ω 2
89
where we again used the simplified notation a ↦ a. We have for the first term in line (601c):
⟨a ∣ Aβ (ω)ρA†α (ω) ∣a ⟩ = ⟨a∣ ∑ Aa′ b′ ,β ∣a′ ⟩⟨b′ ∣ρ ∑ Ab′′ a′′ ,α ∣b′′ ⟩⟨a′′ ∣a⟩ (604a)
ω=b′ −a′ ω=b′′ −a′′
= ∑ Aab′ ,β ρb′ b′′ Ab′′ a,α (604b)
ω=b′ −a
ω=b′′ −a
= ∑ Aab′ ,β pb′ Ab′ a,α = ∑ Aaa′ ,β pa′ Aa′ a,α , (604c)
ω=b′ −a ω=a′ −a
where to go the second line we used ⟨a∣a′ ⟩ = δaa′ and ⟨a′′ ∣a⟩ = δa′′ a , and to go to the third line we used the fact that b′ must
equal b′′ due to the summation constraints.
Similarly,
′ ′
⟨a∣ A†α (ω)Aβ (ω)ρ ∣a⟩ = ∑ ⟨a∣ Ab′ a′ ,α ∣b ⟩⟨a ∣ ∑ Aa′′ b′′ ,β ∣a′′ ⟩⟨b′′ ∣ρ ∣a⟩ (605a)
ω=b′ −a′ ω=b′′ −a′′
= ∑ Aaa′ ,α Aa′ b′′ ,β ρb′′ a (605b)
ω=a−a′
ω=b′′ −a′
= ∑ Aaa′ ,α Aa′ a,β pa , (605c)
ω=a−a′
and
′ ′
⟨a∣ ρA†α (ω)Aβ (ω) ∣a⟩ = ∑ ⟨a∣ ρAb′ a′ ,α ∣b ⟩⟨a ∣ ∑ Aa′′ b′′ ,β ∣a′′ ⟩⟨b′′ ∣a⟩ (606a)
ω=b′ −a′ ω=b′′ −a′′
= ∑ ρab′ Ab′ a′ ,α Aa′ a,β (606b)
ω=b′ −a′
ω=a−a′
= ∑ pa Aaa′ ,α Aa′ a,β , (606c)
ω=a−a′
ṗa = ∑ ( ∑ Aa′ a,α Aaa′ ,β pa′ − ∑ Aaa′ ,α Aa′ a,β pa ) γαβ (ω) . (607)
αβ ω=a′ −a ω=a−a′
90
Since the index a is fixed, the sum over ω really only involves varying a′ . Thus:
ṗa = ∑ ∑ γαβ (a′ − a)Aa′ a,α Aaa′ ,β pa′ − γαβ (a − a′ )Aaa′ ,α Aa′ a,β pa . (608)
αβ a′
Note that W (a∣a′ ) ≥ 0. To prove this, let u be the unitary matrix that diagonalizes γ: γαβ = ∑α′ uαα′ γα′ u∗βα′ . Then:
⎛ ⎞
W (a∣a′ ) = ∑ uαα′ γα′ (a′ − a)u∗βα′ Aa′ a,α Aaa′ ,β = ∑ γα′ (a′ − a) (∑ uαα′ Aa′ a,α ) ∑ u∗βα′ Aaa′ ,β (610a)
α′ αβ α′ α ⎝β ⎠
2
= ∑ γα′ (a′ − a) ∣Ãa′ a,α′ ∣ ≥ 0 , (610b)
α′
where Ãa′ a,α′ = ∑α uαα′ Aa′ a,α , and we used the Hermiticity of Aβ to write Aaa′ ,β = A∗a′ a,β . Eq. (608) can thus be simplified
as:
This represents a closed set of rate equations for the populations {pa }.
If we assume that the KMS condition γαβ (−ω) = e−βω γβα (ω) (for ω > 0) holds, then this allows us to write, for a > a′ :
W (a∣a′ ) = ∑ e−β(a−a ) γβα (a − a′ )Aa′ a,α Aaa′ ,β = e−β(a−a ) ∑ γαβ (a − a′ )Aa′ a,β Aaa′ ,α = e−β(a−a ) W (a′ ∣a) .
′ ′ ′
(613)
αβ αβ
It says that the rate for an “up” transition, from the low energy state ∣a′ ⟩ to the high energy state ∣a⟩, is exponentially less likely
than the reverse, “down” transition, with the exponent given by the energy difference in units of the bath temperature. This is
an extremely important result, since it establishes rigorously the intuition that at very low temperatures (relative to the smallest
energy gap) systems tend to relax towards their ground states. This is a special case of the quantum detailed balance condition
we mentioned in Sec. XVI C.
Finally, we can also reestablish that the Gibbs state is the stationary state (recall that we showed this in Sec. XVI B). For a
stationary state ṗa = 0. It follows from Eq. (611) that in this case:
W (a∣a′ ) pa e−βa
= ′ = e−β(a−a ) = −βa′
′
(615a)
W (a ∣a) pa
′ e
e−βa
⇒ pa = ; Z = ∑ e−βa , (615b)
Z a
All our derivations of the LE so far have assumed the weak coupling limit of system-bath coupling. Somewhat surprisingly,
the opposite limit of strong coupling also allows us to derive the Lindblad equation, while avoiding the use of the RWA.
91
A. Derivation
We can perform a change of variables to τ = 2 τ ′ , and take the limit → 0, so that τ → 0. Then, recalling Eq. (501), the various
terms in Eq. (618) transform as follows:
US (t)Aβ (t − τ )US† (t − τ )ρ(t − τ )US (t − τ )Aα (t)US† (t) = US† (−τ )Aβ ρ(t − τ )US† (τ )Aα
→ Aβ ρ(t)Aα = Aβ ρ(t)A†α (619a)
US (t)Aα (t)Aβ (t − τ )US† (t − τ )ρ(t − τ )US (t − τ )US† (t) = Aα US† (−τ )Aβ ρ(t − τ )US† (τ )
→ Aα Aβ ρ(t) = A†α Aβ ρ(t) (619b)
i2 τ ′ HB /2 −i2 τ ′ HB /2
Bαβ (τ ) = Tr (e Bα e Bβ ρB ) = Tr (UB† (τ ′ )Bα UB (τ ′ )Bβ ρB ) ′
= Bαβ (τ ) (619c)
1 t t −2
∞
′
∫ dτ = ∫ dτ → ∫ dτ ′ . (619d)
2 0 0 0
Thus the → ∞ strong coupling and bath limit, is essentially a Markovian limit, as it allows us to extend the integration limit to
∞ and make ρ time-local. It also removes the time dependence from the Aα system operators.
Applying the transformations in Eq. (619) to Eq. (618) gives:
dρ ∞
= −i[HS , ρ(t)] + g 2 ∑(Aβ ρ(t)A†α − A†α Aβ ρ(t)) ∫ dτ Bαβ (τ ) + h.c. . (620)
dt αβ 0
Note that the SCL keeps only the ω = 0 component out of all the Bohr frequencies, so it is clearly a more “extreme” limit than
the WCL. We can understand this as a consequence of the fact that the SCL is designed to accelerate the internal evolution of
the bath by rescaling the bath Hamiltonian via HB ↦ HB /2 ; this means that all system frequencies are effectively zero relative
to the very high effective bath evolution frequency, and only the static component ω = 0 survives.
92
We assume that
1
HS = − ωz σ z , HSB = gσ z ⊗ B. (623)
2
For the interaction Hamiltonian in Eq. (623), there is only a single system operator Az = σ z = ∣0⟩⟨0∣ − ∣1⟩⟨1∣. The eigenstates are
∣ε0 ⟩ = ∣0⟩ and ∣ε1 ⟩ = ∣1⟩. Considering the RWA-LE (the weak coupling limit case) Eq. (536) and ⟨εa ∣ Az ∣εb ⟩ ∝ δab , there is only
a single Lindblad operator that is non-zero:
Az (0) = σ z , (624)
as given by Eq. (520). This follows since [HS , HSB ] = 0. Therefore, the RWA-LE [Eq. (540)] takes the simple form
d 1
ρ(t) = −i [HS , ρ(t)] + g 2 γ(0) (σ z ρ(t)(σ z )† − {(σ z )† σ z , ρ(t)}) , (625)
dt 2
where we have also used the fact that HLS ∝ I. This form is the same as what is predicted in the SCL, since only the ω =
0 component appears. We have encountered this equation several times before [e.g., Eq. (277)]. After expanding ρ(t) =
∑i,j∈{0,1} ρij ∣i⟩ ⟨j∣, and taking matrix elements in the computational basis (which here is equivalent to the energy eigenbasis)
we obtain:
where
(c) 1
T2 = , (627)
2g 2 γ(0)
where the ‘c’ superscript denotes the computational basis (we shall shortly see a second T2 associated with the energy eigenba-
sis). This is the familiar phase damping channel, where only the off-diagonals elements (transverse magnetization) decay with a
(c)
characteristic timescale T2 . The stronger the coupling to the bath g, the shorter the qubit coherence time. Note that the qubit
(c) (c)
energy gap ωz plays no role in the result for T2 , and T2 here is entirely determined by the spectrum of the bath correlation
function at zero frequency. In this example there is no thermal relaxation (the T1 time is infinite), since the population of the
energy states remains fixed, as a consequence of [HS , HSB ] = 0.
Let us now replace the system Hamiltonian so that [HS , HSB ] ≠ 0. Specifically, consider
1
HS = − ωx σ x , HSB = gσ z ⊗ B . (628)
2
We shall see that there is a sharp contrast between the WCL and SCL, with the WCL resulting in decoherence in the energy
eigenbasis, while the SCL results in decoherence in the computational basis, just as in the previous subsection, when HS and
HSB were commuting.
a. WCL The energy eigenstates of HS are ∣ε0 ⟩ = ∣+⟩ with eigenvalue − 21 ωx (ground state) and ∣ε1 ⟩ = ∣−⟩ with eigenvalues
1
ω (excited state), where ∣±⟩ = √12 (∣0⟩ ± ∣1⟩). Therefore the possible Bohr frequencies are ω ∈ {0, ±ωx }. Since σ z ∣±⟩ = ∣∓⟩,
2 x
we find Az (0) = 0, and the non-zero Lindblad operators are:
Writing ρ(t) = ∑i,j∈{+,−} ρij ∣i⟩⟨j∣, and taking matrix elements of Eq. (631), we find:
⟨−∣ ρ̇ ∣−⟩ = ρ̇−− = −i ⟨−∣ (HS + HLS )ρ − ρ(HS + HLS ) ∣−⟩ − g 2 γ(ωx )ρ−− + g 2 γ(−ωx )ρ++ , (632)
and the first (Hamiltonian) term is easily seen to vanish. Also, note that Tr[ρ(t)] = ρ++ (t) + ρ−− = 1 implies that ρ̇−− = −ρ̇++ .
After a similar calculation for the off-diagonal components, we find that the Lindblad equation for the density matrix components
is:
d d
−ρ++ = ρ−− = −g 2 γ(ωx )ρ−− (t) + g 2 γ(−ωx )ρ++ (t) (633a)
dt dt
d ∗ d 1
ρ (t) = ρ−+ (t) = Ωρ−+ (t) , Ω ≡ −i [Ω(−ωx ) − Ω(ωx )] − g 2 [γ(ωx ) + γ(−ωx )] . (633b)
dt +− dt 2
The solution for the off-diagonal elements [Eq. (633b)] is immediate: ρ−+ (t) = ρ−+ (0)e−iΩt , i.e.:
(e) 1 (e)
T1 = T2 . (637)
2
Moreover, recall that the Gibbs state is
1 −βHS 1 12 βωx σx
ρG = e = e = pG (−)∣+⟩⟨+∣ + pG (+)∣−⟩⟨−∣ , (638)
Z Z
where
e±βωx /2
pG (±) = , Z = Tr(ρG ) = pG (−) + pG (+) = 2 cosh(βωx /2) . (639)
Z
Using this and the KMS condition, we have
γ(−ωx )
d= = PG (−) . (640)
γ(−ωx )(1 + eβωx )
94
where bk is the annihilation operator associated with bosonic mode k, and the system-bath interaction is
HSB = gA ⊗ B , A = σz , B = ∑(gk /g)(bk + b†k ) . (646)
k
There is only a single bath correlation function, because there is only a single bath operator B. For a bath in a Gibbs state at
inverse temperature β it can be shown that the bath correlation function in this case is [21, Appendix H]:
(gk /g)2
⟨B(t)B⟩B = ∑ (e−iωk t + eiωk t−βωk ) . (647)
k 1 − e−βωk
95
10
3 8
6
2
1
2
ω ω
-10 10 20 30 40 50 60 -20 -10 10 20 30 40 50
FIG. 8. The Lindblad rate γ(ω) for an Ohmic spectral density [Eq. (650c)], for η = 1/(2π), ωc = 20, and low temperature β = 10 (left) or high
temperature β = 0.1 (right). It can be checked numerically that the peak is always at ω ≈ ωc for sufficiently large β, or at ω = 0 for sufficiently
small β. Note that γ(0) = 1/β.
12 If J(ω) = ηω ζ e−ω/ωc then the ζ > 1 case is called super-Ohmic, and the 0 < ζ < 1 case is called sub-Ohmic.
96
To close our discussion of the RWA-LE, let us revisit the phase damping model we considered in Sec. X C, but this time for n
qubits. Thus the system Hamiltonian is
n
HS = ∑ εα Zα . (654)
α=1
The eigenstates {∣a ⟩}2a=0−1 are just the computational basis states, i.e., all length-n bit strings.
n
In the collective phase damping case there is a qubit permutation symmetry and the qubits are all coupled to the same bosonic
modes. Thus the system-bath interaction is
n
HSB = ∑ gk Zα ⊗ (bk + b†k ) = gA ⊗ B , A = ∑ Zα , B = ∑(gk /g)(bk + b†k ) . (655)
k,α α=1 k
Since there is only one bath operator, the analysis starting from Eq. (647) holds without any change.
Here each qubit is coupled to a separate bosonic bath. Thus the bath Hamiltonian is
n
HB = ∑ HB,α , HB,α = ∑ ωk,α b†k,α bk,α , (656)
α=1 k
where bk,α is the annihilation operator associated with bosonic mode k and qubit α, and the system-bath interaction is
n
HSB = ∑ gk,α Zα ⊗ (bk,α + b†k,α ) = g ∑ Aα ⊗ Bα , A α = Zα , Bα = ∑(gk,α /g)(bk,α + b†k,α ) . (657)
k,α α=1 k
The bath Gibbs state factors since operators belonging to different qubit indices commute:
1 1 −βHB,α
ρB = e−βHB = ⊗ ρB,α , ρB,α = e , (658)
Z α Z α
where Zα = Tre−βHB,α . In light of this case, consider the bath correlation functions, and recall that Tr(A ⊗ B) = TrA × TrB
for any pair of operators A and B:
⟨Bα (t)Bβ ⟩B = Tr (ρB eiHB,α t Bα e−iHB,α t Bβ ) (659a)
α≠β −iHB,α t
= Tr (ρB,α e iHB,α t
Bα e ) Tr (ρB,β Bβ ) (659b)
=0, (659c)
where the last equality follows since (as in Sec. XIV C) we can always ensure that Tr(ρB B) = 0. If α = β, we recover the
expression we obtained in the collective case but with the bath parameters corresponding to the α-th bath. Thus,
⟨Bα (t)Bβ ⟩B = δαβ ⟨Bα (t)Bα ⟩B . (660)
This, in turn, implies that
∞
γαβ (ω) = ∫ dteiωt ⟨Bα (t)Bβ ⟩B = δαβ γαα (ω) . (661)
−∞
If we again assume an Ohmic spectral density, now of the form
Jα (ω) = ηα ωe−ω/ωc,α , (662)
then the same calculation as in the collective case yields
2πηα ωe−∣ω∣/ωc,α
γαα (ω) = , (663)
1 − e−βω
where we have assumed that all baths are thermally equilibrated at the same inverse temperature β.
97
We can now compare the predictions of the collective and independent dephasing models. Consider the time evolution of the
density matrix elements in the energy eigenbasis, i.e., ρ̇ab . Using the RWA-LE we have:
1
ρ̇ab = ⟨a ∣ρ̇∣b ⟩ = ⟨a ∣ ∑ γαβ (ω) (Aβ (ω)ρA†α (ω) − {A†α (ω)Aβ (ω), ρ}) ∣b ⟩ . (664)
αβ,ω 2
Evaluating this yields, after some algebra:
n
1
ind
1/τab = g 2 ∑ γαα (0)(Aaa,α − Abb,α )2
independent: (665a) ρ̇ab = −ρab /τab
ind
,
2 α=1
1
collective: ρ̇ab = −ρab /τab
col col
, 1/τab = g 2 γ(0)(Aaa − Abb )2 , (665b)
2
where we used the explicit form of the eigenstates of the system Hamiltonian in Eq. (654). We see that, as expected from
single-qubit dephasing case (recall, e.g., Sec. XIV G) that there is no change in the populations, i.e., ρ̇aa = 0. The solution to
these decoupled equations for the off-diagonal elements is of the form ρab (t) = ρab (0)e−t/τab , where τab is the dephasing time.
Let us compare the scaling of this time with the number of qubits n in the independent and collective dephasing settings.
• Independent-dephasing:
Aaa,α = ⟨a ∣Zα ∣a ⟩ = ±1 . (666)
Thus, (Aaa,α − Abb,α )2 = 4 for a ≠ b.
• Collective dephasing:
n
Aaa = ⟨a ∣ ∑ Zα ∣a ⟩ ∈ {−n, −n + 2, . . . , n − 2, n} (667)
α=1
Thus max(Aaa − Abb )2 = 4n2 and min(Aaa − Abb )2 = 0 for even n, or min(Aaa − Abb )2 = 4 for odd n.
There is thus a substantial difference between the two models. In the independent case, using Eq. (665a), we find 1/τab ind
=
O(n), or simply 1/τab = 2nγ(0) if all rates γαα (0) are equal [to γ(0)]. In the collective case, using Eq. (665b), we have a
ind
for the zero-eigenvalue system eigenstates of the collective dephasing operator ∑nα=1 , i.e., states ∣a ⟩ that have an equal number
of 0’s and 1’s in the computational basis. Such states form a conserved subspace under the action of the RWA-LE, and hence are
called a decoherence-free subspace [26–28] (recall also our discussion of non-equilibration in Sec. XVI C). At the other extreme,
the states in the superdecoherent subspace dephase quadratically faster than in the independent dephasing case.
Earlier we asserted that it is permissible to go from Eq. (512) to Eq. (515). Our goal is now to prove this, and in particular to
derive the associated error estimate, O(g 4 τB3 ).
Consider just one of the four (two due to the commutator, times two due to the h.c.) terms in Eqs. (512), and its Markov
approximation [as in Eq. (515)]:
t
true ≡ g 2 ∑ ∫ dτ Bαβ (τ )Aα (t)Aβ (t − τ )ρ̃(t − τ ) (668a)
αβ 0
∞
≈ g2 ∑ ∫ dτ Bαβ (τ )Aα (t)Aβ (t − τ )ρ̃(t) ≡ approx (668b)
αβ 0
∞ ∞
= g2 ∑ ∫ dτ Bαβ (τ )Aα (t)Aβ (t − τ ) (ρ̃(t) − ρ̃(t − τ )) +g 2 ∑ ∫ dτ Bαβ (τ )Aα (t)Aβ (t − τ )ρ̃(t − τ ) (668c)
αβ 0 αβ 0
´¹¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¸ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¶
∆1
∞
= g 2 ∑ ∆1 + true + g 2 ∑ ∫ dτ Bαβ (τ )Aα (t)Aβ (t − τ )ρ̃(t − τ ) . (668d)
αβ t
αβ
´¹¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¸ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¶
∆2
98
1. Bound on ∥∆1 ∥
Now, by the mean value theorem of elementary calculus, there exists a point t′ ∈ [t − τ, t] such that
ρ̃(t) − ρ̃(t − τ ) ˙ ′
= ρ̃(t ) . (672)
τ
Therefore
∥ρ̃(t) − ρ̃(t − τ )∥∞ ≤ τ ˙ ′ )∥∞ ,
sup ∥ρ̃(t (673)
t′ ∈[t−τ,t]
and
∞
∥∆1 ∥∞ ≤ η 2 ∫ ˙ ′ )∥∞ .
dτ τ ∣Bαβ (τ )∣ sup ∥ρ̃(t (674a)
0 t′ ∈[t−τ,t]
t′
≤ 4g 2 ∑ ∫ dτ ∣Bαβ (τ )∣∥Aα (t)Aβ (t − τ )ρ̃(t − τ )∥∞ (675b)
α,β 0
t′
≤ 4g 2 ∑ ∫ dτ ∣Bαβ (τ )∣∥Aα ∥∞ ∥Aβ ∥∞ ∥ρ̃(t − τ )∥1 (675c)
α,β 0
t′
≤ 4(ηg)2 M ∫ dτ ∣Bαβ (τ )∣ , (675d)
0
where in the second line we used the fact that all four terms in the first line (again, after the commutator and h.c.) have the same
operator norm, and where M ≡ ∑αβ 1 is the square of the number of summands in HSB = ∑α Aα ⊗ Bα . Now, since
t′ ∞
sup ∫ dτ ∣Bαβ (τ )∣ ≤ ∫ dτ ∣Bαβ (τ )∣ , (676)
t′ ∈[t−τ,t] 0 0
we have
∞ ∞
∥∆1 ∥∞ ≤ 4M η 4 g 2 ∫ dτ τ ∣Bαβ (τ )∣ ∫ dτ ∣Bαβ (τ )∣ ∼ 4M η 4 g 2 τB3 , (677)
0 0
2. Bound on ∥∆2 ∥
Similarly,
∞
∥∆2 ∥∞ ≤ ∫ dτ ∣Bαβ (τ )∣∥Aα (t)∥∞ ∥Aβ (t − τ )∥∞ ∥ρ̃(t − τ )∥1 (678a)
t
∞
≤ η2 ∫ dτ ∣Bαβ (τ )∣ . (678b)
t
∞
Intuitively, we know that ∫t dτ ∣Bαβ (τ )∣ should be arbitrarily small as long as t ≫ τB , as we assumed in Eq. (513), since the
∞
correlation function decays over a timescale of τB . To formalize this, note that convergence of ∫t dτ ∣Bαβ (τ )∣ is guaranteed if
Thus, we will assume that the correlation function decays no more slowly than this power-law dependence [this is even slower
than the subexponential decay we assumed to get Eq. (518)]. Under this assumption, we have
∞
∞ ∞ τB x τBx 1 τBx
∫ dτ ∣Bαβ (τ )∣ ∼ ∫ dτ ( ) = ∣ = . (680)
t t τ (1 − x)τ x−1 t x − 1 tx−1
Now, to use the assumption that t ≫ τB , let us write t = cτB , where c ≫ 1. Then:
∞ τB
∫ dτ ∣Bαβ (τ )∣ ∼ . (681)
t (x − 1)cx−1
Therefore, even with a power-law decaying correlation function, we have
τB
∥∆2 ∥∞ ≲ η 2 , (682)
(x − 1)cx−1
which can be made arbitrarily small by making c = t/τB large enough.
We have seen that ∥∆1 ∥∞ ≲ 4M η 4 g 2 τB3 = O(g 2 τB3 ) and ∥∆2 ∥∞ can be made arbitrarily small. Thus the dominant contribu-
tion to the error comes from ∥∆1 ∥∞ , which is the error due to replacing all the intermediate-time states (at t − τ ) by the state at
the single time t. Moreover, we need t ≫ τB in order to ensure that ∥∆2 ∥∞ can be neglected.
When accounting for the additional g 2 prefactor in Eq. (668) (as well as ∑αβ , which just gives rise to another factor of M ),
we finally have from Eq. (669):
as claimed.
The RWA we used in Sec. XV E in order to derive the Lindblad equation leaves something to be desired. We simply dropped
terms with different Bohr frequencies, without a rigorous mathematical justification. We will now show that the RWA-LE can be
rigorously derived from the cumulant-LE, in the limit of an infinite coarse-graining timescale. This shows that the cumulant-LE
is truly more general than the (standard) RWA-LE.
1. Quick summary
For convenience, let us collect the main results of each of the two approaches. For simplicity we’ll set λ = g = 1 and also
assume that HSB = A ⊗ B (not a sum), so that we can drop the α index from Eq. (536). The RWA-LE is then:
1
˙
ρ̃(t) = −i [HLS , ρ̃(t)] + ∑ γ(ω) (Aω ρ̃(t)A†ω − {A†ω Aω , ρ̃(t)}) (684)
ω 2
100
with
∞
γ(ω) = ∫ dseiωs B(s, 0) . (685)
−∞
The cumulant-LE is
′ 1
˙
ρ̃(t) = −i [HLS , ρ̃(t)] + ∑ γωω′ (τ ) (Aω ρ̃(t)A†ω′ − {A†ω′ Aω , ρ̃(t)}) , (686)
ω,ω ′ 2
where the rates γ keep a dependence on two different Bohr frequencies ω and ω ′ :
1 τ τ
γωω′ (τ ) = bωω′ (τ ) , bωω′ (τ ) = ∫ ds′ ei(ω s−ωs ) B(s, s′ ) .
′ ′
ds ∫ (687)
τ 0 0
Our goal is to show that in an appropriate sense the cumulant-LE tends to the RWA-LE in the limit as τ → ∞, where τ is the
coarse-graining timescale. More specifically, we will show that limτ →∞ γωω′ (τ ) = γ(ω)δωω′ [17]. We will assume stationarity,
i.e., B(s, s′ ) = B(s − s′ , 0).
2. A useful lemma
Proof. In the RWA we dropped terms with ω ≠ ω ′ , so it makes sense to rewrite ω ′ s − ωs′ in terms of a sum and difference of
Bohr frequencies:
1 1
ω ′ s − ωs′ = (ω ′ − ω)v + (ω ′ + ω)u , (689)
2 2
where u = s − s′ and v = s + s′ . After this change of variables B(s − s′ , 0) = B(u, 0), and since s = (v + u)/2 and s′ = (v − u)/2,
1/2 1/2
the Jacobian of the transformation is ∣( )∣ = 1/2. In terms of the new variables the integration region is diamond
1/2 −1/2
shaped (a square rotated by π/4), bounded between the lines u = v and u = −v for v ∈ [0, τ ] and the lines u = 2τ − v and v − 2τ
for v ∈ [τ, 2τ ]. Thus:
τ v 2τ 2τ −v
1 ω ′ −ω ω+ω ′ 1 ω ′ −ω ω+ω ′
bωω ′ (τ ) = ∫ dv ei 2 v ∫ du ei 2 u B(u, 0) + ∫ dv ei 2 v ∫ du ei 2 u
B(u, 0) . (690)
2 2
0 −v τ −(2τ −v)
To get the integration limits to be the same we make a change of variables from v to 2τ − v in the second double integral:
τ v τ v
1 ω ′ −ω ω+ω ′ 1 ω ′ −ω ω+ω ′
bωω′ (τ ) = ∫ dv ei 2 [(v−τ )+τ ] ∫ du ei 2 u B(u, 0) + ∫ dv e−i 2 [(v−τ )−τ ] ∫ du ei 2 u B(u, 0) (691a)
2 2
0 −v 0 −v
τ v
′
i ω 2−ω τ ω′ − ω ω+ω ′
=e ∫ dv cos ( (v − τ )) ∫ du ei 2 u B(u, 0) . (691b)
2
0 −v
3. The ω = ω ′ case
τ v
1
γωω (τ ) = ∫ dv ∫ du e B(u, 0) .
iωu
(692)
τ
0 −v
101
v
Let U = ∫−v du eiωu B(u, 0). Recall the Leibnitz rule for differentiating a definite integral:
b(z) b(z)
∂z ∫ f (x, z)dx = ∫ ∂z f (x, z)dx + f (b(z), z)b′ − f (a(z), z)a′ . (693)
a(z) a(z)
Therefore dU = (eiωv B(v, 0) + e−iωv B(−v, 0)) dv. Then, integrating by parts (∫0 U dv = [U v]0 − ∫0 vdU ) gives:
τ τ τ
⎡ v ⎤τ τ
1 ⎢⎢ ⎥ 1
γωω (τ ) = ⎢v ∫ du e B(u, 0)⎥⎥ − ∫ dv v (eiωv B(v, 0) + e−iωv B(−v, 0)) .
iωu
(694)
τ⎢ ⎥ τ
⎣ −v ⎦0 0
4. The ω ≠ ω ′ case
For ω ≠ ω ′ we also perform integration by parts of Eq. (688), but we shall see that this time the boundary terms vanish. We
ω ′ −ω ω+ω ′
write γωω′ (τ ) = τ1 ei 2 τ ∫0 dV U (v), where now dV = cos ( ω 2−ω (v − τ )) dv and U (v) = ∫−v du ei 2 u B(u, 0). Then
τ ′ v
2 ω′ − ω
V (v) = sin ( (v − τ )) (697a)
ω′ − ω 2
ω+ω ′ ω+ω ′
dU /dv = ei 2 v
B(v, 0) + e−i 2 v
B(−v, 0) (697b)
τ
[U (v)V (v)]0 = U (τ )V (τ ) − U (0)V (0) = 0 . (697c)
Therefore:
ω ′ −ω τ
τ 2ei 2 τ (ω ′ − ω) ω+ω ′ ω+ω ′
γωω′ (τ ) = − ∫ V dU = − ′ ∫ dv sin ( (v − τ )) [ei 2 v B(v, 0) + e−i 2 v B(−v, 0)] . (698)
0 (ω − ω)τ 2
0
where we used the angle sum identity for the sine in the last equality. Thus:
ω ′ −ω
ei 2 τ ω′ − ω ′ ω′ − ω
lim γ ωω ′ (τ ) = lim [sin ( τ ) (γ(ω) + γ(ω )) + 2 cos ( τ ) (S(ω) − S(ω ′ ))] . (700)
τ →∞ τ →∞ (ω ′ − ω)τ 2 2
∞
where we have used that for Γ(ω) = ∫0 ds eiωs B(s, 0) [recall Eq. (523)], we have γ(ω) = Γ(ω) + Γ∗ (ω) and 2iS(ω) =
Γ(ω) − Γ∗ (ω) [recall Eq. (535)]. Since nothing cancels with the overall τ −1 , we find that the ω ≠ ω ′ term vanishes.
A similar calculation could be done for the Lamb shift term (466). Therefore, the RWA results can be understood as the
τ → ∞ limit of the coarse-graining timescale.
102
The master equations we have developed so far are approximations to the true dynamics. In this section we take a step back
and derive an exact master equation. Since it is exact, it will naturally be non-Markovian.
Consider the total Hamiltonian
H = H0 + αHSB , H0 = HS + HB , (701)
where 0 < α < 1 is a dimensionless parameter. Let us work in the interaction picture, so that the total system-bath state ρ̃ satisfies
where as usual H̃(t) = U0† (t)HSB U0 (t), with U0 (t) = e−iH0 t , and HSB = ∑ Sα ⊗ Bα . We abbreviate ∂t ≡ ∂t ∂
. For the rest
of this section we drop the tilde decoration on states to simplify the notation, so that, e.g., ρ denotes the interaction-picture
system-bath state.
Consider a fixed bath state ρB . As usual, ρS = TrB ρ is the system state of interest. Consider the projection superoperator P
defined via
Q=I −P . (705)
We call Pρ the “relevant” part, and Qρ the “irrelevant part”. This procedure is sometimes called Feshbach P-Q partitioning,
after a method introduced in nuclear scattering theory [32].
We are interested in deriving a master equation for ∂t (Pρ). Now, note that
Likewise:
These are coupled differential equations for the relevant (Pρ) and irrelevant (Qρ) parts. To solve them, let us eliminate the
irrelevant part.
Define
X̂ ≡ PX , X̄ ≡ QX (710)
for any operator X. then Eq. (709) can be rewritten more compactly as:
B. Derivation
We can formally solve the second of these equations and substitute the solution into the first. Consider first ∂t ρ̄ = αL̄ρ̄. This
has the immediate solution ρ̄(t) = T+ exp (α ∫t0 L̄(t′ )dt′ ) ρ̄(t0 ), where T+ denotes the usual forward Dyson time-ordering. We
t
thus define
α ∫tt L̄(t′ )dt′
G(t, t0 ) ≡ T+ e 0 . (712)
Eq. (711b) contains another term, and we can easily guess that the solution integrates over this term, but first applies G, i.e.:
t
ρ̄(t) = G(t, t0 )ρ̄(t0 ) + α ∫ G(t, t′ )L̄(t′ )ρ̂(t′ )dt′ . (713)
t0
´¹¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹¸¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¶
⊛
To verify that this is the formal solution of Eq. (711b), we apply the Leibnitz rule (693) to get ∂t ∫t0 f (t, t′ )dt′ = f (t, t) +
t
∂t f (t, t′ )dt′ , and also note that G(t, t0 ) has the property G(t, t) = I, ∂t G(t, t′ ) = L̄(t)G(t, t′ ). Using all of the above we
t
∫t0
have:
t
∂t ⊛ = αG(t, t)L̄(t)ρ̂(t) + α ∫ ∂t G(t, t′ )L̄(t′ )ρ̂(t′ )dt′ = αL̄(t)ρ̂(t) + αL̄(t)⊛ = αL̄(t) (ρ̂(t) + ⊛) . (714)
t0
• We can show that term (a) can always be made to vanish in a similar way to what we did in Sec. XIV C. To see this, note
that
L̂(t)ρ̂(t) = PL(t)Pρ(t) = PL(t)TrB [ρ(t)] ⊗ ρB = −iP [H̃(t), ρS (t) ⊗ ρB ] (717a)
= −i ∑ TrB (Aα (t)ρS (t) ⊗ Bα (t)ρB ) ⊗ ρB − TrB (ρS (t)Aα (t) ⊗ ρB Bα (t)) ⊗ ρB (717b)
α
= −i ∑ [A(t), ρS (t)] ⟨Bα (t)⟩ ⊗ ρB = 0 (717c)
α
since ⟨Bα (t)⟩ can be made zero in the same way as in Eq. (469), i.e., [HB , ρB (0)] = 0.
• Term (b) is an inhomogeneity that depends on the initial condition and measures how much correlation there is in the
initial state:
ρ̄(0) = (I − P)ρ(0) = ρ(0) − TrB [ρ(0)] ⊗ ρB . (718)
It vanishes for a factorized initial state, i.e., if ρ(0) = ρS (0) ⊗ ρB (the same fixed initial state we chose for the bath at the
beginning of the derivation).
Thus, assuming a factorized initial state Eq. (716) becomes:
t
∂t ρ̂(t) = ∫ K(t, t′ )ρ̂(t′ )dt′ (719a)
t0
and hence:
t
∂t [Pρ(t)] = ∂t ρS (t) ⊗ ρB = α2 ∫ PL(t)QL(t′ )Pρ(t′ )dt′ (722a)
t0
t
= α2 ∫ PL(t)L(t′ )Pρ(t′ )dt′ (722b)
t0
t
= −α2 ∫ P [H̃(t), [H̃(t′ ), ρS (t′ ) ⊗ ρB ]] dt′ (722c)
t0
t
= −α2 ∫ TrB [H̃(t), [H̃(t′ ), ρS (t′ ) ⊗ ρB ]] ⊗ ρB dt′ (722d)
t0
where in the second line we used PL(t)P = 0 [Eq. (717)]. Applying one final TrB to both sides finally gives
t
∂t ρS (t) = −α2 ∫ TrB [H̃(t), [H̃(t′ ), ρS (t′ ) ⊗ ρB ]] dt′ , (723)
t0
which we recognize as the Born master equation [Eq. (512)] discussed in Sec. XV B.
where
t
g1 (t, t0 ) = α ∫ L̄(s)ds . (725)
t0
where
for any odd n and any ordering of the time argument, by appropriately shifting the bath operators. Therefore the order O(α3 )
term vanishes, and the Nakajima-Zwanzig master equation is unchanged at this order, namely:
t
∂t ρS (t) = −α2 ∫ TrB [H̃(t), [H̃(t′ ), ρS (t′ ) ⊗ ρB ]] dt′ + O(α4 ) . (729)
t0
105
The Nakajima-Zwanzig equation (719) contains a convolution with a complicated memory kernel [Eq. (719b)]: ∂t ρ̂(t) =
K(t, t′ )ρ̂(t′ )dt′ . It seems that this is an unavoidable feature of an exact, non-Markovian master equation. In this section we
t
∫t0
will see that it is possible to remove the memory kernel by making a type of short-time approximation, and arrive at a fully time-
local, convolutionless master equation. The main insight we’ll need to achieve this, is that the memory kernel can be removed
by formally back-propagating the system state.
A. Derivation
1. Back-propagation
Let us start again from the Liouville-von-Neumann equation ∂t ρ̃ = −iα[H̃(t), ρ̃(t)] ≡ αLρ̃(t) [Eq. (702)]. Its formal solution
is
which defines the backward time-ordered superoperator U− (t, t′ ). To get an explicitly representation note first that by substituting
ρ̃(t′ ) from Eq. (731) into Eq. (730) we get U+ (t, t′ )U− (t, t′ ) = I. Now, since
t − t′
U+ (t, t′ ) = T+ eα ∫t′ L(s)ds = lim eα∆tL(t−∆t) ⋯eα∆tL(t +∆t) eα∆tL(t )
t
(∆t = lim ),
′ ′
(732)
∆t→0 N →∞ N
in order to have U+ (t, t′ ) and U− (t, t′ ) be each other’s inverse, it must be that U− (t, t′ ) has the opposite order and α is replaced
by −α, so that when multiplied the two products cancel equal and opposite terms. I.e.,
t − t′
U− (t, t′ ) = T− e−α ∫t′ L(s)ds = lim e−α∆tL(t ) e−α∆tL(t +∆t) ⋯e−α∆tL(t−∆t)
t
(∆t = lim ).
′ ′
(733)
∆t→0 N →∞ N
Applying P to both sides of Eq. (731), and again dropping the tilde decoration to simplify the notation (though we continue
to work in the interaction picture) we have ρ̂(t′ ) = Û− (t, t′ )ρ(t), so that Eq. (713) becomes:
Note that the superoperator Σ(t) is not chronologically ordered since it contains both forward [via G(t, t′ ); recall Eq. (712)] and
backward time propagation. For this reason we do not write Σ(t, t0 ), despite the dependence of Σ(t) on t0 , since that notation
is reserved for propagation from t0 to t.13 Equation (734a) has removed the memory kernel and replaced it by (the even more
complicated object) Σ(t). However, in terms of the time-dependence of ρ, it is time-local, i.e., depends only on t (apart from
the initial condition t0 ). Next we solve this equation.
13 We could write Σt0 (t) without danger of confusion, but this more cumbersome notation won’t turn out to be particularly helpful.
106
We can solve this for ρ̄(t) provided I − Σ(t) is invertible, i.e., provided Σ(t) is not too close from identity. Since Σ(t0 ) = 0,
we can conclude that I − Σ(t) is invertible for sufficiently short evolution times. In addition, Σ(t) = O(α), so invertibility
should also hold provided the system-bath coupling is sufficiently weak. Thus, from now we shall assume that I − Σ(t) is indeed
invertible, which is the only assumption we shall make to arrive at the TCL master equation. Then:
−1 −1
ρ̄(t) = (I − Σ(t)) G(t, t0 )ρ̄(t0 ) + (I − Σ(t)) Σ(t)ρ̂(t) , (736)
where in the second line we used the freedom to insert a P and Q in front of ρ̂ and ρ̄, respectively. Note that
−1
I + (I − Σ) Σ = (I − Σ)−1 (I − Σ) + (I − Σ)−1 Σ = (I − Σ)−1 (I − Σ + Σ) = (I − Σ)−1 . (738)
where
−1
J (t) ≡ αL̂ (I − Σ(t)) G(t, t0 )Q inhomogeneity , (740a)
−1
K(t) ≡ αL̂(t)(I − Σ(t)) P TCL generator . (740b)
The most salient feature of the TCL-ME is that (when the inhomogeneity vanishes, e.g., for factorized initial conditions) it is
purely time-local, in stark contrast to the NZ-ME [Eq. (719)].
B. Perturbation theory
Despite the formal appearance of the result we have found so far, it is a convenient starting point for perturbation theory.
1. Matching powers of α
n
Let us write (I − Σ(t))−1 = ∑∞
t
n=0 Σ (t), i.e., as a geometric series. It follows from Eq. (734b) that Σ (t) = α (∫t0 ⋯) , so
n n n
that after substitution into K(t) [Eq. (740b)] we have a series expansion in powers of α:
∞
K(t) = αL̂(t) ( ∑ Σn (t)) P (741a)
n=0
∞
= ∑ αn Kn (t) , (741b)
n=1
where we need to determine the operators Kn (t). To do so we need to first expand Σ(t) in powers of α. It also follows from
α t L̄(t′ )dt′
Eq. (734b) that the expansion must start from α1 , since G(t, t0 ) = T+ e ∫t0 = I + O(α) [Eq. (712)]:
∞
Σ(t) = ∑ αm Σm (t) . (742)
m=1
Matching terms of equal power of α with Eq. (741b) yields, for the lowest four orders:
α1 ∶ K1 (t) = αL̂(t)P = 0 (744a)
α ∶ K2 (t) = αL̂(t)Σ1 (t)P
2
Redfield equation (744b)
α ∶ K3 (t) =
3
αL̂(t) (Σ21 (t) + Σ2 (t)) P =0 (744c)
α4 ∶ K4 (t) = αL̂(t) (Σ31 (t) + {Σ1 (t), Σ2 (t)} + Σ3 (t)) P lowest order non-Markovian (744d)
The vanishing of K1 (t) is for the same reason as in Eq. (717); that of K3 (t) is explained below. First we need to explicitly find
the lowest order Σm (t)’s. The expansions of G(t, s) = T+ eα ∫s L̄(t )dt and U− (t, t′ ) = T− e−α ∫t′ L(s)ds [Eq. (733)] yield:
t ′ ′ t
2
α2 t t
G(t, t′ ) = I + α ∫ L̄(s)ds +
T+ (α ∫ L̄(s)ds) + ⋯ (745a)
t′ 2! t′
t 2 t 2
α
Û− (t, t′ ) = P [I − α ∫ L(s)ds + T− (∫ L(s)ds) + ⋯] . (745b)
t′ 2! t′
We can now collect equal powers of α in Σ(t) = α ∫t0 G(t, t′ )L̄(t′ )Û− (t, t′ )dt′ = ∑∞
t
m=1 α Σm (t):
m
t
α1 ∶ Σ1 (t) = ∫ L̄(t′ )Pdt′ (746a)
t0
t t t t
α2 ∶ Σ2 (t) = − ∫ dt′ L̄(t′ )P ∫ L(s)ds + ∫ dt′ (∫ L̄(s)ds) L̄(t′ )P (746b)
t0 t′ t0 t′
t s
=∫ ds ∫ dt′ [L̄(s)L̄(t′ )P − L̄(t′ )PL(s)] , (746c)
t0 t0
where in the last line we switched the order of integration via ∫t0 dt′ ∫t′ ds = ∫t0 ds ∫t0 dt′ .
t t t s
contains the term L̂(t)L̄(t′ )[PL(s)P] = 0. The final term it contains is L̂(t)L̄(s)L̄(t′ )P = PL(t)(I −P)L(s)(I −P)L(t′ )P =
PL(t)L(s)L(t′ )P = 0, by Eq. (728). Therefore K3 (t) = 0.
Let us consider the lowest non-vanishing order of the TCL-ME, Eq. (739). At this order:
∂t ρ̂(t) = α2 K2 (t)ρ̂(t) , (748)
where we have assumed a factorized initial condition, so that the inhomogeneity vanishes. We already found K2 (t) in Eq. (747),
so what remains is to make it explicit using the definition of the projection to the relevant part:
t
K2 (t)ρ̂(t) = ∫ PL(t)L(t′ )dt′ Pρ(t) (749a)
t0
t
=∫ dt′ TrB [−iH̃(t), [−iH̃(t′ ), (TrB ρ(t)) ⊗ ρB ]] ⊗ ρB . (749b)
t0
This is the Redfield equation, Eq. (514). It is identical to the Born-Markov approximation [Eq. (512)], except for the finite upper
limit of the integral. It is also nearly identical to the second order NZ-ME [Eq. (723)], the only difference being the fact that,
by construction, Eq. (750) is time-local, in the sense that the argument of ρS is t rather than t′ . This is an important difference:
whereas when we derived the RWA-LE we had to just assume that we can replace t′ by t [in going from Eq. (512) to Eq. (514)],
here this is a systematic result of our derivation.
108
As an application of the TCL-ME we now consider a qubit in a cavity. This is an analytically solvable model subject to a
simplifying assumption about the initial condition. As such, it will allow us to compare the predictions of the TCL to an exact
result.
Consider as usual the total Hamiltonian H = H0 + HSB , where H0 = HS + HB , with
HS = ω0 ∣1⟩⟨1∣ = ω0 σ+ σ− , HB = ∑ ωk b†k bk = ∑ ωk nk , (751a)
k k
HSB = σ+ ⊗ B + σ− ⊗ B † , B = ∑ gk bk . (751b)
k
Here σ+ = ∣1⟩⟨0∣ and σ− = ∣0⟩⟨1∣ are the qubit raising and lowering operators, while bk and b†k are the bosonic lowering and raising
operators for mode k, satisfying the canonical bosonic commutation relations [bk , b†k′ ] = δkk′ . The gk are coupling constants
with dimensions of energy, and nk is the number operator for mode k. This Hamiltonian describes a qubit (the system) with
ground state ∣0⟩ of energy 0 and excited state ∣1⟩ with energy ω0 coupled to a QHO bath. The coupling either excites the qubit
and removes excitations from the bath, or v.v. It will be useful to think of the bath in this case as electromagnetic modes of cavity.
As usual, let us transform to the interaction picture wrt H0 , so that
H̃(t) = U0† (t)HSB U0 (t) = σ+ (t) ⊗ B(t) + h.c. (752a)
−iωk t
σ+ (t) = e iωt
σ+ , B(t) = ∑ e gk bk . (752b)
k
Then the joint system-bath state ∣φ(t)⟩ (assume it is pure) in the interaction picture is given by ∣φ(t)⟩ = U (t) ∣φ(0)⟩, where
U (t) = T+ exp (−i ∫0 H̃(t′ )dt′ ).
t
This model is not analytically solvable in general. However, we shall assume that the cavity supports at most one photon.
Under this assumption the model becomes analytically solvable, as we shall see.
a. The 1-excitation subspace is conserved Let ∣0⟩B denote the vacuum state of the bath and consider the following joint
system-bath states:
∣ψ0 ⟩ = ∣0⟩S ⊗ ∣v⟩B , (753a)
∣ψ1 ⟩ = ∣1⟩S ⊗ ∣v⟩B , (753b)
∣ψk ⟩ = ∣0⟩S ⊗ ∣k⟩B , (753c)
where ∣k⟩ = b†k ∣v⟩B = ∣01 , . . . , 0k−1 , 1k , 0k+1 , . . .⟩ denotes the state with one photon in mode k (∣k⟩ is not to be confused with
the usual labels for the computational basis of a qubit). Assume that the initial joint system-bath state contains at most a single
excitation, i.e.:
∣φ(0)⟩ = c0 ∣ψ0 ⟩ + c1 (0) ∣ψ1 ⟩ + ∑ ck (0) ∣ψk ⟩ . (754)
k
We wish to show that under the Hamiltonian above this remains true for all times, i.e., for all t:
∣φ(t)⟩ = c0 (t) ∣ψ0 ⟩ + c1 (t) ∣ψ1 ⟩ + ∑ ck (t) ∣ψk ⟩ . (755)
k
This is intuitively clear, since the system-bath coupling either excites the qubit while removing a photon, or v.v., and H0 creates
no new excitations. Nevertheless, let us give a formal argument for completeness.
Define the excitation number operator by
N = σ+ σ− ⊗ I + I ⊗ ∑ b†k bk . (756)
k
N ∣ψ1 ⟩ = (σ + σ− ∣1⟩) ⊗ ∣v⟩ + ∣1⟩ ⊗ ∑ b†k bk ∣v⟩ = ∣1⟩ ⊗ ∣v⟩ = 1 ⋅ ∣ψ1 ⟩ (757b)
k
N ∣ψk ⟩ = (σ + σ− ∣0⟩) ⊗ ∣v⟩ + ∣0⟩ ⊗ ∑ b†k′ bk′ ∣k⟩ = ∣0⟩ ⊗ ∑ δkk′ ∣k⟩ = 1 ⋅ ∣ψk ⟩ , (757c)
k′ k′
109
= σ+ ⊗ B − σ− ⊗ B † − σ+ ⊗ B + σ− ⊗ B † = 0 . (758c)
This means that N is a conserved quantity, i.e., its eigenvalues are conserved under the evolution generated by H, or by H̃(t)
in the interaction picture. It also means that H and N share a common set of eigenvectors, which can be indexed using the
eigenvalues of both H and N . Eigenvectors with different eigenvalues of N don’t mix under the dynamics generated by H
or H̃(t). This explains why, assuming the initial state is Eq. (754), the state subsequently must be as in Eq. (755): the state
∣ψ0 ⟩ has eigenvalue 0 under N [Eq. (757a)] and evolves as a separate one-dimensional subspace, and the states ∣ψ1 ⟩ and ∣ψk ⟩
have eigenvalue 1 under N [Eqs. (757b), (757c)], and also evolve as a separate two-dimensional subspace. U (t) evolves each
subspace separately and does not couple different subspaces labeled by different eigenvalues of N .
Note that i∂t ∣ψ0 ⟩ = H̃(t) ∣ψ0 ⟩ = 0, which means, since ∣ψ0 ⟩ evolves separately, that ∣ψ0 (t)⟩ = ∣ψ0 (0)⟩. Therefore c0 (t) =
c0 (0).
Even though the subspace spanned by {∣ψ0 ⟩ , ∣ψ1 ⟩ , ∣ψk ⟩} contains both 0 and 1 excitations, we loosely refer to it as the
1-excitation subspace.
b. Schrödinger dynamics in the 1-excitation subspace Substituting Eq. (755) into the Schrödinger equation, we have:
= H̃(t) ∣φ(t)⟩ = (σ+ (t) ⊗ B(t) + σ− (t) ⊗ B † (t)) (c0 (0) ∣ψ0 ⟩ + c1 (t) ∣ψ1 ⟩ + ∑ ck (t) ∣ψk ⟩) (759b)
k
= [σ+ (t) ⊗ B(t)] ∑ ck (t) ∣0⟩ ⊗ ∣k⟩ + c1 (t)[σ− (t) ⊗ B † (t)](∣1⟩ ⊗ ∣v⟩) (759c)
k
= ∣1⟩ ⊗ ∑ gk ∣v⟩ ck (t)eiω0 t−iωk t + c1 (t) ∣0⟩ ⊗ ∑ gk∗ ∣k⟩ e−iω0 t+iωk t (759d)
k k
Multiplying by ⟨ψ1 ∣ and ⟨ψk ∣ gives us two coupled differential equations for the amplitudes c1 and ck :
For simplicity, let us assume that the cavity starts in the vacuum state, i.e., ck (0) = 0. Then, after substituting the above into
Eq. (760a) we obtain:
t
c˙1 (t) = − ∫ dt′ f (t − t′ )c1 (t′ ) , (762)
0
where J(ω) is the bath spectral density, formally given as usual by J(ω) = ∑k ∣gk ∣2 δ(ω − ωk ).
110
since the Laplace transform of a convolution of two functions is the product of their Laplace transforms:
t
Lap[∫ dt′ f (t − t′ )c1 (t′ )] = fˆ(s)ĉ1 (s) . (765)
0
= [∣c0 ∣2 + ∑ ∣ck (t)∣2 ]∣0⟩⟨0∣ + c0 c∗1 (t)∣0⟩⟨1∣ + c∗0 c1 (t)∣1⟩⟨0∣ + ∣c1 (t)∣2 ∣1⟩⟨1∣ . (769c)
k
d. Exact master equation To connect the analytical solution to the master equation framework, let us now find the exact
master equation satisfied by ρ(t). To do so, we differentiate Eq. (768), to find:
−∂ ∣c ∣ c0 ċ∗1 (t)
2
ρ̇ = ( ∗ t 1 2) . (770)
c0 ċ1 (t) ∂t ∣c1 ∣
The system-bath Hamiltonian describes an excitation and relaxation process. Therefore, recalling Eq. (282), a reasonable ansatz
for the exact master equation in the interaction picture is of the form
i 1
ρ̇ = − S(t)[σ+ σ− , ρ(t)] + γ(t) (σ− ρ(t)σ+ − {σ+ σ−, ρ(t)}) , (771)
2 2
where the first term represents the Lamb shift and the second term represents relaxation. We will shortly verify this ansatz.
Meanwhile, note that unlike Eq. (282), the relaxation rate γ is now time-dependent. This is an important difference, since there
is now no guarantee that the rate is always positive and finite.
Let us now check and confirm the ansatz. Note that
ρ11 0 ∣c (t)∣2 0
σ− ρσ+ = ( )=( 1 ) (772a)
0 0 0 0
0 −ρ01 0 −c0 c∗1 (t)
[σ+ σ− , ρ] = ( )=( ∗ ) (772b)
ρ10 0 c0 c1 (t) 0
0 ρ01 0 c0 c∗1 (t)
{σ+ σ− , ρ} = ( )=( ∗ ), (772c)
ρ10 2ρ11 c0 c1 (t) 2∣c1 (t)∣2
111
where we used Eq. (768) for the second equality in each line. If Eq. (771) holds then it must be true, using the first equality in
each line of Eq. (772), that:
Comparing the off-diagonal elements of Eqs. (770) and Eq. (773) we find that they agree provided ċ1 = − 12 c1 (t)[γ(t) + iS(t)],
i.e.:
ċ1 (t)
S(t) = −2I ( ) (774a)
c1 (t)
ċ1 (t)
γ(t) = −2R ( ). (774b)
c1 (t)
We have thus identified the Lamb shift rate and relaxation rate from the exact master equation (771).
But, to ensure that the ansatz is correct we still need to confirm that this identification also works for the diagonal elements.
Let c1 (t) = r(t)eiθ(t) . Then ∂t ∣c1 ∣2 = 2ṙr, and also ċ1 = ṙeiθ(t) + iθ̇c1 , which implies ċ1 /c1 = ṙ/r + iθ̇, i.e., R(ċ1 /c1 ) = ṙ/r.
Therefore, if Eq. (774b) holds then:
γ(t)∣c1 (t)∣2 = −2(ṙ/r)r2 = −2ṙr = −∂t ∣c1 (t)∣2 , (775)
as required if Eqs. (770) and Eq. (773) are to agree.
e. Connection with the TCL formalism Note that Eq. (771) is in the form of the TCL-ME, since it is time-local. Namely,
we can introduce a time-local generator and rewrite it as
ρ̇ = KS (t)ρ(t) = TrB [K(t)ρ(t) ⊗ ρB ] , (776)
where K(t) is the TCL generator [Eq. (739)], which can be computed directly from the time-local generator KS (t), which we
identify here as KS (t) = − 2i S(t)[σ+ σ− , ⋅] + γ(t) (σ⋅ σ+ − 12 {σ+ σ−, ⋅}). Next, recall that K(t) = ∑∞
n=1 α Kn (t) [Eq. (741b),
2n
∞
where we have shifted the bath operators so all odd orders vanish]. Correspondingly, KS (t) = ∑n=1 α Kn (t), and therefore
2n
also
∞ ∞
γ(t) = ∑ α2n γ2n (t) , S(t) = ∑ α2n S2n (t) . (777)
n=1 n=1
To make the connection between the exact solution of the qubit-in-cavity model and this perturbative expansion of the TCL-
ME, recall that we started from the Liouville-von-Neumann equation in the form ∂t ρ̃ = −iα[H̃(t), ρ̃(t)] ≡ αLρ̃(t) [Eq. (730)].
This means that if we were to introduce the dimensionless parameter α into the formulation of the qubit-in-cavity model, it
would multiply the coupling constants gk , and hence we would need to replace f (t) with α2 f (t) in Eq. (763). Then Eq. (762)
is replaced by
t
c˙1 (t) = −α2 ∫ dt′ f (t − t′ )c1 (t′ ) . (778)
0
If we consider the Laplace transform solution for c1 (t), given by the inverse Laplace transform of Eq. (767), then to lowest
order in α we simply have c1 (t) = c1 (0). The reason is that the inverse Laplace transform of c1 (0)/s [where have taken α → 0
in Eq. (767)] is c1 (0). Therefore to lowest order in α, Eq. (778) yields c˙1 (t) = −α2 c1 (0) ∫0 dt′ f (t − t′ ) + O(α3 ), and it follows
t
Having derived the exact master equation for a qubit in a cavity, we can now apply it to compare the predictions of various
master equations to the exact solution. To do so we need to specify the bath spectral density J(ω). We will consider the Jaynes-
Cummings model on resonance, a model in which the cavity supports a single mode with a frequency ω0 equal to that of the
qubit. First we consider the case where the cavity is completely isolated from the external world, then we consider the case
where the cavity is coupled to the external electromagnetic field.
112
1. Isolated cavity
Assume that the cavity has opaque walls that act as infinitely tall barriers, so that no radiation can leak into or out of the cavity.
In this case, with ω0 being the qubit transition frequency, since the cavity only has a single mode, at this frequency, the spectral
density becomes
and the amplitude of the qubit excited state, c1 (t), then satisfies
t t
c˙1 (t) = − ∫ dsf (t − t′ )c1 (s) = −∣g∣2 ∫ dt′ c1 (t′ ) . (782)
0 0
Rather than using the Laplace transform solution, it is simpler to differentiate both sides to get
where A and B are constants. Thus, the population of the excited state is ρ11 (t) = ∣c1 (t)∣2 , which oscillates with a period given
by π/∣g∣, as expected from a qubit coupled to an oscillator resonant with it.
2. Leaky cavity
Next we consider the case where, instead of opaque walls, the cavity allows photons to leak out or in. It can be shown that in
this case the memory function is
1
f (t) = e−t/τB (785)
2τM τB
where τM is a Markovian timescale whose exact meaning will become apparent below, and τB is the usual bath correlation time
(decay time of ⟨B(t)B(0)⟩B ). Moreover, it can be shown that α2 = τB /τM , where α is the dimensionless system-bath coupling
strength we have used as a dimensionless prefactor for HSB in the TCL-ME.
The excited state amplitude c1 (t) then satisfies
t 1 t
c˙1 = − ∫ dt′ f (t − t′ )c1 (t′ ) = − dt′ e−(t−t )/τB c1 (t′ ) .
′
∫ (786)
0 2τM τB 0
It is again simpler to differentiate once more rather than use the Laplace transform:
1 1
c̈1 + ċ1 + c1 = 0 , (787)
τB 2τM τB
a simple second order differential equation. Its solution is:
− 2τt tδ 1 tδ
c1 (t) = c1 (0)e B [cosh ( )+ sinh ( )] , (788)
2 τB δ 2
where
√
1 2 1 √
δ= − = 1 − 2α2 . (789)
τB2 τM τB τB
The excited state population is ρ11 (t) = ∣c1 (t)∣2 . We thus have two distinct cases:
113
a. Weak coupling This is the case when α2 = τB /τM ≤ 1/2, so that δ ∈ R. Then S(t) = 0 [Eq. (774a)] and Eqs. (774b)
and (788) yield:
2
τM τB
cosh ( tδ
2
)
γ(t) = (790a)
δ cosh ( 2 ) + τB sinh ( tδ
tδ 1
2
)
− τt tδ 1 tδ 2
ρ11 (t) = ρ11 (0)e B ∣[cosh ( )+ sinh ( )]∣ . (790b)
2 τB δ 2
In this case the population decays, i.e., the dynamics is Markovian-like.
b. Strong coupling This is the case when α2 = τB /τM > 1/2, so that δ ∈ I. Then S(t) ≠ 0 [Eq. (774a)] and Eq. (774b)
and (788) yield:
2
τM τ B
cos ( t∣δ∣
2
)
γ(t) = (791a)
δ cos ( t∣δ∣
2
)+ 1
τB
sin ( t∣δ∣
2
)
2
− τt t∣δ∣ 1 t∣δ∣
ρ11 (t) = ρ11 (0)e B ∣[cos ( )+ sin ( )]∣ . (791b)
2 τB ∣δ∣ 2
In this case the population exhibits damped oscillations, i.e., the dynamics is non-Markovian.
With this analytical solution in hand, we are ready to compare to the predictions of the TCL-ME.
Recall that the TCL-ME expansion is, in the present case, equivalent to an expansion of γ(t) and S(t) in powers of α, as in
Eq (777). We can thus obtain the γ2n (t) terms for the weak coupling case by expanding γ(t) of Eq. (790b) in powers of α, and
similarly for the strong coupling case.
We can also use Eq. (779b), so that:
t t 1 1
γ2 (t) = 2R (∫ dt′ f (t − t′ )) = R (∫ dt′ e−(t−t )/τB ) = (1 − e−t/τB ) ,
′
(792)
0 0 τM τB τM
which is clearly an example of the weak coupling case (as expected for a low-order-in-α expansion) since the rate exhibits no
oscillations. Note that γ2 (t) has a rise time of τB to its asymptotic value of 1/τM .
Recall that the TCL-2 result is exactly the Redfield equation, as we showed in Sec. XX B 2. Moreover, if we take the upper
limit of the integral to infinity we have the Markov limit. Therefore:
which explains the subscript M notation we used all along in this example. We already know the solution in the Markovian
limit: ρ11 (t) = ρ11 (0)e−t/τM .
By doing the expansion to fourth order in α we find:14
1 τM
γ4 (t) = (1 − e−t/τB + [sinh(t/τB ) − t/τB ]e−t/τB ) , (794)
τM τB
which has the limiting behavior γ4 (∞) = τ1M + 2τ1B > γ2 (∞).
What about the NZ-ME? It can be shown that to second order in α, the NZ-ME yields exactly the same result as TCL-2,
except that two changes are needed: (1) e−t/τB is replaced by e−t/(2τB ) in the results for ρ11 (t), and (2) δ is replaced by
√
′ 1 4 1 √
δ = − = 1 − 4α2 . (795)
τB2 τM τB τB
Figure 9 shows these various results in terms of the deviation of the excited state population from the exact result. Focusing
on panels (a)-(c), which report results for the weak coupling case, it illustrates a number of points:
14 Note that the result given in the book [1] differs from Ref. [33][Eq. (69)]; the latter is the correct one.
114
• All approximations, except Markov, are good for very short times (shorter than τB ).
• The Markov approximation initially overestimates the depopulation of the excited state, the underestimates it for longer
times. It is a particularly poor approximation for times shorter than τB , which is the rise-time of the curves in panel (a).
• TCL-2 (Redfield) underestimates the depopulation of the excited state for intermediate times.
• TCL-2 converges to Markov in the long-time limit.
• NZ-2 overestimates the depopulation of the excited state for intermediate times.
• TCL-4 is a better approximation than both the TCL-2 and the Markov approximation. Its rate γ4 (t) goes above the Markov
rate, as expected since γ4 (∞) − =1 2τ1B .
What about the strong coupling case? The exact result is shown in Fig. 9(d), and exhibits damped oscillations. The second
order NZ-ME also exhibits damped oscillations, but the excited state population becomes negative! This result is physically
non-sensical and is a clear example of violation of complete positivity of the evolution map. The TCL-4 approximation is good
for short times but fails to capture the oscillations. To understand this let us take a step back and recall that the TCL-ME requires
FIG. 9. Damped Jaynes-Cummings model on resonance. Exact solution (exact), TCL-ME to second (TCL 2) and fourth order (TCL 4),
NZ-ME to second order (GME 2), and the RWA-LE (Markov). (a) Decay rate of the excited state population, (b) the population of the excited
state, including a stochastic simulation of the TCL-ME with 105 realizations (diamonds for TCL 2and stars for TCL 4), and (c) deviation of the
approximate solutions from the exact result, for 1/γ0 ≡ τM = 5τB (weak coupling). (d) Population of the excited state for 1/γ0 ≡ τM = 0.2τB
(strong coupling). Source: Ref. [33].
115
the invertibility of the operator I − Σ. The present example serves to illustrate how this invertibility condition can be violated,
and how therefore the TCL can break down.
(1) (2) (3)
Assume that for different initial conditions {ρ11 (0), ρ11 (0), ρ11 (0), . . . } there is a common time t0 at which the exact
11 (t0 ) = 0. This is indeed the case shown in Fig. 9(d), as is easy to verify from Eq. (791b): solving for its roots
solution gives ρexact
we have:
2
tan(∣δ∣tn /2) = −∣δ∣τB Ô⇒ tn = (arctan(∣δ∣τB ) + nπ) Ô⇒ t0 = min tn , (796)
∣δ∣ n
where n runs over the integers. Now, since the TCL-ME is time-local, i.e., it only “knows” about the current time t, this means
that for t ≥ t0 it is impossible to invert the evolution back to the initial condition, as this information is lost in a time-local
description. We therefore expect the TCL-ME to give unreliable results when the exact solution predicts a vanishing population.
This is precisely what is seen in Fig. 9(d).
Mathematically, we can see this another way. Eq. (791a) tells us that γ(t) diverges at the same times t = tn where ρexact
11 (t) = 0.
More fundamentally, this is because c1 (t) = 0 implies via Eq. (774b) that γ(t) diverges (unless ċ1 (t) = 0 at the same time). But
if γ(t) diverges then it does not have a Taylor series, so the various γ2n (t) are undefined, and the TCL-ME expansion does not
exist.
We have seen a variety of approaches to describing the reduced system dynamics via master equations, ranging from the
exact Nakajima-Zwanzig equation, via the time-convolutionless, to the Markovian limit. In this section we will review a master
equation approach that naturally interpolates between the Markovian limit and the limit of exact dynamics, as expressed in
terms of CP maps via the Kraus OSR [34]. The key idea will be to understand both limits as arising from a non-selective
measurement process of the bath state. The exact dynamics corresponds to a single measurement at the final time, whereas
Markovian dynamics corresponds to the limit of infinitely many measurements. The interpolation will thus limit the number of
measurements in order to arrive at an non-Markovian approximation.
Consider the usual setup of open system evolution, with the initial state ρ(0) = ρS (0) ⊗ ρB evolving under a joint unitary U
to the final state ρ(t) = U (t)ρ(0)U † (t). The reduced system state at the final time is ρS (t) = TrB [ρ(t)]. We wish to show
that this can be understood equivalently as a projective measurement of the bath at the final time, as depicted schematically in
Fig. 10.
Suppose that we measure the bath at the final time t via the complete set of projection operators {Pk = ∣k⟩⟨k∣}. Thus, if
outcome k was observed, then the joint state transforms as
Pk (IS ⊗ Pk )ρ(t)(IS ⊗ Pk )
ρ(t) z→ ≡ ρ(k) (t) (797)
pk
with probability pk = Tr[(IS ⊗ Pk )ρ(t)]. The reduced system state for this outcome is
Assuming we do not keep track of the measurement outcome, i.e., the measurement is non-selective, the final system state is the
(k)
mixed state ensemble [recall Eq. (68)] {pk , ρS (t)}, i.e.,
(k)
ρS (t) = ∑ pk ρS (t) = ∑ ⟨k∣ ρ(t) ∣k⟩ = TrB [ρ(t)] , (799)
k k
i.e., exactly the Kraus OSR result. Thus we can indeed understand the Kraus OSR as joint unitary evolution followed by a single
non-selective measurement of the bath at the final time t.
In other words, we have shown that the following two evolutions are equivalent:
U (t) TrB
ρ(0) z→ ρ(t) z→ ρS (t) (800a)
U (t) PB TrB (k) non-selective
ρ(0) z→ ρ(t) z→ ρ(k) (t) z→ ρS (t) z→ ρS (t) , (800b)
116
P M (a)
0 t
P M M M M (b)
0 τ 2τ t −τ t
FIG. 10. Measurement approach to open system dynamics. P=preparation, M= measurement, time proceeds from left to right. (a) Exact Kraus
operator sum representation. (b) Markovian approximation.
where τ → 0 and
1
K0 = I + (−iH − ∑ L Lα )τ
†
conditional evolution (803a)
2 α≥1 α
√
Kα = Lα τ , (α ≥ 1) jumps . (803b)
Since we have just shown that each CP map can be understood as a projective measurement of the bath, we see that the LE can
also be understood as representing an infinite sequence of such measurements, taking place in intervals of length τ . Since each
such measurement disentangles the system and bath state, it can be viewed as a preparation step of a new product state between
the system and bath; see Fig. 10.
Having seen that the exact Kraus OSR and the fully Markovian LE are two measurement limits, it is natural to consider an
intermediate scenario, of a finite number of intermediate measurements between the initial and final times. Consider the simplest
case, of a single projective measure of the bath at a random time t′ ∈ (0, t), and note that the more measurements we introduce,
the more Markovian the evolution becomes. We assume that the same CPTP map Λ governs the evolution in the period [0, t′ )
and (t′ , t], as shown in Fig. 11. The measurement produces a random system state ρ(t′ ) (where we from here on we drop the
subscript S since we are interested only in the system dynamics), which is then propagated to ρ(t), i.e., ρ(t) = Λ(t − t′ )ρ(t′ ).
But since we do not know the outcome, nor the time t′ , we introduce a weighting function k(t − t′ , t) (the choice to make the
argument depend on the remaining time interval t − t′ rather than t′ is for later convenience). The final state ρ(t) can then be
represented in the following form:
t
ρ(t) = ∫ k(t − t′ , t) Λ(t − t′ )ρ(t′ )dt′ (804)
0 ´¹¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹¸ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹¶
weight (kernel)
Λ (t ′) ρ (0) Λ (t − t ′) ρ (t ′)
P M M
0 t′ t
FIG. 11. A single projective measurement of the bath is preceded and followed by a CPTP map Λ. For that specific outcome ρ(t′ ) = Λ(t′ )ρ(0)
and ρ(t) = Λ(t − t′ )ρ(t′ ). To account for all possible outcomes each such trajectory is weighted as in Eq. (804).
117
Our purpose is to arrive at a master equation, so let us differentiate Eq. (805) with respect to t:
∂ρ ∂ t
= ∫ k(s, t)Λ(s)ρ(t − s)ds (806a)
∂t ∂t 0
t ∂k(s, t) ∂ρ(t − s)
= k(t, t)Λ(t)ρ(0) + ∫ ( Λ(s)ρ(t − s) + k(s, t)Λ(s) ) ds . (806b)
0 ∂t ∂t
The first term corresponds to performing the bath measurement at t = 0 and then evolving from ρ(0) via Λ(t). This term can thus
be dropped [formally, by setting k(t′ = 0, t) = k(s = t, t) = 0] since we assumed that the intermediate measurement weighted
by k occurs in the open interval (0, t). To make further progress let us specify the form of the CP map Λ. For simplicity, let us
assume that the intermediate evolutions are themselves Markovian:
where L is a Lindbladian, since this is the unique way to ensure that Λ is CPTP in the Markovian case. Then
∂ρ(t − s) ∂eL(t−s)
= ρ(0) = LeL(t−s) ρ(0) = Lρ(t − s) , (808)
∂t ∂t
so that Eq. (806b) simplifies to:
∂ρ t ∂k(s, t)
=∫ ( + k(s, t)L) eLs ρ(t − s)ds . (809)
∂t 0 ∂t
We now seek to ensure that this evolution is trace-preserving. This requires the RHS to be traceless, since then 0 = Tr∂t ρ =
∂t Trρ = 0, so that Trρ(t) = const. It is sufficient to this end to demand that ∂t k(s, t) = 0, since the second term is already
traceless:
t t
Tr [∫ ds k(s, t)LeLs ρ(t − s)] = ∫ ds k(s, t)Tr [LeLs ρ(t − s)] = 0 , (810)
0 0
Now, since ∂t k(s, t) = 0, it follows that k(s, t) = ck(s), where c is a constant we can choose to be 1. Therefore
where in the second line ∗ denotes a convolution. Equation (813) is the PMME.
Now consider two special cases of Eq. (813):
• k(s) = δ(s): In this case the PMME reduces to ∂ρ
∂t
= Lρ(t), which is the standard Lindblad equation. Therefore the
PMME includes the LE as a special case.
t
• Expanding the exponential to zeroth order in L (assuming ∥Lt∥ ≪ 1), the PMME reduces to ∂ρ ∂t
= L ∫0 k(s)ρ(t − s)ds,
which is a form that has been proposed heuristically in the literature on non-Markovian master equations.
• Since the PMME involves a convolution, it can be viewed as a special case of the NZ-ME. Namely, we can write the
PMME in the NZ-ME form ∂t ρ̂(t) = ∫0 K(t, t′ )ρ̂(t′ )dt′ [recall Eq. (719b)], where K(t, t′ ) is directly obtainable from
t
Eq. (813a).
118
To solve the PMME (813) we can use the Laplace transform (764). Recall that the Laplace transform of the convolution of
two functions is the product of their Laplace transforms: Eq. (765), and also recall the result for the Laplace transform of a
derivative in Eq. (766). Therefore, upon taking the Laplace transform of the PMME we find:
but to use it requires a few extra steps, since it is not immediately clear how to deal with eLt in this context. Thus, we find that
it is most convenient to work in the eigenbasis of L. Since L is not normal ([L, L† ] ≠ 0 in general), it can have distinct right
and left eigenvectors, i.e., we can find a set of operators {Ri } and {Li } such that LRi = λi Ri and Li L = λi Li . Both sets are
complete, and they are mutually orthonormal in the sense that after normalization Tr[Li Rj ] = δij .
We can therefore expand ρ in this so-called “damping basis” (the basis of right eigenvectors of L), to get:
where we used eLs Ri = eλi s Ri . Multiplying both sides by Li and taking the trace yields:
∂µi t
= λi ∫ k(s)eλi s µi (t − s)ds. (819)
∂t 0
At this point we can take the Laplace transform of both sides and use the shifting property (815), to get:
Therefore:
1
µ̃i (s) = µi (0). (821)
s − λi k̃(s − λi )
Finally, taking the inverse Laplace transform:
where:
1
ξi (t) ≡ Lap−1 [ ], µi (0) = Tr[Li ρ(0)] . (823)
s − λi k̃(s − λi )
This completes the exact solution of the PMME.
To summarize, given L we need to compute its eigenvalues λi and associated left and right eigenvectors, and given the kernel
k(t) we need to compute its Laplace transform. Using the initial condition ρ(0), we can then compute ξi (t) and µi (t), from
which we obtain ρ(t) using Eq. (816). The kernel k(t) was assumed to satisfy the condition k(0) = 0.
119
where
Let us assume that ξi (t) ≠ 0 ∀t. If this is the case then Φ is invertible, i.e., if we let
then
Φ−1 ○ Φ[X] = ∑ ξi−1 (t)Tr[Li Φ(X)]Ri = ∑ ξi−1 (t)ξj (t)Tr[Lj X]Tr[Li Rj ]Ri = ∑ Tr[Li X]Ri = X , (827)
i ij i
as required. Therefore, using ρ(t) = Φ[ρ(0)] we can write ρ(t − t′ ) = Φ(t − t′ )[ρ(0)] = Φ(t − t′ )Φ−1 (t)[ρ(t)], and so we have:
∂ρ t
= [L ∫ k(t′ )eLt Φ(t − t′ )Φ−1 (t)dt′ ] ρ(t) ≡ K(t)ρ(t) ,
′
(828)
∂t 0
where K(t) is now a convolutionless generator, and Eq. (828) is time-local. Therefore, despite the appearance of the convolution
in the PMME (813), it can be written in TCL-ME form. This is similar to what we did to transform the NZ-ME into the TCL-ME,
where an invertibility assumptions was likewise assumed (recall Sec. XX A 2). It is an interesting open problem to identify the
conditions under which the TCL-ME reduces to the PMME.
Due to the freedom in choosing the kernel k(t), complete positivity is not a guaranteed feature of the PMME. The following
theorem provides us with a way to construct a complete positivity test.
Consider a linear map Φ ∶ Cd×d ↦ Cd×d , i.e., Φ acts on operators represented by d × d matrices, acting on the Hilbert space
H = span{∣i⟩}di=1 . Let us pick ∣i⟩ as a column vector of zeroes, except for a single 1 in position i. Let C = {Φ[∣i⟩⟨j∣]}ij =
∑ij ∣i⟩⟨j∣ ⊗ Φ[∣i⟩⟨j∣]. I.e., C, known as the Choi matrix, is a d × d matrix of the d × d matrices Φ[∣i⟩⟨j∣], meaning that C is d2 × d2 .
Theorem 3 (Choi’s theorem [4]). Φ is completely positive if and only if C > 0.
Constructing the Choi matrix C for the PMME using Eq. (825) we have:
C = ∑ ∣i⟩⟨j∣ ⊗ ∑ ξk (t)Tr[Lk ∣i⟩⟨j∣]Rk = ∑ ξk (t) ∑ ∣i⟩⟨j∣ ⊗ ⟨j∣ Lk ∣i⟩ Rk = ∑ ξk (t) ∑ ∣i⟩⟨j∣(LTk )ij ⊗ Rk . (829a)
ij k k ij k ij
Hence:
Eq. (830) is the complete positivity for the kernel k(t), for a given Lindbladian L and its set of left and right eigenvectors.
To find the left and right eigenvectors of L, consider its action on the Pauli matrices {I, X, Y, Z}:
γ γ
(ZIZ − I) = 0 , LZ = (Z 3 − Z) = 0
LI = (832a)
2 2
γ γ
LX = (ZXZ − X) = −γX , LY = (ZY Z − Y ) = −γY . (832b)
2 2
Thus the Pauli matrices {Ri } = {I, X, Y, Z} are L’s right eigenvectors, with corresponding eigenvalues {λi } = {0, −γ, −γ, 0}.
Representing the Pauli matrices as vectors, i.e., as I = (1, 0, 0, 0)T , X = (0, 1, 0, 0)T , etc., we can write L as a diagonal matrix
with diagonal entries {0, −γ, −γ, 0}. It is then clear that the left eigenvectors are again the Pauli matrices, i.e., in this example
Li = Ri for i ∈ {I, X, Y, Z}, and the condition Tr(Li Rj ) = δij is automatically satisfied.
Let us express the density matrix in terms of the Bloch vector: ρ(t) = 12 (I + v⃗(t) ⋅ σ
⃗ ). The initial condition can then be written
as
1
µi (0) = Tr[Li ρ(0)] = vi (0) (833)
2
where vI (0) = 1.
Let us now assume that the kernel k(t) is:
k(t) = Ae−at . (834)
Recall that Lap(eat ) = 1/(s − a), so that after the Laplace transformation we have
A
k̃(s) = . (835)
s+a
Using Eq. (823) we thus find:
⎡ ⎤
⎢
−1 ⎢ 1 ⎥
ξi (t) = Lap ⎥.
⎢ ⎥ (836)
⎢ s − λi s−λi +a ⎥
A
⎣ ⎦
The L eigenvectors I and Z have the eigenvalue λ = 0, so that:
1
ξI,Z (t) = Lap−1 [ ] = e0t = 1 . (837)
s
The L eigenvectors X and Y have the eigenvalue λ = −γ, so that:
⎡ ⎤
⎢ ⎥
ξX,Y (t) = Lap−1 ⎢⎢
1 ⎥ = e− 21 (a+γ)t (cos ωt + a + γ sin ωt) ,
A ⎥
(838)
⎢ s + γ s+γ+a ⎥ 2ω
⎣ ⎦
√
where ω = 1
2
4γA − (γ + a)2 . Thus the density matrix is
1
ρ(t) = ∑ µi (0)ξi (t)Ri = [I + (vX (0)X + vY (0)Y )ξX,Y (t) + vZ (0)Z] . (839)
i 2
This describes a Bloch vector with fixed Z-component but with X and Y components undergoing damped oscillations with
frequency ω. This is clearly non-Markovian dynamics. The condition for oscillation is 4γA > (γ +a)2 ; otherwise the oscillations
become exponential decay.
Finally, we can use the complete positivity criterion we found above. The Choi matrix is:
C = ∑ ξk (t)LTk ⊗ Rk = ξI I T ⊗ I + ξX X T ⊗ X + ξY Y T ⊗ Y + ξZ Z T ⊗ Z (840a)
k
⎛ 1 0 0 ξX ⎞
⎜ 0 0 0 0 ⎟
= 2⎜
⎜
⎟.
⎟ (840b)
⎜ 0 0 0 0 ⎟
⎝ξ 0 0 1 ⎠
X
Its eigenvalues are easily found to be {0, 0, 2(1 + ξX ), 2(1 − ξX ). Therefore the PMME in this case corresponds to a CP map iff
∣ξX ∣ = ∣ξY ∣ < 1 , (841)
which is a condition on the problem parameters A, a, γ.
121
Since both L and k(t) are phenomenological in the PMME, is there a way we can determine them experimentally? To do so,
we need to express the kernel in terms of measurable quantities. Let us assume that we ρ(t) can be determined via quantum
state tomography, let us guess L (perhaps based on physical intuition as to the prevalent noise). Then we know ρ(t), the initial
condition ρ(0), and the left and right eigenvector sets {Li , Ri }, so that we can compute ξi (t):
which gives us way to compute ξi (t) from purely experimentally measurable quantities. But at the same time ξi (t) is related to
the kernel via Eq. (823). We can invert the latter for k(t) as follows:
˜ = 1 1 1
ξ(s) Ô⇒ k̃(s − λi ) = (s − ), (843)
s − λi k̃(s − λi ) λi ˜
ξi (s)
e−λi t 1
k(t) = Lap−1 [s − ]. (844)
λi ξ˜i (s)
Note that in this expression only the RHS depends on the eigenvalue index i. This gives us an opportunity to optimize the choice
of the Lindbladian by minimizing the deviation for different i values, since they must all agree in order to give a unique result
for k(t). The experimental determination of L and k(t) is thus an iterative process involving this minimization.
122
Everything in this Appendix is about the finite dimensional case, unless explicitly noted otherwise.
1. Inner Product
The inner product of two vectors is a function operating on two copies of a vector space V that outputs a complex number,
f ∶ V × V ↦ C. By definition it must satisfy the following conditions:
n ⎛ w1 ⎞
f (∣v⟩, ∣w⟩) ≡ ∑ vi∗ wi = (vi∗ , ..., vn∗ ) ⎜
⎜ ...
⎟ = ⟨v∣w⟩.
⎟ (A3)
i=1 ⎝ wn ⎠
2. Orthonormal Bases
Two vectors ∣v⟩ and ∣w⟩ are orthogonal if and only if their inner product is zero: ⟨v∣w⟩ = 0 ⇐⇒ ∣v⟩ ⊥ ∣w⟩. The norm of a
vector is
√
∥∣v⟩∥ ≡ ⟨v∣v⟩. (A4)
A unit vector is normalized: ∥∣v⟩∥ = 1. A set of vectors forms a basis if it spans the vector space and is linearly independent.
Using the previous definitions, we can then say that an orthonormal basis is a set of normalized orthogonal vectors that span the
vector space V and are linearly independent:
⎧
⎪
⎪1, if i = j
Orthonormal set: {∣vi ⟩}ni=1 , ⟨vi ∣vj ⟩ = δij , δij = ⎨
⎪
⎩0, if i ≠ j
⎪
3. Linear Operators
Another concept important to our formulation of quantum mechanics is that of linear operators. Consider an operator A that
maps one vector space to another:
A∶V ↦W (A5)
In words, the operator A acting on a linear combination of vectors in the space V produces a linear combination of the operator
acting on each vector individually, and this sum is an element of space W . A good example of a linear operator is the outer
product.
123
4. Outer Product
If we consider vectors ∣v⟩ , ∣z⟩ V and ∣w⟩ W , the outer product of ∣v⟩ and ∣w⟩ is defined as follows:
One important use of the outer product is in the case of expansion in an orthonormal basis. Consider a vector ∣v⟩ V and a set
of vectors {∣i⟩}ni=1 which forms an orthonormal basis set for V . We can equivalently write ∣v⟩ = ∑ni=1 ci ∣i⟩, in which ci is an
arbitrary constant. The inner product of some vector ∣j⟩ with ∣v⟩ produces the coefficient of ∣v⟩ in the given basis:
If we take the outer product of ∣v⟩ with itself, we generate an n × n identity matrix:
⎛ 1 0 ... 0 ⎞
n ⎜ 0 1 ... 0 ⎟
∑ ∣i⟩⟨i∣ = I = ⎜
⎜
⎟
⎟ (A9)
i=1 ⎜ ... ... ... ... ⎟
⎝ 0 0 ... 1 ⎠
We can confirm this is true by applying this inner product as an operator on a vector ∣v⟩:
n n n
(∑ ∣i⟩⟨i∣) ∣v⟩ = ∑ ∣i⟩ ⟨i∣v⟩ = ∑ ci ∣i⟩ = ∣v⟩ (A10)
i=1 i=1 ± i=1
ci
The operator acting on the vector returned the vector, and is known as the “resolution of the identity”. This special case of the
outer product is used to generate a matrix representation of an operator in the appropriate basis. If we consider an operator A
that preserves the space, A ∶ V ↦ V , multiplication of the operator by the identity matrix produces a matrix with elements that
perform the operation A in the following way:
A = IAI (A11)
n ⎛ ⎞ n
= (∑ ∣i⟩⟨i∣) A ∑ ∣j⟩⟨j∣ (A12)
i=1 ⎝j=1 ⎠
= ∑ ∣i⟩ ⟨i∣A∣j⟩⟨j∣ (A13)
i,j ´¹¹ ¹ ¹ ¸ ¹ ¹ ¹ ¶
aij
The scalar aij is known as a matrix element of the operator A. Recall that since the vectors ∣i⟩ and ∣j⟩ are members of an
orthonormal basis, aij ∣i⟩⟨j∣ is actually a matrix with all but the ijth element equal to zero and the (i, j)th element equal to aij :
⎛ 0 ... 0 ⎞
aij ∣i⟩⟨j∣ = ⎜ ⎟
⎜ .. aij ... ⎟ (A15)
⎝ 0 ... 0 ⎠
The sum over all combinations of i and j therefore produces a matrix with elements aij :
It helps us make powerful statements about the properties of vectors in Hilbert space that define the domain of quantum mechan-
⃗ ⋅ ⃗b = ∥⃗
ics. In its elementary form it states that, from the definition of the inner product a a∥∥⃗b∥ cos θ, it follows that the magnitude
of the inner product of those vectors is less than or equal to the product of their norms: ∣⃗ a ⋅ ⃗b∣ ≤ ∥⃗
a∥∥⃗b∥.
We can prove this for Hilbert spaces while demonstrating the power of Dirac notation.
Proof. Pick an orthonormal basis whose first element is ∣1⟩ = ∣w⟩ /∥ ∣w⟩ ∥ (we can always do this using the Gram-Schmidt process
to complete the basis). Then, using the resolution of identity we have
n
⟨v∣w⟩ ⟨w∣v⟩ n
⟨v∣v⟩⟨w∣w⟩ = ⟨v∣I∣v⟩⟨w∣w⟩ = ∑⟨v∣i⟩⟨i∣v⟩⟨w∣w⟩ = ⟨w∣w⟩ + ∑ ∣⟨v∣i⟩∣2 ∥∣w⟩∥2
i=1 ∥∣w⟩∥ ∥∣w⟩∥ i=2 ´¹¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¸ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¶
´¹¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¸ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¹ ¶ ≥0
i=1
= ⟨v∣w⟩⟨w∣v⟩ + positive number (A18)
6. Trace equalities
The following are some useful equalities satisfied by the trace operation. They are easily provable by the rules of matrix
multiplication. A and B are arbitrary matrices of matching dimensions.
7. Positive operators
An operator is positive definite (or positive, for short) if all its eigenvalues are positive. An operator is positive semi-definite
if all its eigenvalues are non-negative. To test this for a given operator A, it suffices to prove that for all vectors ∣v⟩, the diagonal
matrix elements ⟨v∣ A ∣v⟩ are positive or non-negative, respectively. The reason is that this will obviously include the eigenvectors
of A.
8. Pauli matrices
1 0 0 1 0 −i 1 0
σ0 = I = ( ), σ 1 = σx = X = ( ), σ 2 = σy = Y = ( ), σ 3 = σz = Z = ( ). (A20)
0 1 1 0 i 0 0 −1
The last three are traceless by inspection. The Pauli matrices satisfy the identity
where δkl is the Kronecker symbol (it is 1 if k = l, otherwise it is 0), and εklm is the completely anti-symmetric Levi-Civita
symbol [it is 1 if (klm) is an even permutation of (123), −1 if it is an odd permutation, and 0 if any index is repeated].
Since the Pauli matrices are traceless we also have the useful identity
√
Let V an inner product space equipped with the Euclidean norm ∥x∥ ≡ ∑i ∣xi ∣2 ⟨ei , ei ⟩, where x = ∑i xi ei ∈ V and V =
Span{ei }. Let A ∶ V ↦ V. Define
√
∣A∣ ≡ A† A . (B1)
Unitarily invariant norms are norms that satisfy, for all unitary U, V [35]:
∥U AV ∥ui = ∥A∥ui . (B2)
We list some important examples.
1. The trace norm:
∥A∥1 ≡ Tr∣A∣ = ∑ si (A) , (B3)
i
where si (A) are the singular values of A (i.e., the eigenvalues of ∣A∣). If A = ρ is a normalized quantum state, then
∥ρ∥1 = Trρ = 1.
2. The operator norm:
∥Ax∥
∥A∥∞ ≡ sup = max si (A) . (B4)
x∈V ∥x∥ i
Therefore ∥Ax∥ ≤ ∥A∥∞ ∥x∥. Also note that, by definition ∥A∥∞ ≤ ∥A∥1 , since the largest singular value is one of the
summands in ∥A∥1 .
3. The Hilbert-Schmidt norm:
√ √
∥A∥2 ≡ TrA† A = ∑ s2i (A) . (B5)
i
√ √
Again, by definition ∥A∥∞ ≤ ∥A∥2 , since ∑i s2i (A) ≥ maxi s2i (A) = ∥A∥∞ . In addition, ∥A∥21 = ∑i,j si (A)sj (A) ≥
∑i s2i (A) = ∥A∥22 .
We have thus established the ordering
∥A∥∞ ≤ ∥A∥2 ≤ ∥A∥1 . (B6)
All unitarily invariant norms satisfy the important property of submultiplicativity:
∥AB∥ui ≤ ∥A∥ui ∥B∥ui . (B7)
It follows that
∥AB∥∞ ≤ ∥A∥∞ ∥Bi ∥ , ∥B∥∞ ∥A∥i i = 1, 2, ∞ . (B8)
The norms of interest to us are also multiplicative over tensor products:
∥A ⊗ B∥i = ∥A∥i ∥B∥i i = 1, 2, ∞ . (B9)
As an application of unitarily invariant norms, let us revisit the convergence of the iterative expansion we saw in Eq. (448).
We have, for the nth order term:
t t1 tn−1
∥(−iλ)n ∫ dt1 ∫ dt2 ⋯ ∫ dtn [H̃(t1 ), [H̃(t2 ), . . . [H̃(tn ), ρSB (0)]] . . . ] ∥∞ (B10a)
0 0 0
t t1 tn−1
≤ λn ∫ dt1 ∫ dt2 ⋯ ∫ dtn ∥ [H̃(t1 ), [H̃(t2 ), . . . [H̃(tn ), ρSB (0)]] . . . ] ∥∞ (B10b)
0 0 0
t t1 tn−1
≤ λn 2n ∫ dt1 ∫ dt2 ⋯ ∫ dtn ∥H̃(t1 )∥∞ ∥H̃(t2 )∥∞ ⋯∥H̃(tn )∥∞ ∥ρSB (0)∥1 (B10c)
0 0 0
t t1 tn−1
= (2λ)n ∫ dt1 ∫ dt2 ⋯ ∫ dtn ∥HSB ∥n (B10d)
0 0 0
n
t
= (2λ∥HSB ∥)n . (B10e)
n!
To go from Eq. (B10a) to Eq. (B10b) we used the triangle inequality; to go from Eq. (B10b) to Eq. (B10c) we used the fact that
∥[A, B]∥ = ∥AB − BA∥ ≤ ∥AB∥ + ∥BA∥ ≤ 2∥A∥∥B∥ for any unitarily invariant norm; to go from Eq. (B10c) to Eq. (B10d) we
used the fact that ∥H̃(tj )∥ = ∥HSB ∥, since H̃(tj ) = U0† (t)HSB U0 (t) and U0 is unitary. Thus, the norm of the nth order term is
O[(∥HSB ∥t)n ].
126
Consider two quantum states represented by their density matrices ρ and σ. Suppose we perform a POVM measurement with
operators Ei , and obtain measurement outcome i with probability pi for state ρ, and qi for state σ:
pi = Tr(Ei ρ) (C1)
qi = Tr(Ei σ) (C2)
How close are the two outcomes, or equivalently, how close are the two distributions? We address this next.
The total variation distance between two classical probability distributions p = {pi } and q = {qi } is defined as
1
D(p, q) ≡ ∑ ∣pi − qi ∣. (C3)
2 i
The total variation distance measure forms a metric on the space of classical probability distributions, as it satisfies all the three
properties of a metric, viz. the distance between the same variables is zero, it is symmetric, and it satisfies the triangle inequality:
D(x, x) = 0 (C4a)
D(x, y) = D(y, x) (C4b)
D(x, y) ≤ D(x, z) + D(z, y) (C4c)
The trace-norm distance can then be realized as a quantum analogue of the total variation distance.
1
D(ρ, σ) ≡ ∥ρ − σ∥1 (C5)
2
Here we have introduced the one-norm, also called the trace norm, which we define for an arbitrary matrix A:
While we’re at it, there is a useful inequality relating the trace norm and the operator norm [35]:
4. It can be shown that if p and q are the probability distributions of ρ and σ for some POVM, D(ρ, σ) ≥ D(p, q). In other
words, the trace-norm distance is always an upper bound on the corresponding total variation distance. Moreover, there
always exists a POVM which saturates the bound.
Hence, an equivalent definition of the quantum distance measure is
D(ρ, σ) = sup D(p, q) (C10)
{POVM}
This is very useful since we’d like to find a measurement which makes the two states as distinguishable as possible. The
trace-norm distance automatically tells us how far apart the states would be if we could find such a measurement.
2. Fidelity Measures
A fidelity measure can be thought of as an overlap of two states, or the inner product between them. The classical fidelity is
defined as
√ √ √ √
F (p, q) = ∑ pi qi = ( ⃗p, ⃗q) , (C11)
i
√ √ √ √ √ √
i.e., it is the inner product between two vectors ⃗p = ( p1 , p2 , . . .) and ⃗q = ( q1 , q2 , . . .), whose elements are given
by square roots of the elements of classical probability distribution. The fidelity is not a metric since it doesn’t satisfy the
triangle inequality. However, arccos(F ) is a distance, also known as the Bures angle, or Bures length (related to the Bures or
Fubini-Study metric).
A quantum fidelity measure was first introduced by Uhlmann. The Uhlmann’s fidelity between two distribution ρ and σ is
clearly inspired by the classical fidelity, and is given by
√ √
F (ρ, σ) ≡ ∥ ρ σ∥1 . (C12)
Fidelity and distance both give us a sense of how close two states are. While the distance gives us the separation between
two states, fidelity measure the amount of overlap, or similarity of two states. We use two such measures, as while the distance
measure has a nice interpretation as resulting from the optimal POVM, the fidelity measure is often easier to calculate. The two
measures are related by the following inequality [36]:
√ √
1 − F ≤ D ≤ 1 − F 2 ⇐⇒ 1 − D ≤ F ≤ 1 − D2 . (C13)
4. Uhlman’s Theorem
Uhlman’s theorem gives a nice operational interpretation for the fidelity. Consider two states ρ and σ, acting on the same
Hilbert space H1 . Next consider the “doubled” Hilbert space given by H1 ⊗ H2 , where H2 = H1 .
One can always find two pure states ∣Ψ⟩ , ∣Φ⟩ ∈ H1 ⊗ H2 such that
ρ = TrH2 ∣Ψ⟩⟨Ψ∣ (C14a)
σ = TrH2 ∣Φ⟩⟨Φ∣. (C14b)
√ √
Indeed, if the spectral decomposition of ρ is ∑i ri ∣i⟩ ⟨i∣, then ∣Ψ⟩ = ∑i ri ∣i⟩ ⊗ ∣i⟩ yields TrH2 ∣Ψ⟩⟨Ψ∣ = TrH2 ∑ij ri rj ∣i⟩ ⟨j∣ ⊗
√
∣i⟩ ⟨j∣ = ∑ij ri rj ∣i⟩ ⟨j∣ Tr(∣i⟩ ⟨j∣) = ρ, and similarly for σ.
This procedure is called “purification”, and ∣Ψ⟩ is called a purification of ρ. While the purification of a state is not unique
√
(e.g., we could have picked ∣Ψ⟩ = ∑i eiθi ri ∣i⟩ ⊗ ∣i⟩ instead), it can clearly always be found. Uhlman’s theorem states that
F (ρ, σ) = sup ∣ ⟨Ψ∣Φ⟩ ∣ , (C15)
{∣Ψ⟩,∣Φ⟩}
i.e., the fidelity has the appealing interpretation of being the largest possible overlap among the purifications of the two states.
Thus it is also an inner product, just like the classical fidelity in Eq. (C11). Moreover, since ∣ ⟨Ψ∣Φ⟩ ∣ = ∣ ⟨Φ∣Ψ⟩ ∣, clearly
F (ρ, σ) = F (σ, ρ) . (C16)
128
Using the definition of the trace norm and the positivity of ρ and σ, we have
√ √ √ √ √ √ √
∥ ρ σ∥1 = Tr ( ρ σ)† ( ρ σ) (C17a)
√√ √ √ √
= Tr σ ρ ρ σ (C17b)
√√ √
= Tr σρ σ (C17c)
√√ √ √ √
= Tr ρσ ρ = ∥ σ ρ∥1 , (C17d)
where the last line follows from Eq. (C16).
It turns out that, just like the trace distance is the maximum of the classical distance of the probability distributions from
arbitrary POVMs [Eq. (C10)], the quantum fidelity is the minimum of the classical fidelity of the probability distributions from
arbitrary POVMs [3][p.412]:
F (ρ, σ) = inf F (p, q). (C18)
{POVM}
Suppose a pure state ∣ψ⟩ passes through a noise channel N , as depicted below, and we wish to compare the resultant mixed
state ρ = N (∣ψ⟩⟨ψ∣) with the original.
√
In this case we can simplify the expression for the fidelity (note that ∣ψ⟩⟨ψ∣ > 0, (∣ψ⟩⟨ψ∣)2 = ∣ψ⟩⟨ψ∣ ⇒ ∣ψ⟩⟨ψ∣ = ∣ψ⟩⟨ψ∣):
√√ √
F (ρ, ∣ψ⟩⟨ψ∣) = Tr ∣ψ⟩⟨ψ∣ρ ∣ψ⟩⟨ψ∣ (C19a)
√
= Tr ∣ψ⟩⟨ψ∣ρ∣ψ⟩⟨ψ∣ (C19b)
√
= ⟨ψ∣ρ∣ψ⟩Tr(∣ψ⟩⟨ψ∣) (C19c)
√
= ⟨ψ∣ρ∣ψ⟩. (C19d)
It turns out that in this case we can also obtain a tighter inequality than (C13),
1 − F 2 (ρ, ∣ψ⟩⟨ψ∣) ≤ D(ρ, ∣ψ⟩⟨ψ∣). (C20)
If we rotate ρ and σ by the same unitary transformation U , the Fidelity measure doesn’t change, i.e.
F (ρ, σ) = F (U ρU † , U σU † ) (C21)
To prove this, we note that
√the √
trace norm is a unitarily invariant norm, and hence is submultiplicative [recall Eq. (??)]. Also,
√ √ √ √
if A is positive, U AU = (U AU † )2 = U A AU † , so that
†
√ √
U AU † = U AU † . (C22)
Consequently,
√ √
F (U ρU † , U σU † ) = ∥ U ρU † U σU † ∥ (C23a)
1
√ √
= ∥U ρU † U σU † ∥1 (C23b)
√ √
= ∥U ρ σU † ∥1 (C23c)
√ √
= ∥ ρ σ∥1 (C23d)
= F (ρ, σ). (C23e)
129
Consider a noise channel N that is completely positive and trace preserving (CPTP). Such maps can be represented by a set
of Kraus operators {Ki }, such that N (ρ) = ∑i Ki ρKi† and ∑i Ki† Ki = I. CPTP maps are contractive, i.e., they can only make
states become less distinguishable:
D(N (ρ), N (σ)) ≤ D(ρ, σ) (C24a)
F (N (ρ), N (σ)) ≥ F (ρ, σ) (C24b)
As a heuristic justification of these inequalities, consider a completely depolarizing noise channel which maps all states to
identity: N (ρ) = I ∀ρ. Then D(N (ρ), N (σ)) = 0 and F (N (ρ), N (σ)) = 1. At the other extreme, if N is a unitary rotation
(no decoherence), i.e., N (ρ) = U ρU † , then D(N (ρ), N (σ)) = D(ρ, σ) and F (N (ρ), N (σ)) = F (ρ, σ). Other CPTP maps lie
in between these two extremes.
Since the fidelity can only increase under a CPTP map it makes sense to define the fidelity of a noise channel by taking the
minimum over all input states ρ:
F (N ) ≡ inf F (ρ, N (ρ)). (C25)
ρ
Actually we can simplify this somewhat: we can show that the minimization doesn’t require general mixed states, but instead
pure states suffice. The reason that the fidelity satisfies “strong-concavity”, i.e., for any two convex combinations of mixed states
defined over the same index set,
√
F (∑ pi ρi , ∑ qi σi ) ≥ ∑ pi qi F (ρi , σi ) . (C26)
i i i
With this result, and the spectral decomposition ρ = ∑i λi ∣i⟩⟨i∣, we have from Eq. (C25)
F (N ) = inf F (∑ λi ∣i⟩⟨i∣, N (∑ λi ∣i⟩⟨i∣)) (C27a)
ρ
i i
√
≥ inf ∑ λi λi F (∣i⟩⟨i∣, N (∣i⟩⟨i∣)) (C27b)
ρ
i
where in the penultimate line we used the fact that all terms of the form F (∣i⟩⟨i∣, N (∣i⟩⟨i∣) are non-negative, so eliminating all
but the smallest among them certainly makes the expression smaller.
Since every mixed state has a spectral decomposition, the infimum will be achieved for some pure state belonging to the
spectral decomposition of some mixed state. Hence the fidelity of a CPTP noise channel can be redefined as (∣ψ⟩ is a pure state)
√
F (N ) = inf F (∣ψ⟩⟨ψ∣, N (∣ψ⟩⟨ψ∣)) = inf ⟨ψ∣N (∣ψ⟩⟨ψ∣)∣ψ⟩ . (C28)
∣ψ⟩ ∣ψ⟩
Consider a channel that flips the phase of a qubit with probability p, and acts as identity otherwise.
NPD (ρ) = (1 − p)ρ + pZρZ (C29)
The fidelity of this channel can be calculated using Eq. (C28) as
F (NPD ) = inf F (∣ψ⟩⟨ψ∣, NPD (∣ψ⟩⟨ψ∣) (C30a)
∣ψ⟩
√
= inf ⟨ψ∣ NPD (∣ψ⟩⟨ψ∣) ∣ψ⟩ (C30b)
∣ψ⟩
√
= inf (1 − p) ⟨ψ∣ψ⟩ ⟨ψ∣ψ⟩ + p ⟨ψ∣Z∣ψ⟩ ⟨ψ∣Z∣ψ⟩ (C30c)
∣ψ⟩
√
2
= inf (1 − p) + p ⟨Z⟩ (C30d)
∣ψ⟩
130
In this case the minimization is trivial, since, e.g., ⟨+∣Z∣+⟩ = 0. Therefore we have
√
F (NPD ) = 1 − p = 1 − p/2 + O(p2 ) (C31)
We see that the fidelity has been degraded by a term of order p. In other words, the pure-dephasing channel introduces an error
of order O(p) on the system.
If ∣ψ⟩ = a ∣0⟩+b ∣1⟩, we obtain ⟨σ z ⟩ = ∣a∣2 −∣b∣2 , ⟨σ x ⟩ = 2 R(a∗ b) and ⟨σ y ⟩ = 2 I(a∗ b). The minimization over all a and b, subject
to ∣a∣2 + ∣b∣2 = 1, yields a = 1 and b = 0 as one possible solution (the easiest way to see this is to realize that the depolarizing
channel is completely symmetric, so any state will do, e.g., ∣0⟩). Thus,
√
p p
F (NDep ) = 1 − p + = 1 − + O(p2 ) (C34)
3 3
Thus, the error is again O(p).
131
[1] H.-P. Breuer and F. Petruccione, The Theory of Open Quantum Systems (Oxford University Press, Oxford, 2002).
[2] R. Alicki and K. Lendi, Quantum Dynamical Semigroups and Applications (Springer Science & Business Media, 2007).
[3] M. A. Nielsen and I. L. Chuang, Quantum computation and quantum information (Cambridge University Press, 2010).
[4] M.-D. Choi, Linear Algebra and its Applications 10, 285 (1975).
[5] A. Peres, Physical Review Letters 77, 1413 (1996).
[6] M. Horodecki, P. Horodecki, and R. Horodecki, Physics Letters A 223, 1 (1996).
[7] P. Horodecki, M. Horodecki, and R. Horodecki, Physical Review Letters 82, 1056 (1999).
[8] H. Ollivier and W. H. Zurek, Physical Review Letters 88, 017901 (2001).
[9] C. A. Rodrı́guez-Rosario, K. Modi, A.-M. Kuah, A. Shaji, and E. C. G. Sudarshan, J. Phys. A 41, 205301 (2008).
[10] F. Buscemi, Phys. Rev. Lett. 113, 140502 (2014).
[11] J. M. Dominy, A. Shabani, and D. A. Lidar, Quant. Inf. Proc. 15, 1 (2015).
[12] J. M. Dominy and D. A. Lidar, Quant. Inf. Proc. 15, 1349 (2016).
[13] D. A. Lidar, Z. Bihary, and K. Whaley, Chem. Phys. 268, 35 (2001).
[14] L.-M. Duan and G.-C. Guo, quant-ph/9703036.
[15] C.W. Gardiner and P. Zoller, Quantum Noise, Springer Series in Synergetics, Vol. 56 (Springer, Berlin, 2000).
[16] T. A. Brun, Am. J. Phys. 70, 719 (2002).
[17] C. Majenz, T. Albash, H.-P. Breuer, and D. A. Lidar, Phys. Rev. A 88, 012103 (2013).
[18] R. Horn and C. Johnson, Topics in Matrix Analysis, Topics in Matrix Analysis (Cambridge University Press, 1994).
[19] P. Gaspard and M. Nagaoka, Journal of Chemical Physics 111, 5668 (1999).
[20] R. S. Whitney, Journal of Physics A: Mathematical and Theoretical 41, 175304 (2008).
[21] T. Albash, S. Boixo, D. A. Lidar, and P. Zanardi, New J. of Phys. 14, 123016 (2012).
[22] M. Reed and B. Simon, Methods of Modern Mathematical Physics: Fourier analysis, self-adjointness, Vol. 2 (Academic Press, 1975).
[23] J. H. Mathews and R. W. Howell, Complex Analysis: for Mathematics and Engineering, sixth ed. (Jones and Bartlett Pub. Inc., Sudbury,
MA, 2012).
[24] R. Haag, N. M. Hugenholtz, and M. Winnink, Comm. Math. Phys. 5, 215 (1967).
[25] R. Alicki, Physica A: Statistical Mechanics and its Applications 150, 455 (1988).
[26] P. Zanardi and M. Rasetti, Phys. Rev. Lett. 79, 3306 (1997).
[27] D. A. Lidar, I. L. Chuang, and K. B. Whaley, Phys. Rev. Lett. 81, 2594 (1998).
[28] D. A. Lidar and K. B. Whaley, Irreversible Quantum Dynamics, edited by F. Benatti and R. Floreanini, Lecture Notes in Physics, Vol.
622 (Springer Berlin / Heidelberg, 2003) pp. 83–120.
[29] T. Albash and D. A. Lidar, Reviews of Modern Physics 90, 015002 (2018).
[30] W. A. Majewski and R. F. Streater, Journal of Physics A: Mathematical and General 31, 7981 (1998).
[31] R. Alicki, D. A. Lidar, and P. Zanardi, Phys. Rev. A 73, 052311 (2006).
[32] H. Feshbach, Annals of Physics 5, 357 (1958).
[33] H.-P. Breuer, B. Kappler, and F. Petruccione, Physical Review A 59, 1633 (1999).
[34] A. Shabani and D. A. Lidar, Physical Review A 71, 020101 (2005).
[35] R. Bhatia, Matrix Analysis, Graduate Texts in Mathematics No. 169 (Springer-Verlag, New York, 1997).
[36] C. A. Fuchs and J. van de Graaf, IEEE Transactions on Information Theory, IEEE Transactions on Information Theory 45, 1216 (1999).