(Shane Cloude) Polarisation Applications in Remot (BookFi)
(Shane Cloude) Polarisation Applications in Remot (BookFi)
S. R. CLOUDE
1
3
Great Clarendon Street, Oxford OX2 6DP
Oxford University Press is a department of the University of Oxford.
It furthers the University’s objective of excellence in research, scholarship,
and education by publishing worldwide in
Oxford New York
Auckland Cape Town Dar es Salaam Hong Kong Karachi
Kuala Lumpur Madrid Melbourne Mexico City Nairobi
New Delhi Shanghai Taipei Toronto
With offices in
Argentina Austria Brazil Chile Czech Republic France Greece
Guatemala Hungary Italy Japan Poland Portugal Singapore
South Korea Switzerland Thailand Turkey Ukraine Vietnam
Oxford is a registered trade mark of Oxford University Press
in the UK and in certain other countries
Published in the United States
by Oxford University Press Inc., New York
© S. R. Cloude 2010
The moral rights of the author have been asserted
Database right Oxford University Press (maker)
First Edition 2010
All rights reserved. No part of this publication may be reproduced,
stored in a retrieval system, or transmitted, in any form or by any means,
without the prior permission in writing of Oxford University Press,
or as expressly permitted by law, or under terms agreed with the appropriate
reprographics rights organization. Enquiries concerning reproduction
outside the scope of the above should be sent to the Rights Department,
Oxford University Press, at the address above
You must not circulate this book in any other binding or cover
and you must impose the same condition on any acquirer
British Library Cataloguing in Publication Data
Data available
Library of Congress Cataloging in Publication Data
Cloude, Shane.
Polarisation : applications in remote sensing / S.R. Cloude.
p. cm.
ISBN 978–0–19–956973–1 (hardback)
1. Electromagnetic waves—Scattering. 2. Polarimetric remote sensing.
3. Interferometry. I. Title.
QC665.S3C56 2009
539.2—dc22
2009026998
Typeset by Newgen Imaging Systems (P) Ltd., Chennai, India
Printed in Great Britain
on acid-free paper by
CPI Antony Rowe, Chippenham, Wiltshire
An alternative title considered for this book was Which Way is Up? Questions
and Answers in Polarisation Algebra. On advice it was rejected in favour of a
more conventional approach. Still, it is a good question. Which way is up? A
question with a literal scientific interpretation—namely, how to define vertical
in a free reference frame for electromagnetic waves, but also one with a col-
loquial interpretation about the best route to progress. At a technical level this
book is concerned with the answer to the former, but hopefully will serve to
promote in the reader some idea of the latter. It arises from over twenty years’
personal experience of research in the topic, but also through the privilege of
having met and collaborated with many of those who made fundamental con-
tributions to the subject. Much of this original work remains, unfortunately,
scattered in the research literature over different years and journals. This book,
then, is an attempt to bring it all together in a didactic and coherent form suitable
for a wider readership.
The book aims to combine—I believe for the first time—the topics of wave
polarisation and radar interferometry, and to highlight important developments
in their fusion: polarimetric interferometry. Here indeed we shall see that the
whole is greater than the sum of the parts, and that by combining the two we
open up new possibilities for remote sensing applications.
It is intended as a graduate level text suitable for a two-semester course for
those working with radar remote sensing in whatever context, but is also aimed
at working scientists and engineers in the broad church that is remote sensing.
Hopefully it will also appeal to those working in optical physics—especially
polarimetry and light scattering—and to mathematicians interested in aspects
of polarisation algebra.
Before reviewing the structure of the book, certain spelling requires clarifi-
cation. Polarisation or Polarization? The usual response is that British English
uses ‘s’, and American ‘z’. However, in this text we reserve spelling with ‘s’ for
the property of a transverse wave, while we use ‘z’ for the effect of electromag-
netic fields on matter. Hence waves remain polarised while matter is polarized.
In this way we take advantage of both forms.
Chapter 1 first provides an introduction to the physical properties of polarised
waves using the formal machinery of electromagnetic wave theory. The idea is
to provide motivation and a foundation for many concepts used in later chapters.
For example, the concepts of matrix decomposition, the use of the Pauli matrices
in wave propagation and scattering and, most importantly of all, the idea of
using unitary matrices to form a bridge between mathematical descriptions of
polarisation in terms of complex and real numbers, are all introduced in this
chapter. This is in addition to the more prosaic elements of polarisation theory,
such as the polarisation ellipse, the Stokes vector, and the Poincaré sphere, all
of which are covered. The chapter is organized around three main themes: how
vi Preface
One key conclusion we will see from the first four chapters is that entropy or
‘loss of memory’ about polarisation is often linked directly to the randomness
of the scattering medium, and that the remote sensing ‘observer’ has little con-
trol over this. This is problematic for applications, for example, in vegetation
remote sensing, where randomness in the volume leads to loss of polarisation
information. A key idea for the second part of the book is therefore how to
achieve some kind of entropy control in remote sensing of random media. One
way to do this is to employ interferometry. Radar interferometry is a mature
established topic, so in Chapter 5 we provide only a brief introduction for those
not familiar with the key concepts. However, the chapter also contains one or
two novel developments required in later chapters. In particular we develop a
Fourier–Legendre series approach to a description of coherent volume scatter-
ing in interferometry. This then provides a bridge between the two halves of the
book, and allows us to consider, in Chapter 6, the combination of polarisation
diversity with interferometry.
The combination of polarisation diversity with radar interferometry has been
a key development over the past decade. It was first made possible from an
experimental point of view by late additions to the NASA Shuttle imaging radar
mission SIR-C in 1994, and since then has evolved through a combination of
theoretical studies and airborne radar experiments. In Chapter 6 we outline
the basic theory of the topic, showing how to form interferograms in different
polarisation channels before considering mathematically the idea of coherence
optimization, whereby we seek the polarisation that maximizes the coherence
(or minimizes the entropy). In this way we provide a link with earlier chapters by
showing how polarimetric interferometry leads to a form of ‘entropy control’,
even in random media applications.
In Chapter 7 we therefore revisit the ideas of surface and volume scattering
first introduced in Chapter 4, but this time we investigate their properties in both
interferometry and polarimetry. This is built around the idea of a coherence loci,
a geometrical construct to bound the variation of interferometric coherence with
polarisation, and closely related to the coherence region, the latter taking into
account spread due to statistical estimation of coherence from data. Given the
importance of surface/volume decompositions in microwave remote sensing,
we treat in some detail the two-layer scattering problem of a volume layer on
top of a surface and use it to review several model variations that are found in
the literature.
In Chapter 8 we use these ideas to investigate the inverse problem: the esti-
mation of model parameters from observed scattering data. We concentrate on
the two-layer geometry and investigate four classes of problem. We start with
the simplest: estimation of the lower bounding surface position, which is a
basic extension of conventional interferometry and allows us, for example, to
locate surface position beneath vegetation and hence remove a problem called
vegetation bias in digital elevation models (DEMs). We then look at estimating
the top of the layer, which corresponds in vegetation terms to finding forest
height. This is an important parameter for estimating forest biomass, for exam-
ple, and in assessing the amount of carbon stored in above-ground vegetation.
We then look at the possibility of imaging a hidden layer using polarimetric
interferometry. In this case we wish to filter out the scattering from a volume
layer to image a surface beneath. The next logical step is to image the vertical
viii Preface
variation of scattering through the layer itself, and this we treat as the topic
of polarisation coherence tomography or PCT, which combines the Fourier–
Legendre expansion of coherent volume scattering with decomposition theory
in an interesting example of what can happen when two of the major themes of
this book—polarisation and interferometry—are fused.
Finally, in Chapter 9 we turn attention to illustrative examples of these theo-
retical concepts. By far the most important current application area is in radar
imaging or synthetic aperture radar (SAR), and so we begin by reviewing
the basic concepts behind this technology, always highlighting those issues
of particular importance to polarisation. We treat a hierarchy of such imaging
systems, from SAR to POLSAR and POLInSAR, and then consider illustra-
tive current applications in surface, volume, and combined surface and volume
scattering.
We then present supportive material in three Appendices. In the first we
provide a basic introduction to matrix algebra. This is used extensively in
descriptions of polarised wave scattering, and is provided here to help those
not familiar with the terminology and notation employed.
As mentioned earlier, one key idea in this book is the role played by uni-
tary matrix transformations in linking (or mapping) different representations
of polarisation algebra. For this reason, in Appendix 2 we provide a detailed
mathematical treatment of the algebra behind such relationships, introducing
concepts from Lie algebra, group theory, and matrix transformations to illus-
trate the fundamental relationships between complex and real representations
of polarised wave scattering.
Finally, in Appendix 3 we provide a short treatment of stochastic signal
theory as it relates to polarisation and interferometry. Here we treat aspects of
speckle noise in coherent imaging, and show how estimation errors impact on
estimation of scattered field parameters in remote sensing.
This book is the culmination of many years of study and research, and
acknowledgement must be given to those many colleagues and students who
provided the impetus and curiosity to study and develop these topics. Acknowl-
edgements and thanks are extended to the European Microwave Scattering
Laboratory (EMSL) at Ispra, Italy, for their permission to use data from their
large anechoic chamber facility; to the German Aerospace Centre (DLR) in
Oberpfaffenhofen, Germany, for provision of airborne radar data from their
E-SAR system; and to Michael Mishchenko of NASA Goddard Space Center,
USA, for provision of his latest numerical simulations of multiple scattering
from particle clouds. Thanks also to the Japanese Space Agency (JAXA) for
provision of the PALSAR satellite data used in Chapter 9. All these datasets
play a vital role in illustrating the theory outlined in this book, and, I believe,
help enormously in clarifying what would otherwise remain abstract concepts.
Key personal thanks go to five colleagues in particular. Firstly, to Profes-
sor Wolfgang Boerner of the University of Illinois, Chicago, USA. His early
vision and boundless energy have inspired several generations of researchers
in these topics, including my own early studies as a PhD student. Secondly,
thanks to Professor Eric Potter of the University of Rennes, France. Our early
collaboration on radar polarimetry, and particularly on decomposition theory,
was inspiring, and has lead, I am pleased to say, to a lifelong friendship and
collaboration. Thanks also to Drs Irena Hajnsek and Kostas Papathanassiou of
the German Aerospace Centre, DLR. Their support and their contributions to
Preface ix
Bibliography 435
Index 451
Polarised
electromagnetic waves 1
The term ‘wave polarisation’ is relatively recent in the history of optics. It was
first used by Étienne Malus (1775–1812) in 1809, although the ‘orientabil-
ity’ of optical waves was certainly known by Isaac Newton (1643–1727) and
Christiaan Huygens (1629–1695). They were concerned with a description of
the strange phenomenon of double refraction in Iceland spar (calcite), first pre-
sented by Rasmus Bartholin (1625–1698) in 1670, and an explanation was set
to challenge the best minds in optics for the ensuing 150 years. (For an introduc-
tion to the historical importance of polarisation in optics and its role in nature,
see Collet, 1993; Iniesta, 2003; Konnen, 1985.)
It was, however, Thomas Young (1773–1829) who first suggested, in 1817,
that polarisation may arise due to a transverse wave component of light—a
controversial suggestion at the time, but an idea that was further developed
and quantified by Augustin-Jean Fresnel (1788–1827) in 1821, with the devel-
opment of the Fresnel equations for polarisation by surface reflection. This
was followed in 1852 by the development of a mathematical theory of par-
tially polarised waves by George Gabriel Stokes (1819–1903), based on the
concept of a four-element Stokes vector (Stokes, 1852). However, it was
only with the development of the electromagnetic wave theory of James Clerk
Maxwell (1831–1879) in 1861 that light and indeed all electromagnetic waves
were formally shown to be transverse and thus ‘carry a memory of orien-
tation’ in propagating from source to observer (Jones, 1989). The reader
should note, however, that Maxwell’s theory caused some controversy at the
time, and an interesting and readable account of the (sometimes turbulent)
evolution of what we now call Maxwell’s equations can be found in Hunt
(1991).
In this book we concentrate on this orientation ‘memory effect’ and inves-
tigate ways in which it can be used for remote sensing. In a more general
sense, this can be considered a subset of the wider, more formal topic of vector
electromagnetic inverse problems (Boerner, 1981, 1992; Hopcraft, 1992).
In the post-Maxwell era there were four main developments of historical
interest in the description of polarised waves. Firstly we mention the work
of Henri Poincaré (Poincaré, 1892), who formalized many useful concepts in
polarisation optics using a strongly geometrical approach. This was followed
in 1941 by the first use of formal matrix algebra to describe the propagation of
vector waves, by R. Clark Jones of the Polaroid Corporation and Harvard Uni-
versity. At about the same time, Hans Mueller, at the Massachussetts Institute
of Technology, developed a matrix calculus for dealing with partially polarised
2 Polarised electromagnetic waves
Equation (1.1) then relates the radiated vector fields E, B, D, and H to source
vector currents J and scalar charge density ρ. The explicit differential equations
relating these quantities can then be derived from equation (1.1) by treating the
∇ operator as a vector of partial derivative operations and using the following
results from linear algebra:
i j k
∂ ∂ ∂
∇= , , a × b = ax ay az (1.3)
∂x ∂y ∂z bx by bz
The cross-product of ∇ with a vector is called the ‘curl’ (or sometimes ‘rot’ for
rotor) operator, and the dot product the divergence or ‘div’. In this book we
shall be primarily concerned with the ‘memory’ these fields have for the vector
nature of their source (that is, its orientation in space and structure in time), and
how this may be used for remote sensing purposes.
The vector currents J and scalar charges ρ are sources of the fields in equation
(1.1). To demonstrate these as equivalent to a time derivative of current, we
generate a vector wave equation by first forming a secondary vector product
as ∇ × ∇ × E and then using the ∇ × H Maxwell equation plus constitutive
relations to eliminate B. The result is shown on the right-hand side of equation
(1.1). Note that on the left of this equation we have mixed second time and space
derivatives of the electric field vector, while on the right we have the source of
these fields, localized as the time derivative of vector currents. As current itself
is caused by the time derivative of charge, it follows that radiation is caused by
the second time derivative or charge acceleration. This acceleration is a vector
quantity, the orientation of which is transferred into the radiated fields in the
4 Polarised electromagnetic waves
form of propagating waves. While the form of these waves can be very general
(see Cloude, 1995a), it is useful to start with a special type of solution: namely,
vector plane waves.
For plane wave solutions we postulate electric field and driving current vec-
tors of the form shown in equation (1.4). By adopting these simple plane wave
solutions the space and time derivatives take on the simplified form shown on
the right-hand side of this equation:
i ωt−β.r ∂
≡ iω
E = E0e ∂t
⇒ (1.4)
J = J 0 eiωt ∇ ≡ −iβ
Where ω = 2πf and β = 2π/λ, f is the frequency of the √ wave in Hertz, λ its
wavelength in metres, and throughout this text we set i = −1. Note that our
notation, with a positive sign for the time derivative, is chosen by convention
and leads to a complex refractive index for lossy material with a negative
imaginary part (see Section 3.1.1.1). Be aware, however, that other notations
exist in the literature, with some authors choosing E* for the plane wave, which
changes the sign of the time derivative and leads to a complex refractive index
with positive imaginary part (and which also impacts on the sense of circular
polarisations, as we shall see). By direct substitution
we then find that the vector
wave equation in (1.1) has a solution when β = β = ω2 ε0 µ0 , and we obtain
a vector Helmholtz equation of the form shown in equation (1.5), where we
have now eliminated explicit time dependence.
∇ × ∇ × E − β 2 E = iωµo J (1.5)
The importance of such simple vector plane waves follows from the linearity of
Maxwell’s equations since, by superposition, the field at any location x can then
be obtained as a sum of contributions from all the source currents at locations
y. Hence we can express the solution of equation (1.5) as an integral or sum of
the form shown in equation (1.6) (Chen, 1985).
i 1
G(x, y) = I − r r g + (I − 3r r)g − 2 2 (I − 3r r)g (1.7)
kR k R
Here I is the 3 × 3 unit dyad and has the form of a unit matrix, R = |x − y|,
r = (x − y)/|x − y|, and the scalar Green’s function g accounts for causality
1.1 The generation of polarised waves 5
As we move further away from the source currents, then R → ∞ and the
first term of G dominates. Hence in the far-field, the dyadic Green’s function
simplifies by definition to the following form:
e−iβR
G∞ = I − r r (1.9)
4πR
The first part of this expression shows that only the components of J transverse
to the direction of propagation r contribute to the radiated field. It follows from
this that the radiated fields are transverse to the direction of propagation of the
wave (called transverse electromagnetic or TEM waves). The electric field is
defined from an integral sum over all currents, but the resultant must always
lie in a plane perpendicular to r. This is called the plane of polarisation and
the resultant time locus of the electric field in this plane, the polarisation of the
radiated wave.
To illustrate this, consider the fields radiated by elementary dipoles. In
electromagnetic theory there are two types to consider: electric and magnetic
(Jackson, 1999). For the electric dipole, current is localized at the origin, and
an electric dipole moment p0 generates an effective current distribution of the
form shown in equation (1.10). Now evaluating the integral using the far field
Green’s dyadic (equations (1.6) and (1.9)), we obtain the fields radiated by
the dipole as shown in equation (1.10). In the far field, all components have
the structure of transverse electromagnetic (TEM) waves for which the electric
and magnetic field amplitudes are related by the free space wave impedance
Zo ≈ 377 as shown. Note, for example, that the radiation in the direction
r = p 0 is zero (the cross-product is zero), producing the characteristic dumb-
bell radiation pattern. The radiated magnetic field vector can always be derived
from the electric field as shown.
A magnetic dipole, on the other hand, can be generated by a small loop carrying
a uniform current I. The magnetic dipole moment m is then defined from the
product of current and loop area, and is a vector normal to the plane of the loop,
6 Polarised electromagnetic waves
β 2 e−iβR
m H (r) = r × (r × m)
4π µ0 R (1.11)
βωe−iβR
E(r) = − (r × m)
4π R
φ
k
j
φ = −sinφ i − cosφ j
φ
Fig. 1.3 Ludwig system II: spherical polar
i coordinates
This is polarised in the same direction as the antenna current vector. In this way
we can consider the EM wave as transferring a ‘memory’ of the orientation of
the dipole source into the far field, with zero crosspolarisation. Such a con-
venient result does not, however, apply in all coordinate systems, as we now
demonstrate.
pβ 2 e−iβR pβ 2 e−iβR
E= r× r×i = cos θ cos φθ − sin φφ (1.14)
4π ε0 R 4π ε0 R
Here we see that although our source has a well-defined orientation, the radi-
ated field has components that vary with direction and hence are not so neatly
constrained as in the Cartesian case. Although providing a convenient general
format for three-dimensional radiated fields, the spherical polar system is not the
only choice for describing general polarised systems. An alternative, favoured
in the antenna measurement community, is based on a hybrid combination of
Cartesian and polar concepts, considered as follows.
z
a y = sinφ θ + cosφ φ
θ a x = cosφ θ − sinφ φ
y
φ
Fig. 1.4 Ludwig system III: hybrid measure-
ments coordinates x
pβ 2 e−iβR
copolarfield = E.ax = (cos θ cos2 φ + sin2 φ)
4π ε0 R
(1.15)
pβ 2 e−iβR
crosspolarfield = E.ay = sin φ cos φ(cos θ − 1)
4π ε0 R
Note that in the principal planes (when θ and φ are zero) there is zero crosspolar-
isation. However, for radiation in other directions the ratio of cross- to copolar
fields (the XPOL ratio) is given by equation (1.16), which can rise to a maximum
of −15 dB when φ = θ = π/4.
sin φ cos φ(cos θ − 1)
XPOL = 20 log10 (1.16)
cos θ cos2 φ + sin2 φ
This level is often too high for radar and communication applications, and hence
more sophisticated antennas with even lower crosspolarisation have been devel-
oped. To illustrate how such a low crosspolar antenna might be constructed,
consider the case of radiation by a Huygens source (Collin, 1985). This can be
considered a ’patch’ of a plane wave. According to Huygens’ principle, such
a patch radiates elementary secondary wavelets, the superposition of which
marks the advance of the wave front. Figure 1.5 shows such a patch of plane
wave of square dimension 2a, where the fields are constant across the aperture
and zero elsewhere.
The field radiated by such a structure can be obtained from Maxwell’s equa-
tions by employing equivalent electric and magnetic currents Jes , Jms in the
aperture (Collin, 1985; Cloude, 1995a). These are defined from the transverse
components of the field, as shown in equation (1.5). The radiation is then defined
by the expression shown. Note that with a distributed current source such as
1.2 The propagation of polarised waves 9
J es = n × H J ms = n × E
Ex
Es = (1 + cosθ ) f (cosφθ − sin φ φ )
e –ibR
R
Hy sin(b sin θ cos φ a) sin( bsin θ sin φ a)
f=
bsin θ cos φ a b sin θ sin φ a
Fig. 1.5 Radiation by a Huygens patch: the
Huygens Source ideal zero cross-polarised source
this, the radiation integral can be explicitly evaluated and produces a Fourier
Transform relation between the aperture distribution and the far field. In this
case the rectangular distribution gives rise to a SINC function. However, for
our purposes, interest centres more on the polarisation properties of the radiated
field.
From the polarisation point of view we observe a very interesting result. The
radiation from this ‘aperture antenna’has zero crosspolarisation in all directions.
This shows that in theory, low crosspolarisation can be obtained, although
in practice securing the right kind of symmetric aperture distribution can be
difficult to engineer, especially over a broad band of frequencies (Collin, 1985;
Mott, 1992).
Having established the influence of coordinate systems on the definition
of co- and crosspolarised waves in free space, we now turn to consider the
propagation of waves in more complex environments. In particular, we consider
constraints posed by the presence of the medium on the allowed polarisation
states of the propagating field, and thus establish a calculus for dealing with
the distortion of the ‘memory’ effect in the transfer of orientation information
from source to far field.
As when considering the radiated field, we first generate a set of vector homoge-
neous wave equations from the Maxwell curl equations. The resulting systems
10 Polarised electromagnetic waves
for each of the three classes of material are shown in equation (1.18).
∂ 2E
∇ × ∇ × E + εµ0=0 Case I
∂t 2
∂ 2E
∇ × ∇ × E + εµ0 2 = 0 Case II (1.18)
∂t
∂ ∂ 2E
∇ × ∇ × E − µ0 (η + γ ) ∇ × E + εµ0 2 = 0 Case III
∂t ∂t
With these two assumptions we can simplify the homogeneous wave equations
to the general form shown in equation (1.20):
∇ × ∇ × E − ω2 εµ0 E = 0 Case I
∇ × ∇ × E − ω2 εµ0 E = 0 Case II (1.20)
∇ × ∇ × E − iωµ0 (η + γ ) ∇ × E − ω2 εµ0 E = 0 Case III
∇ × E = ei(ωt−βz) iβ(ey i − ex j)
(1.21)
∇ × ∇ × E = ei(ωt−βz) β 2 (ex i + ey j)
D = εE → ∇ × ∇ × E − ω2 µεE = 0 (1.22)
In order for the plane wave to be a solution, its components must then satisfy
the following equation set obtained by explicit evaluation of equation (1.22):
ex ex 0
β 2 ey − ω2 εµ0 ey = 0 ⇒ β 2 = ω2 εµ0 = β0 n ⇒
0 ez 0
1 c c
f =√ √ =√ = (1.23)
εr ε0 µ0 λ εr λ nλ
Note that ez = 0; that is, these plane wave solutions represent transverse elec-
tromagnetic (TEM) waves. These waves are also non-dispersive; that is, they
all propagate with the same phase velocity, itself determined from the free
space velocity c = 2.997 × 108 m/s, and the refractive index n of the medium,
which is related to the square root of the dielectric constant εr , as shown in
equation (1.23).
These constraints do not specify ex and ey . In fact, any complex pair will
satisfy the wave equation. This we call a C2 symmetry, in that any element of
a two-dimensional complex space is a solution. Note, however, that the pair
(ex , ey ) are independent of time and space, and therefore represent a spatio-
temporal invariant of the wave. They define the polarisation of the plane wave.
Since their resultant always lies in the xy plane transverse to the propagation
direction, this is now called the plane of polarisation of the wave.
Without loss of generality we can write the pair as a column vector in C2—
the space of two-dimensional complex numbers—as shown in equation (1.24),
where m is the amplitude of the wave, and the trigonometric factors arise directly
from the requirement that w itself has unit amplitude, or is unitary.
ex 2 cos αw eiφx cos αw eiφx
E0 = = |ex | + ey
2 =m = mw
ey sin αw eiφy sin αw eiφy
(1.24)
We often wish to compare waves with the same amplitude, and therefore set
m = 1. In this case the column vector is unitary and has three free parameters.
Importantly, each choice of unitary vector w then defines a new class of vectors
w⊥ , being orthogonal to the first. As is conventional for complex vector spaces,
orthogonality is based on the Hermitian inner product of column vectors, as
shown in equation (1.25):
cos αw eiφx ∗T iχ − sin αw e
iφx π
w= ⇒ w ⊥ · w = 0 ⇒ w ⊥ = e ⇒ α⊥ = αw +
sin αw eiφy cos αw eiφy 2
(1.25)
We see that the orthogonal state is not uniquely defined. There is a phase angle
χ left undetermined from w by the combined Hermitian and unitary constraints.
This problem can be resolved by considering how the pair w and w⊥ are to be
combined to provide a coordinate system or polarisation basis or frame for the
representation of arbitrary wave states.
12 Polarised electromagnetic waves
We must still deal with the free parameter χ . One way to resolve this issue is
to force the matrix U to be special unitary; that is, to have unit determinant.
This not only establishes a consistent method for change of base but, as shown
in Appendix 2, links directly via group theory to the geometry of the real space
of the Poincaré sphere and Stokes vector. With this added condition we obtain
the following constraint equation for χ :
Consequently the general special unitary change of base matrix can be written
as shown in equation (1.28):
cos αw eiφx − sin αw e−iφy
[U2 ] = (1.28)
sin αw eiφy cos αw e−iφx
This is generally made complicated because the ε tensor is full. In this case
it is more convenient to rewrite equation (1.31) in terms of the electric dis-
placement vector D rather than E. We then obtain the modified form shown in
equation (1.32):
−1
ε11 ε12 ε13 dx dx 0
∇ × ∇ × ε21 ε22 ε23 . dy − ω2 µ0 . dy = 0 (1.32)
ε31 ε32 ε33 dz dz 0
For plane wave solutions, the vector on the far left of this expression has only x
and y components, from which it follows that dz = 0; that is, that the D vector
(not the E vector) is always transverse to the direction of propagation. For this
reason D is often preferred to the electric field E when describing the polar-
isation of waves in anisotropic media. Now assuming that our external wave
system xyz corresponds to abc we obtain the following simplified dispersion
relation:
1
εa 0 dx dx 0
β 2
1 − ω 2 µ0 . = (1.33)
0 εb
dy dy 0
14 Polarised electromagnetic waves
We see that in this case we no longer have the C2 freedom of isotropic mate-
rial, and that for a wave to propagate it must be polarised along the a (x) or
b (y) directions. Furthermore, the velocity of propagation is different for the
two waves—a phenomenon that leads to differential phase shifts between com-
ponents of the wave, and is known as birefringence. Any general polarisation
state can be expressed as a linear mixture of a and b through the basis projec-
tion matrix of equation (1.28). Thus, when a polarisation state is launched at
z = 0 then its a and b components will propagate at different velocities (and
also in general with different extinction rates), and hence as it progresses into
the material it will change its polarisation state. The only exceptions to this
are the states a and b themselves. If they are launched into the material then
they progress without distortion. If we represent the effect of propagation up
to a plane z = z0 as a 2 × 2 complex matrix [Mz0 ] we can write the following
eigenvalue problem:
E (z0 ) = Mz0 .E (0) = λE (0) ⇒ Mz0 − λ [I2 ] E(0) = 0 (1.34)
We then see that the states that remain unchanged due to propagation are eigen-
vectors of the matrix [Mz0 ]. Consequently we refer to these as eigenpropagation
states, or simply eigenstates, of the material. We now show how [Mz0 ] can be
related to the electric field wave equation.
Returning to equation (1.31) for the electric field, and now imposing the
constraint that dz = 0, we can remove the ez dependence and obtain a pair of
equations for ex and ey only. The following equation is then obtained for an
arbitrary polarisation state, where in the last step we have expressed the spatial
term as an ordinary derivative with respect to z, itself obtained from integration
of the second derivative appearing from the vector wave equation (assuming
[Kz ] does not depend on z).
|ε13 |2 ε13 ε23
2 ex
ε11 − ε ε12 −
ε ex 0
β − ω µ0
2 · e =
2 33 33
ε ∗ ε∗
ey |ε | y 0
ε12∗ − 13 23
23
ε22 −
ε33 ε33
0
⇒ β 2 E − ω2 µ0 [Kz ] .E =
0
d 2E
⇒ = −ω2 µ0 [Kz ] · E
dz 2
dE
⇒ = [N ] · E = −iω µ0 [Kz ] · E (1.35)
dz
The most important part of the above analysis is the derivation of a simple matrix
differential equation governing the propagation of the C2 column vector E in
terms of a differential matrix [N ], which may be easily integrated to obtain the
[M ] matrix at distance z0 , as shown in equation (1.36).
dE z0
= [N ] E ⇒ Mz0 = [M0 ] exp [N ]dz (1.36)
dz 0
1.2 The propagation of polarised waves 15
[A]2 z 2 [A]n z n
exp([A]z) = I + [A]z + + ··· + ··· (1.38)
2! n!
We shall now make use of the following six important properties of the
matrix exponential function, where the matrix commutator bracket is defined
as [A, B] = AB − BA. We see from property II that the eigenvectors of [Mz ]
and [N ] are identical, and that the eigenpolarisation states are determined by
the eigenvectors of the reduced dielectric tensor [Kz ] in equation (1.35).
In the special case of zero absorption by the material, [Mz ] must be unitary
(norm-preserving). If this is the case then its inverse is just its conjugate trans-
pose, and from property IV it follows that [N ] = i[H ] where [H ] is Hermitian.
If the matrix [Mz ] is special unitary (that is, with unit determinant) then from
property III it follows that the matrix [N ] must also be traceless. Note that we
can always factor a determinant phase term from a unitary propagation matrix
[Mz0 ] to leave a special unitary form, as shown in equation (1.40):
e−iβa z0 0
[Mz0 ] = [U2 ] · . [U2 ]∗T
0 e−iβb z0
(βa −βb )
−i(βa +βb ) e−i 2 z0 0
=e 2
z0
[U2 ] · (βa −βb ) z · [U2 ]∗T (1.40)
0 ei 2 0
cos αw eiφx − sin αw e−iφy
[M2 ] = [U2 ] = = exp(iH )
sin αw eiφy cos αw e−iφx
h1 h2 − ih3
⇒ [H ] = (1.41)
h2 + ih3 −h1
where the three coefficients h1 , h2 and h3 are all real. This last result introduces
the idea of matrix decomposition to polarimetry. In principle, we take a com-
plex matrix (such as [M2 ]) and express it as the sum of component matrices,
each of which has some simpler physical interpretation. In this way we can
‘model’ the processes giving rise to the observed matrix in terms of a combina-
tion of elementary physical mechanisms. To see this, note that the matrix [H ]
can be formally expressed as a linear combination of elementary matrices as
follows:
!
3
1 0 0 1 0 −i
[H ] = hl [σl ] = h1 + h2 + h3 (1.42)
0 −1 1 0 i 0
l=1
The triplet of matrices σ =[σ 1 , σ 2 , σ 3 ] are called the Pauli spin matrices, as
they were first applied to problems of spin in quantum mechanics by Wolfgang
Pauli (1900–1958). More generally, as we shall see, they are useful for decom-
posing classical vector wave scattering problems involving complex matrix
transformations.
Considering each elementary Pauli matrix at a time, we can use the series
expansion of equation (1.38) to derive the corresponding unitary matrices. The
key stage is to derive the square of the elementary matrix, and we note that for
all three Pauli matrices we have the following relation:
1 0
σi2 = = σ0 (1.43)
0 1
θ 2 σ12 θ n σ1n
exp(iθ σ1 ) = σ0 + iθ σ1 − + · · · (i)n + ···
2! n!
θ2 θ3
= σ0 1 − + · · · + iσ1 θ − + ···
2! 3!
cos θ + i sin θ 0
= cos θ σ0 + i sin θ σ1 =
0 cos θ − i sin θ
iθ
1 0 e 0 1 0
= . .
0 1 0 e−iθ 0 1
θ 2 σ22 θ n σ2n
exp(iθ σ2 ) = σ0 + iθ σ2 − + · · · (i)n + ···
2! n!
θ2 θ3
= σ0 1 − + · · · + iσ2 θ − + ···
2! 3!
(1.44)
cos θ i sin θ
= cos θ σ0 + i sin θ σ2 =
i sin θ cos θ
iθ
1 1 −1 e 0 1 1
= . .
2 1 1 0 e−iθ −1 1
θ 2 σ32 θ n σ3n
exp(iθ σ3 ) = σ0 + iθ σ3 − + · · · (i)n + ···
2! n!
θ2 θ3
= σ0 1 − + · · · + iσ3 θ − + ···
2! 3!
cos θ sin θ
= cos θ σ0 + i sin θ σ3 =
− sin θ cos θ
iθ
1 1 i e 0 1 −i
= · ·
2 i 1 0 e−iθ −i 1
[H ]2 = (h1 σ1 + h2 σ2 + h3 σ3 ) . (h1 σ1 + h2 σ2 + h3 σ3 )
from which we see it is convenient to define the scalar amplitude of the matrix
[H ] as θ and to normalize the vector of coefficients h = θn where n · n = 1.
With this modification the series again simplifies into elementary trigonometric
18 Polarised electromagnetic waves
functions as follows:
θ 2 (n.σ )2 θ n (n.σ )n
exp(iθn.σ ) = σ0 + iθn.σ − + · · · (i)n + ···
2! n!
θ2 θ3
= σ0 1 − · · · + in.σ θ − + ···
2! 3!
= cos θ σ0 + i sin θn.σ
cos θ + i sin θn1 i sin θ (n2 − in3 )
=
i sin θ (n2 + in3 ) cos θ − i sin θn1
eiθ 0
= [U2 ] [U2 ]∗T (1.46)
0 e−iθ
This represents the most general special unitary matrix and an alternative
parameterization to that used in equation (1.28). We shall see in Section 1.3
that there is a simple geometrical interpretation of both sets of parameters in
terms of spherical trigonometry on the Poincaré sphere. From the form of the
eigenvalue decomposition we can see that the general unitary matrix repre-
sents birefringence between a pair of orthogonal elliptical polarisations. Such
a propagation channel is called a retarder, and θ is called the retardence of the
channel.
where Ne is the electron number density in the material (between 1010 and
1012 m−3 for the ionosphere) and e/m is the charge-to-mass ratio for an elec-
tron. Such a material, although frequency-dispersive, is isotropic, and therefore
does not distort the polarisation of the propagating wave. However, in the pres-
ence of an applied DC magnetic field the situation changes. Here we restrict
attention to the case where the DC field is applied along the z-direction (along
the direction of propagation for our plane wave). In this case the effect of an
electric field depends on its polarisation, and the medium becomes gyrotropic
with a dielectric tensor of the form shown in equation (1.48) (Ishimaru 1991,
1.2 The propagation of polarised waves 19
Chapter 8):
ωp2
εa iεb 0
εa = 1 −
ω2 − ωc2
ε = ε0 −iεb εa 0 ⇒ (1.48)
−ωc ωp2
0 0 εr
εb =
ω(ω2 − ωc2 )
eB0
ωc = (1.49)
m
To give a typical example, the Earth’s magnetic field strength is around 5 ×
10−4 Tesla, which leads to a cyclotron frequency of 1.4 MHz. Considering
propagation in the z direction, we can now use equation (1.48) to generate the
2 × 2 [Kz ] matrix directly from this tensor, as shown in equation (1.50):
εa iεb
[Kz ] = ε0
−iεb εa
ε0 1 i ε − εB 0 1 −i
= . a . (1.50)
2 i 1 0 εa + εB −i 1
x
Fig. 1.6 Definition of left- and right-hand
circular polarisations
and finally we obtain the propagation matrix [Mz ] using the exponential function
as shown in equation (1.52):
We can see that the right circular wave has a resonance when ω = ωc . This
wave is forcing the electrons to move in their ‘natural’ direction about the
magnetic field (according to the Lorentz force equation F = q(E + v × B)). For
this reason it is called the extraordinary wave. The left circular wave, on the
other hand, forces the electrons in the opposite direction and therefore shows
no resonance. It is termed the ordinary wave. Note that when ω is less than
some critical frequency ω1 then βL becomes imaginary and the ordinary wave
does not propagate. The cut-off frequency can be easily obtained from equation
(1.53), as shown in equation (1.54):
"
ω2 ωc
ω1 = ωp2 + c − (1.54)
4 2
Importantly, the extraordinary wave can propagate at low frequencies when the
ordinary wave is below cut-off. Hence low-frequency waves can penetrate the
ionosphere along lines of the Earth’s magnetic field. This is the main mechanism
behind the low-frequency whistler mode of atmospheric propagation. These
results are summarized in Figure 1.7. Here we show typical dielectric constant
variation with frequency and polarisation.
We see that the ordinary wave has a relatively simple behaviour with a cut-off
frequency of ω1 . The extraordinary wave shows more complex behaviour, with
1.2 The propagation of polarised waves 21
4
Ordinary wave (left handed)
Extraordinary wave (right handed)
3
Dielectric constant
vp
1+4
2
vc
c/2
1
vc vc
v1 = v2p + –
4 2
0
1 2 v2 vc
v2 = v2 + +
4 2
–1
0.5 1 1.5 2 2.5 3
Normalized frequency v
vc
Fig. 1.7 Vector propagation modes in
gyrotropic media
two branches to its propagation behaviour, one at low frequencies, and one
at high. Note that at high frequencies (compared to ωc ) the medium becomes
isotropic and transparent with εr = 1. There is, however, a second impor-
tant polarisation phenomenon arising from this result: the distortion of linear
polarisations as they propagate via Faraday rotation, as we now discuss.
Surprisingly, the rotation is in the same direction as for +z; that is, if the wave
first propagates through the medium and is then returned to its starting point
then the Faraday rotation is not cancelled but doubled, since
2
cos θF − sin θF
[Mz ][M−z ] = = e−iθF σ3 e−iθF σ3
sin θF cos θF
cos 2θF − sin 2θF
= e−i2θF σ3 = (1.58)
sin 2θF cos 2θF
This can be traced to the presence of the DC magnetic field, which has a
polarity of its own and causes this lack of reciprocity. This is in contrast to
a second type of circularly polarised wave propagation that occurs in many
natural media, such as sugar solutions (optical activity) and in manmade chiral
materials such as helical microwave dielectric composites. Here again, circular
eigenpolarisations are generated, but this time the effect does not double with
space reversal and has a fundamentally different physical origin, as we now
consider.
D = εE + ηB
(1.59)
H = γ E + µ−1
0 B
For simplicity we here consider the case of lossless chiral media where η = γ
and both are purely imaginary, so the constitutive equations have the special
form shown in equation (1.60):
D = εE − iγ B
(1.60)
H = −iγ E + µ−1
0 B
These show that D not only depends on the local value of E at a point in the
material, but also on neighbouring values through the local spatial derivative of
E. This is termed spatial dispersion, and is characteristic of this type of material.
With this notation established we now return to the vector wave equation for
plane waves in such media, and obtain
Performing the spatial derivatives for the plane wave we obtain the following
matrix equation for the electric field components:
ex ey ex 0
β 2 ey − 2iβωµ0 γ −ex − ω2 εµ0 ey = 0 (1.64)
0 0 ez 0
24 Polarised electromagnetic waves
from which see that, unlike the case for anisotropic material, ez = 0 is always
true and so these are TEM waves. We can now obtain the [Kz ] matrix by
inspection, as shown in equation (1.65):
2iβγ
1 εω 1 iεb
[Kz ] = ε =ε
− 2iβγ
εω 1 −iεb 1
ε 1 i 1 − εb 0 1 −i
= · . (1.65)
2 i 1 0 1 + εb −i 1
The eigenvector decomposition again yields left and right circular eigenpo-
larisations and differential propagation phase (circular birefringence) due to a
splitting of the eigenvalues. Note that because of spatial dispersion this matrix
is itself a function of the desired unknown wavenumber β. The [N ] matrix can
be obtained by taking the square root of [Kz ], as shown in equation (1.66):
√
[N ] = −iω µ0 [Kz ]
√ √
iω µ0 ε 1 i 1 − εb −i
=− . √ 0 .
1
(1.66)
2 i 1 0 1 + εb −i 1
Finally, by using the exponential function we obtain the [Mz ] matrix for
propagation to z, as shown in equation (1.67):
>0 βR > βL and the phase velocity for RHC is slower than LHC.
<0 βR < βL and the phase velocity for LHC is slower than RHC.
The parameter β0 is typically small, lying in the range 10−6 < |β0 | < 10−4
for natural materials. In practice we can therefore expand the square root and
obtain the following simplified relationships:
βL ≈ β0 (1 − β0 ) n ≈ (1 − β0 ) nR − nL
⇒ L ⇒ β0 ≈ (1.69)
βR ≈ β0 (1 + β0 ) nR ≈ (1 + β0 ) nR + nL
1.2 The propagation of polarised waves 25
where nr is the refractive index of the material for right-hand circular polarisa-
tion. Thus can be obtained experimentally by measuring the refractive index
of the material for right and left circularly polarised wave transmission. Alter-
natively, the chiral parameter can be estimated by measuring the rotation of a
linearly polarised wave. This rotation effect is similar in form to that for Fara-
day rotation, and can be made explicit by using the Pauli matrix representation
of [N ] which leads to an [Mz ] matrix of the form
−i l 2 r z cos θA − sin θA
β +β βl +βr
[Mz ] = e = e−i 2 z e−iθA σ3 (1.70)
sin θA cos θA
where
1 π
θA = (βL − βR )z ≈ −β02 z = z (nR − nL ) (1.71)
2 λ
This rotation is clockwise when is positive, and anticlockwise when is
negative (looking into the source of the wave). To compare different materials,
the specific rotatory power is defined as θA /z. This has a value ranging from
around 20 degrees/mm for solids such as quartz, down to 0.4 degrees/mm for
liquids such as turpentine.
Note that a key difference between this and Faraday rotation is the behaviour
under inversion of space coordinate. In this case, when we consider propaga-
tion in the −z direction, θA changes sign, and hence this type of rotation is
cancelled with propagation back through the medium. We now turn to consider
a generalization of these ideas to enable us to model propagation in arbitrary
media.
!
3
h0 + h1 h2 − ih3
[N ] = i[H ] = i hj σj = i (1.73)
h2 + ih3 h0 − h1
j=0
The study of general solutions of this equation was first carried out in optics in
Jones (1941, 1948), and has subsequently been termed the Jones calculus. In this
section we outline the general structure of this method and detail a classification
of different propagation channels based on the eigenvector decomposition of
the propagation matrix (Azzam, 1987; Lu, 1994).
To complete our study of wave propagation, we first consider generalization
to the case where the propagation channel involves losses and the matrix [Mz ]
26 Polarised electromagnetic waves
cosh θ + sinh θn1 sinh θ (n2 − in3 )
|M2 | · exp(θn · σ ) = |M2 | ·
sinh θ (n2 + in3 ) cosh θ − sinh θn1
θ
−φ e 0
= e . [U2 ] . · [U2 ]∗T
0 e−θ
cos θ + i sin θn1 i sin θ (n2 − in3 )
|M2 | · exp(iθn.σ ) = |M2 | ·
i sin θ (n2 + in3 ) cos θ − i sin θn1
iθ
−iφ e 0
= e . [U2 ] . · [U2 ]∗T
0 e−iθ
(1.75)
In microwave remote sensing such a situation can also arise for propaga-
tion through an oriented grid or volume of aligned scatterers, such as occurs
in the growth of many aligned agricultural crops such as wheat stalks (see
Section 3.5.2).
Hence, in general, a propagation channel will show both retardence and diatten-
uation and the [M2 ] matrix will be neither pure Hermitian nor unitary. However,
by employing a polar decomposition of the [M2 ] matrix we can always express
a general channel as an equivalent cascade of just two elements: a pure retarder
[U ] followed by a pure diattenuator [H ] (or vice versa) (Lu, 1996). In matrix
z2 zN
E in E out
[M1] [M2] [M3] [MN]
where the Hermitian diattenuator matrices are uniquely defined from [M2 ], as
shown in equation (1.81):
If [M2 ] is singular then [U ] is not uniquely defined, but this case can be accom-
modated by employing the singular value decomposition, as shown below. In
the most general case we can write [M2 ] in terms of a singular value decom-
position (see Appendix 1) as a product of unitary matrices [V ] and [W ] and a
diagonal matrix of singular values [D], as shown in equation (1.83):
[HR ] = [W ] · [D] · [W ]∗T
[M2 ] = [V ] · [D] · [W ]∗T ⇒ [HL ] = [V ] · [D] · [V ]∗T (1.83)
[U ] = [V ] · [W ]∗T
which is the same as the general lossy matrix shown in equation (1.74) with
the substitution θ = φ + χ. However, this does not exhaust the possibilities in
terms of possible types of propagation channel, as we now show.
If [M2 ] has non-orthogonal eigenvectors, then [V ] and [W ] are no longer
equal, and [HR ] and [HL ] are therefore distinct diattenuators with the same
1.2 The propagation of polarised waves 29
The eigenvectors of [M2 ] are not orthogonal and are easily calculated from the
eigenvalues, as shown in equation 1.86:
sin θ
λ1 = 0 ⇒ e 1 =
cos θ
⇒ e∗T
1 e2 = 0 (1.86)
1
λ2 = cos θ ⇒ e2 =
0
These are propagation states that remain unchanged after transmission through
the channel. However, as we shall now show using the SVD, they do not have
the optimum transmittance. The singular value decomposition is easily obtained
by inspection, as follows:
∗T 1 0 1 0 cos θ − sin θ
[M2 ] = [V ] . [D] . [W ] = · · (1.87)
0 1 0 0 sin θ cos θ
from which it follows that the polar decomposition yields an equivalent retarder
and pair of diattenuators of the form shown in equation (1.88):
cos θ − sin θ 1 0 cos2 θ − sin θ cos θ
[U ] = [HL ] = [HR ] =
sin θ cos θ 0 0 − sin θ cos θ sin2 θ
(1.88)
The extrema of this function define the maximum and minimum transmittance of
the channel. This is a classical Rayleigh quotient (see Appendix 1) for which the
maximum and minimum values are obtained for the eigenvalues of [M2 ]∗T [M2 ].
From the above we see that these are just the square of the singular values of
[M2 ]. We further see that we can decompose the matrix product, as shown in
equation (1.90):
In this way we see that the singular value decomposition is more relevant to
channel gain optimization studies than the eigenvalue expansion of [M2 ]. Note,
however, that the price to pay for maximization of transmittance is distor-
tion of polarisation state. For homogeneous systems there is little difference
between expansions, but for inhomogeneous propagation channels the SVD
or eigenvalue decompositions must be selected only after consideration of the
application in mind.
∆z
∆z
z axis
r
(x,y)
x2 + y2
f0 f1 r= x2 + y2 + ∆z 2 ≈ ∆ z + Fig. 1.10 Derivation of the propagation
2∆z matrix in scattering media
At present there is some ambiguity with the coordinates defined for this equa-
tion. So far we have dealt only with propagation in the z-direction, but clearly
the scattering particle generates a spherical wave, which propagates in three
dimensions. We delay discussion of the definition of coordinates for the scat-
tering matrix until Section 1.4.2. At the moment we are interested only in the
effect of the particle in the same (z) direction as that of the incident wave. This
is termed forward scattering, and later we shall see it as a special case of the
general scattering problem.
The matrix [S] contains four complex numbers: two diagonal copolar scat-
tering coefficients which change the amplitude and phase of the wave but
maintain its polarisation in the same state as that of the incident plane wave.
The off-diagonal terms represent the possibility for crosspolarisation. Here the
amplitude and phase of the wave is modified by the particle, which now also
scatters radiation into the orthogonal polarisation to that incident. We can use
this scattering matrix to derive a Fresnel approximation for the change in elec-
tric field with z, and so determine the form of the [N ] matrix as follows (van
de Hulst, 1981). Figure 1.10 shows a component of the electric field at two
positions: f0 is the field at z while f1 is the new field at position z + z. We
are interested in establishing the change in field f1 − f0 for small z (but still
large enough so that βz 1, otherwise more rigorous diffraction theory is
required). The starting point is to decompose both fields into incident (fi ) and
scattered (fs ) field components so that we can write
f0 = fi
fi = e−iβz (1.93)
f1 = f1i + fs
depending on our choice of polarisation channel. We now again face the problem
that [S] relates a plane wave to a spherical wave, and so formally we must
consider variations in x, y and z. However, we are concentrating on forward
scattering only, and so any consideration of x,y variations must be constrained
to a paraxial or small angle approximation so that only the forward scattering
amplitude is appropriate. To secure this, we replace r in equation (1.95) with its
approximation shown on the right-hand side of Figure 1.10. We also consider
a large number of particles with density N0 per unit volume. In this case the
summation becomes an integral, and we can rewrite equation (1.95) in the form
shown in equation (1.96), where we have included analytical evaluation of the
resulting Fresnel integral with infinite limits:
2 2)
+y
−iβz −1 − iβ(x2z
fs = e sxy fi N0 z e dxdy dz
∞ 1 (1.96)
βx2
−i 2z 2π z 2
−iβz 2π
e dx = ⇒ fs = e sxy fi N0 −i dz
iβ β
−∞
2π No
fs ≈ −i fi zsxy (1.97)
β
In practice, the particles contributing to this integral will not have the same
orientation and size, and we therefore need one further modification to account
for variation over these parameters. Note that changing the size or orientation
does not alter the Fresnel integral, but requires only that we now replace the
constant factor sxy by a configurational average over distributions of size and
shape. So, for example, for ellipsoidal particles with dimensions a, b and c and
major-axis orientation defined by three Euler angles A, B and C (Goldstein,
1980) we can write:
∞ ∞ ∞ 2π π 2π
where p(…) is the probability density function (PDF) of the distribution; and
so finally we have the following expression for the scattered field:
2π No
fs ≈ −i S fi z (1.99)
β
This is the desired result which allows us to write the total change of field for
each of the four scattering channels separately as shown in equation (1.100),
where we note that the crosspolarised channels contain, by definition, no
1.2 The propagation of polarised waves 33
This provides a formal connection with the Jones calculus developed in the
previous section. Again we have two eigenstates for any such material, and
these may or may not be orthogonal, depending now on the distribution of
particle size, shape and orientation. In general, the medium will act as a mixed
retarder/diattenuator, which changes the polarisation of the incident wave as
it progresses into the material. As an example, consider propagation through
a cloud of dipoles. We assume all particles have the same size and shape and
that all are located in the xy plane, and hence their orientation distribution is
controlled by a single parameter function p(θ ). The forward scattering matrix
for each particle is then a function of just two parameters of the form shown in
equation (1.102):
cos θ sin θ ε 0 cos θ − sin θ cos2 θ − sin θ cos θ
S= · . =ε
− sin θ cos θ 0 0 sin θ cos θ − sin θ cos θ sin2 θ
(1.102)
In the special case that the distribution is azimuthally symmetric in the xy plane
we have
1 ε 1 0
p (θ) = ⇒ S = (1.104)
2π 2 0 1
34 Polarised electromagnetic waves
which leads to isotropic propagation with zero diattenuation D and the same
extinction rate for all polarisations depending on the particle density, size and
dielectric constant. This simple example will be very important when we con-
sider coherent volume scattering in Chapter 7, and demonstrates how it is not
only the particle shape but also the orientation distribution that determines the
propagation channel in random media problems.
We have seen that wave polarisation can be transformed by propagation
through a channel. It is therefore of interest to consider the complete set of
possible wave polarisation states with a view to investigating what fraction of
all states are generated by a particular transforming channel or system. Such a
complete set is geometrically represented by the surface of a sphere called the
Poincaré sphere, as we now consider.
where ax and ay are the amplitudes of the components, and φ is the relative phase
shift. We can eliminate time from these equations using standard trigonometric
1.3 The geometry of polarised waves 35
y
θ π π
− ≤θ <
2 2
τ π π
x − ≤τ ≤
4 4
ωt = π/2
q
b a
Change in time ωt = φ
origin
Fig. 1.11 Geometry of the polarisation
p ωt = 0 ellipse
where a and b are vectors in the x–y plane lying along the major and minor axes
of the polarisation ellipse. For a general time origin we then have the following
relation:
/
iωt iφ p = a cos φ − b sin φ
E = a + ib e e = p + iq e ⇒ iωt
⇒ p.q = 0
q = a sin φ + b cos φ
(1.109)
where the vectors p and q again lie in the xy plane, but are not orthogonal and
correspond to what are called the conjugate semidiameters of the polarisation
ellipse. The relationship between a, b and p, q is summarized in Figure 1.11.
1
φ= arg(p2 − q2 + 2ip.q) + nπ (1.110)
2
The importance of this result is that it relates absolute phase to spatial vectors
p and q. For example, points of circular polarisation are then defined from
singularities of φ, for which we have the following geometrical conditions,
which corresponds in general to a line given by the intersection of two surfaces
in space:
p.q = p2 − q2 = 0 (1.111)
These circular polarisation lines are termed C-lines, and are important in a full
characterization of fields based on their singularities as used, for example, in
the study of catastrophe optics (see Nye (1999) for more details).
0 1
E r, t = P(r) + iQ r eiωt (1.112)
where P and Q are now three-element vectors. From this, the wave polarisation
is still an ellipse but now lying in the plane defined by P and Q with a normal
to the plane n = P × Q. Special cases arise when P × Q = 0, in which
case the polarisation is linear and n is not defined. Again this will generally
occur along special lines in space called L-lines. Circular polarisation occurs
for P·Q = P2 −Q2 = 0, and so there exist C-lines also in the three-dimensional
case, although these are in general not the same as those for the two-dimensional
case. Note that the direction of wave propagation is not so clearly defined in
three dimensions, and some care is required in the definition of sense of elliptical
polarisation and absolute phase. Although resort can always be made for general
fields to the Poynting vector S = E × H, which is the direction of energy flow,
this involves a combination of electric and magnetic field components which is
undesirable for polarisation studies. To resolve this we again define the absolute
phase from the definition of major and minor axes, as shown in equation (1.113),
where A · B = 0:
P + iQ = eiφ A + iB (1.113)
The problem is now to relate the phase of an elliptical wave at one point
to the phase at a neighbouring point where the ellipse has a different size,
shape and orientation vector n. One elegant way to resolve this is to use the
Pancharatnam geometric phase (Pancharatnam, 1956), defined from the inter-
ferogram between the two complex signals. In this way the phase difference
between points is defined from maxima and minima in a (real) intensity pattern.
Mathematically it is derived from the Hermitian product, as shown in equation
(1.114):
ψ = arg(E ∗1 E 2 ) (1.114)
where the last term results since A·B = 0 implies that d (A·B) = dA·B+d B·A =
0. This then confirms that the change in phase can be decomposed into the
gradient of the phase plus a contribution from the change in geometry. Note that
there still remains a problem with this definition, as ψ is not formally integrable,
causing problems4when we try to define terms such as δψ/δr, although for the
second term, ∂φ ∂r = ∇φ is acceptable. Hence there is no suitable unique
indicator of propagation direction over the whole field. However, we can define
38 Polarised electromagnetic waves
2AB ∂ε ∂φ 2AB
t=− + =− 2 t + ∇φ (1.116)
A2+ B ∂r
2 ∂r A + B2 ε
where A and B are the magnitudes of the vectors, and the angle ε is the angle
of rotation of the axes A, B about n, defined as B · dA = ABd ε. The component
of t parallel to n represents the twist of the ellipse about n, while the transverse
component represents the bend of the axial directions of the ellipse in its own
plane. This expression is well behaved at most points in the field, and in partic-
ular it can be made to yield sensible definitions even along C-lines, when grad
φ is singular and lines of linear polarisation when tε is not defined (see Nye
(1999) for more details). This means that we can take t as a general definition
of the propagation direction and then use it to define handedness at each point
in the wave field as the sign of the triple product [A B t].
Shortly we shall see a simple geometrical interpretation of this result, but for
the moment consider a change of ellipticity of the wave from τ to τ + τ . This
can be represented by a matrix equation as shown in equation (1.120):
cos (τ + τ ) cos τ i sin τ cos τ
= . (1.120)
i sin (τ + τ ) i sin τ cos τ i sin τ
Combining this with the rotation matrix involving θ we can extend the analysis
by representing not only unitary vectors in C2, but also special unitary matrix
transformations, so that we can transform between any pair of ellipses by first
rotating the major axis and then changing the ellipticity. The general compound
unitary transformation can then be written as shown in equation (1.121):
e−iφ 0 cos τ −i sin τ cos θ − sin θ
[U2 ] = . . (1.121)
0 eiφ −i sin τ cos τ sin θ cos θ
This should be compared with the alternative forms shown in equations (1.28)
and (1.46). Note that in this case we have obtained a product decomposition
with a strong geometrical interpretation. We can also rewrite equation (1.121)
as a product of matrix exponential functions as defined in equation (1.42). The
result is shown in equation (1.122):
a real three-vector r. In the context of polarimetry this is called the Stokes vec-
tor representation, after the nineteenth-century English mathematician George
Gabriel Stokes (1819–1903) (Born and Wolf, 1989; Bickel, 1985; Schmeider,
1969).
The first stage is to define a suitable real vector r. As there are three matri-
ces in the cascade of equation (1.122) we anticipate a three-dimensional real
space
R3 and so set r = (x, y, z). The next key idea is to recognize that
r = x2 + y2 + z 2 (amplitude) should be a transformation invariant, so that
we deal with orthogonal transformations in R3. As we are going to deal with
matrix products we need to generate a matrix from r with a metric that is a
function of the desired invariant. One way to do this is to construct a matrix
using the Pauli spin matrices, as shown in equation (1.123):
x y − iz
[R] = xσ1 + yσ2 + zσ3 = ⇒ det([R]) = −(x2 + y2 + z 2 )
y + iz −x
(1.123)
R ϕ, n = exp iϕσ .n . [R] . exp −iϕσ .n (1.124)
This equation preserves the determinant of [R] on both sides, as the matrix
exponentials both have unit determinant and the determinant of a matrix product
is just the product of determinants (which was the reason why we focused on
special unitary matrices in equation (1.26)). We have therefore constructed a
transformation of r with the correct metric.
We must now determine other invariants under this transformation. In par-
ticular we seek the axis of any plane rotations. This will allow us to finally
determine a mapping from the 2 × 2 complex matrix into a real 3 × 3 orthogo-
nal matrix. To do this we need only consider infinitesimal transformations, so
we can conveniently use only the first few terms of the series expansion of the
exponential function (see equation
5 (1.38)). Equation (1.125) shows this expan-
sion. We note that since [R] = 3j=1 rj σj , invariants can then be conveniently
identified from products of the Pauli matrices that commute; that is, for which
σi σj − σj σi = 0.
[R (ϕ)] = σ0 + iϕσ · n + · · · · [R] · σ0 − iϕσ · n + · · ·
≈ [R] + iϕ [R] σ · n − σ · n [R] (1.125)
of scalar and vector
products of the original real vectors as a · σ b · σ =
a · b + i a × b σ . Hence it follows from this and equation (1.125) that the only
other invariant is given by equation (1.126):
We can formalize this correspondence by defining the groups SU(2) and O+ 3 as,
respectively, the group of all 2 × 2 complex unitary and 3 × 3 real orthogonal
matrices with determinant +1 (see Appendix 2). If two groups X and Y have a
1-to-1 correspondence between elements they are said to be isomorphic, while
if the correspondence is many-to-one they are homomorphic. Hence we have
demonstrated that SU(2) and O+ +
3 are homomorphic. This SU(2)–O3 homo-
morphism is fundamental to the development of polarisation geometry. One of
42 Polarised electromagnetic waves
z
P
2α P 2φ
2τ
δ 2τ
2θ y
2θ
Here we see that by adding the amplitude as a multiple of the identity matrix
we can secure a new invariant: namely, the difference between the amplitude
squared and the square of the length of the vector r. For a polarised wave this
metric will always be zero. This modifies the geometry to that of a Lorentz
1.3 The geometry of polarised waves 43
transformation, where equation (1.128) arises naturally through the use of the
Minkowski metric. The Lorentz spin transformation is then obtained as a gen-
eralization of equation (1.124) as shown in equation (1.129), where [L] is called
a Lorentz spin matrix, det([L]) = 1, but γ can be complex, and [L] does not
have to be unitary.
R γ , n = [L γ , n ] · [R] · [L γ , n ]∗T [L] = exp iγ σ · n γ = γr + iγi
(1.129)
Figure 1.13 shows some examples of the Stokes vectors for linearly and cir-
cularly polarised waves. Equation (1.130) defines the Stokes vector in terms
of the geometrical parameters of the ellipse. An alternative definition is based
on the wave coherency matrix [J ]. This is a 2 × 2 Hermitian matrix, and is
defined from the product of the complex electric field vector with its conjugate
transpose. Using α and δ parameters this matrix can be defined as shown in
equation (1.131):
cos αw
E=a
sin αw eiδ
∗T cos2 αw sin αw cos αw e−iδ
⇒ [J ] = EE =a 2
= [R]
sin αw cos αw eiδ sin2 αw
(1.131)
To make this consistent with equation (1.128) the matrix [J ] must be related
to the Stokes vector g as shown in equation (1.132), where I , Q, U and V are
44 Polarised electromagnetic waves
Horizontal Polarisation
g = (1,1,0,0)
Vertical Polarisation
g = (1,–1,0,0)
symbols widely used in the optics literature for the Stokes parameters, elements
of the Stokes vector.
Ex Ex∗ + Ey Ey∗ I
Ex Ex∗ Ex Ey∗ 1 g0 + g1 g2 − ig3 E E∗ − E E∗ Q
x x y y
[J ] = = ⇒g= =
Ey Ex∗ Ey Ey∗ 2 g2 + ig3 g0 − g1 Ex Ey∗ + Ey Ex∗ U
∗ ∗
i(Ex Ey − Ey Ex ) V
(1.132)
• Det([J ]) = 0. This follows from the Minkowski metric, and is true for
all polarised monochromatic waves.
• [J ] = [J ]∗T ; that is, [J ] is complex Hermitian, and therefore has orthog-
onal eigenvectors and real eigenvalues (see Appendix 1). For polarised
waves it follows that the eigenvector decomposition of [J ] can be written
as shown in equation (1.133):
cos αeiφ − sin αe−iδ λ 0 cos αe−iφ sin αe−iδ
[J ] = . 1 .
sin αeiδ cos αe−iφ 0 0 − sin αeiδ cos αeiφ
(1.133)
The matrices [G] and [G]−1 may then be used to transform between the two
conventions. The use of matrices to represent changes in Stokes vectors forms
the basis of an important calculus. In equation (1.135) we show two other impor-
tant examples of matrices used in Stokes algebra. The first [M ]conj represents
the physical process of conjugation, or taking the complex conjugate of the
wave. As shown, this has the effect of changing the sign of the fourth Stokes
parameter or changing the sense of the polarisation ellipse. On the right we
show the transformation matrix that converts a Stokes vector into its orthogo-
nal polarisation state. This involves reflection of the point in the origin to obtain
the coordinates of the antipodal point on the Poincaré sphere.
1 0 0 0 1 0 0 0
0 1 0 0 0 −1 0 0
[M ]conj = 0 0 1 0
[M ]⊥ = 0 0 −1 0 (1.135)
0 0 0 −1 0 0 0 −1
Finally, we note that an extension of this Stokes approach to cover the case of
three-dimensional waves, when the wave coherency matrix [J ] becomes 3 × 3,
has been developed by Roman (1959a) as a set of nine Hermitian matrices anal-
ogous to the Pauli spin matrices, the coefficients of which define a generalized
set of Stokes parameters ri for the representation of waves with arbitrary form
(Roman 1959b). In this case the generalization of equation (1.132) is shown in
equation (1.136):
Ex Ex∗ Ex Ey∗ Ex Ez∗
[J3 ] = Ey Ex∗ Ey Ey∗ Ey Ez∗
Ez Ex∗ Ez Ey∗ Ez Ez∗
r0 + r1 + r4 r2 + ir3 r7 + ir8
= r2 − ir3 r0 r2 − r5 + i(r3 − r6 ) (1.136)
r7 − ir8 r2 − r5 − i(r3 − r6 ) r0 − r1 + r4
Here we concentrate on the simplified 2×2 case, which is the most commonly
used in remote sensing. However, special applications that require analysis
of dynamic vector near field effects should make use of this more general
formalism.
We have now considered the generation and geometry of polarised waves and
the distortion effects caused by propagation channels. Now we turn to consider
the final stage in Figure 1.1: namely, the scattering of polarised waves and the
interpretation of a scatterer as a transformer of polarisation state.
46 Polarised electromagnetic waves
Scatterer
S
x'
x bi
bs
Α y' Β
y
Incident wave bisectrix b scattered wave
Fig. 1.14 General wave scattering coordi-
nates for the amplitude scattering matrix
1.4 The scattering of polarised waves 47
Note that sometimes in the literature (van de Hulst, 1981; Hovenier, 2004) a
normalizing factor of 1/iβ is used on the right-hand side in equation (1.137). This
is supported by the discussion leading up to equation (1.99), but readers should
be warned that there exist slight differences of definition for the scattering
amplitude matrix.
Cleary, knowledge of this mapping (that is, knowledge of all four complex
numbers in [S]) enables full polarimetric characterization of the scattering prob-
lem from A to B. We can then determine the amplitude and polarisation of the
scattered wave at B for any incident polarisation state Ei on the Poincaré sphere
by combining the C2 freedom with matrix multiplication arising from system
linearity. This is called polarisation synthesis or virtual polarisation adaptation
(Giuli, 1986; Poelman, 1991). The key idea is that measurement of [S] can
lead to adaptive selection of polarisation to enhance features based on signal
processing rather than on physical antenna changes. The matrix [S] therefore
plays an important role in polarimetry theory (Luneburg, 1996, 1997), and we
now turn to discuss its properties in more detail.
E i = [U2A ] .E i
(1.140)
E s = [U2B ] .E s
With this flexibility, the scattering matrix in the new base pair defined by U 2A
and U 2B is related to that in xy and x y by a unitary similarity transformation
as shown in equation (1.141):
L = E ∗T ∗T ∗T
2 [S] E 1 + λ1 (E 1 E 1 − 1) + λ2 (E 2 E 2 − 1) (1.143)
Here we see in the first term the received complex voltage when we transmit a
state E 1 and receive with an antenna with polarisation E 2 . In addition we have
two constraint equations (and two corresponding unknown Lagrange multipli-
ers) to ensure that the antenna polarisation vectors E are unitary. To solve for the
maximum received power LL∗ we can sidestep the need to expand the product
directly and instead set the partial derivatives of L and L∗T separately to zero,
as shown in equation (1.144). (We shall use this same strategy when solving
coherence optimization in polarimetric interferometry in Section 6.2.)
∂L
= [S] E + λ E = 0
∂E ∗T
1 2 2
2 [S]∗T [S] E 1 = λ∗1 λ2 E 1
⇒ (1.144)
∂L∗T
[S] [S]∗T E 2 = λ∗1 λ2 E 2
∗T ∗
= [S] E 2 + λ1 E 1 = 0
∂E ∗T
1
symmetry operation of the system. We have already seen one such example:
the problem of changing the incident wave polarisation. We have seen that we
can easily predict the polarisation of the scattered wave using C2 symmetry
coupled to matrix multiplication. However, this does not exhaust all possibil-
ities. There are several such symmetry transformations (van der Hulst, 1982;
Cloude, 1995b). As a trivial example we can translate the transmitter A or
receiver B along the vectors β i and β s and effect only a scalar phase change in
[S] → eiϕ [S], where φ is related to the change of range between A and B. This
forms the basis for a discussion of interferometry with polarised waves, to be
discussed at length in Chapter 6. Other interesting transformations are offered
by considering rotations and reflections of the original problem in the bisectrix
and scattering planes as we now consider.
We start by considering the matrix S (dispensing with square-bracket nota-
tion for the moment, for convenience) to be the ‘mother’ configuration, with
complex elements a, b, c, and d . There are then three important ‘daughter’
matrices we can write for completely different scattering problems, as shown
in equation (1.145):
a −c
Sα =
−b d
a b a −b
S= ⇒ Sβ = a, b, c, d ∈ C (1.145)
c d
−c d
Sγ =
a c
b d
The first of these, S α , follows from the vector reciprocity theorem for electro-
magnetic waves. This theorem relates the exchange of transmitter and receiver
positions, and states that if we transmit a polarisation state PA from A, then the
component polarised in the PB direction at B is equal to the PA component of
the scattered radiation when we illuminate the same object from B with polar-
isation PB . Figure 1.15 shows the reciprocal problem to that in Figure 1.14.
Scatterer
x'
x y'
y –bs
–bi
Scattered Β Incident
Α Bisectrix b
wave wave
Fig. 1.15 Definition of the reciprocal scatter-
ing problem
50 Polarised electromagnetic waves
Note how the incident and scattered wave vectors have changed direction. The
scattering matrices for the two cases are then related by the reciprocity theorem
as shown in equation (1.146):
0 1
E AS = S β i β s · P A
0 1 P B · E As = P A · E Bs
E S = S −β i − β s · P B Reciprocity Theorem
B
0 1 0 1T
⇒ S βi βs = S −β i − βs (1.146)
Note that the y and y’coordinates have been reversed in direction for the problem
in Figure 1.15 compared to the master problem in Figure 1.14. To account for
this we need to invert the y axis so that equation (1.146) becomes
1 0 T 1 0
S βi βs = S −β i −β s
0 −1 0 −1
a b a −c
⇒S= ⇒ Sα = (1.147)
c d −b d
This leads directly to S α in equation (1.145). The other two daughter problems
are more easily derived: S β follows from an inversion of one coordinate, while
S γ follows from successive application of α and β.
The reciprocity relation is especially important for backscatter problems,
when the transmitter and receiver are at the same position. In this case it follows
that S = S α , and from equation (1.147) this is only possible if the original
‘mother’ scattering matrix has the following form:
a b
[S]backscatter = (1.148)
−b d
This is called the backscatter theorem, and demonstrates that in this special
case the S matrix has only three independent elements. Note that while there
exists a class of non-reciprocal backscatter problems, these are not common
in the remote sensing of natural land and sea surfaces (an important exception
being propagation through the ionosphere when Faraday rotation occurs; see
Section 1.2.2). The reciprocity theorem is therefore an important basic symme-
try, widely employed in radar backscatter systems for information extraction
and calibration.
It is of interest to now reconsider an SVD of the backscatter matrix of equation
(1.148) and to relate the problem to that of matched antenna illumination. To do
this we must first consider a correction to account for the differences between
transmitter and receiver antenna coordinates—a generalization of that following
from equation (1.147).
Voc = h T*.E i
Fig. 1.16 Complex effective length of an
h = hθ θ + hϕ ϕ complex effective length of the antenna antenna
1
1 h1 h2 h2 = [1 1]T
[1 1]
T 2
h1 =
2
Antenna 2
Antenna 1
Receiver
Transmitter
1 0 1 0
Ei = h ⇒ Voc = h*T h1 = 0
*
0 −1 1 2 .E = h 2
0 −1
1
[1 i]
T
1 T h2 =
h1 = [1 i] h1
h2 2
2
Antenna 1 Antenna 2
Transmitter Receiver
∗ 1 0 a b
Voc = h∗T . . .h1 = hT2 [S]sensor h1 (1.149)
2 0 −1 c d
Voc = hT [S]sensor .h
1 0 1 0 a b a b
⇒ [S]BSA = [S]FSA = . =
0 −1 0 −1 c d −c −d
(1.150)
Also shown in equation (1.150) is the relationship between the scattering ampli-
tude matrix in the sensor coordinates (which becomes the backscatter alignment
system or BSA system in backscatter) and the wave or FSA coordinates. Note
the phase change of elements of [S]. Care is therefore required in always speci-
fying which coordinate system is being employed in a description of scattering.
In the engineering literature the sensor or BSA system is almost universally
used, while in optics and physics the conventional FSA or wave system is
preferred.
In equation (1.148) we saw that from the backscatter theorem, b = −c in
backscatter. It follows from equation (1.150) that the BSA matrix becomes
complex symmetric; that is, b = c, and the received backscatter voltage
for reciprocal problems can be written in the most general case, as shown
in equation (1.151):
T a b
Voc = h2 [S]BSA h1 = h2
T
h (1.151)
b −d 1
This is the most common form of this equation used in radar backscattering
theory. Some examples will be used to illustrate its application.
match and mismatch: namely, that circular polarisation yields zero received
voltage, while ±45-degree linear is matched for maximum signal strength (as
incidentally are all linear incident polarisations).
1
h = ⇒ Voc = 0
√1
1 0 ±i 2
[S]BSA =A ⇒ (1.154)
0 1
1
h = √1 ⇒ Voc = A
2 ±1
B
1.4.2.2 Example 2: dihedral scattering matrix
Es = [R] E t
A second important way in which specular backscattering can occur is in dihe-
dral retroreflection, as shown schematically in Figure 1.19. In this case we have
a 90-degree corner formed from two perfect conductors (PC), and the ray paths
indicate a strong specular return to the transmitter through two 45-degree reflec-
A
tions at each surface. From a polarisation point of view the reflection matrix is
formed from the product of the reflection matrices at A and B. In general, the
Fig. 1.19 Dihedral retro-reflection Fresnel coefficients change with angle of incidence onto the surface (see Section
3.1.1), and away from normal incidence are not given by the simple form of
[Ref ] in equation (1.152). We shall consider the more general case later, but
here concentrate on perfect conductors only. For this special case, [Ref ] applies
for all angles, and using a cascade of matrices of the form of equation (1.152)
we can generate the normalized reflection matrix shown in equation (1.155):
1 0 1 0 1 0
Ref = Ref Ref = . = (1.155)
A B 0 −1 0 −1 0 1
The voltage received for an antenna with effective length h in the BSA system
is then of the following form:
1 0 1 0 1 0
Voc = hT .E i = hT . · E t = hT · ·h (1.156)
0 −1 0 1 0 −1
and so the scattering matrix for a dihedral has the form shown in equation
(1.157):
1
h = ⇒ Voc = A
√1
1 0 ±i 2
[S]BSA = A ⇒ (1.157)
0 −1
1
h = √1 ⇒ Voc = 0
2 ±1
We now see that a circularly polarised antenna is matched to the return signal,
while a 45-degree linearly polarised antenna is mismatched, rather like the
communication system of Figure 1.17. Clearly the pattern is now set for higher-
order scattering processes. The BSA scattering matrix for order N scattering
from a set of PC surfaces is denoted as follows:
2N +1
1 0
[S]BSA = (1.158)
0 −1
1.4 The scattering of polarised waves 55
so for N odd (1,3,5 and so on) there is zero phase difference between copolar
elements, while for N even (2,4,6 and so on) there is π phase difference. The
generalization of these ideas to dielectric surfaces is dealt with in Section 3.1,
and leads us eventually to use of the scattering alpha parameter to accommodate
the changes in boundary conditions for general surfaces.
As an important example of equation (1.158), consider the case N = 3, which
occurs for a trihedral corner reflector, widely used for radar backscatter cali-
bration, as it combines high radar cross-section with a broad radiation pattern,
so simplifying alignment with the sensor. (Contrast this with a flat plate N = 1,
which can always be made bright by increasing its size but only at expense of the
radiation pattern, which becomes more localized around the specular direction
in a narrow pencil beam.) Figure 1.20 shows an image of a metallic trihedral N
= 3 reflector. From a polarisation point of view, this scatterer behaves the same
as specular normal surface reflection (N = 1), and has a BSA scattering matrix
equal to the 2 × 2 identity matrix. Hence it too is ‘blind’ to circularly polarised
antennas, regardless of its physical size.
On the other hand, for complex symmetric matrices the two unitary components
[U ] and [V ] of the SVD are related, since we have the following identity:
and so we see from the SVD that a symmetric [S] matrix can always be diagonal-
ized by a congruent unitary transformation, just as we found for the scattering
matrix in the sensor coordinates (equation (1.159)). Some examples will help
illustrate how this formalism can be used in practice. Firstly we consider the
simple case of rotation of the reference plane of polarisation for backscattering
by a horizontal dipole and a dihedral reflector. The scattering amplitude matri-
ces for these two cases can be derived from standard similarity transformations
involving a plane rotation matrix, as shown in equation (1.161):
cos θ − sin θ 1 0 cos θ sin θ cos2 θ cos θ sin θ
[Sθ ] = . . =
sin θ cos θ 0 0 − sin θ cos θ cos θ sin θ sin2 θ
cos θ − sin θ 1 0 cos θ sin θ cos2 θ − sin2 θ 2 cos θ sin θ
[Sθ ] = . . =
sin θ cos θ 0 −1 − sin θ cos θ 2 cos θ sin θ sin2 θ − cos2 θ
(1.161)
For example, consider again the rotated dihedral matrix now expressed in the
circular base (substitute the lower example in equation (1.161) into (1.162)).
The result is shown in equation (1.163). Now we see a much-simplified diagonal
form with rotation influencing only the phase between LL and RR polarisations.
This is an example of diagonalization via the SVD.
cos2 θ − sin2 θ + i2 sin θ cos θ 0
[S]circ =
0 sin2 θ − cos2 θ + i2 sin θ cos θ
i2θ
cos 2θ + i sin 2θ 0 e 0
= =
0 − cos 2θ + i sin 2θ 0 −e−i2θ
(1.163)
1.4 The scattering of polarised waves 57
P
XPOL null
COPOL null
P⊥ Fig. 1.21 Schematic representation of the
Scatterer S XPOL and COPOL nulls
The main conclusion is that for any given symmetric complex scattering matrix
we can always find an orthogonal base ‘ab’ such that [S] is diagonal in that
base; that is, there is zero crosspolarisation. Incidentally, this is reminiscent
of eigenpolarisations in wave propagation (Section 1.2.2), except here we
are dealing with the physically distinct case of wave scattering. These states
are sometimes confusingly called eigenstates, but should more accurately be
called crosspolar nulls, or often abbreviated to XPOL nulls of the scattering
matrix, as originated by Kennaugh (1952) and further developed by Huynen
(1987) and Boerner (1981). Note that for symmetric matrices there are always
two orthogonal XPOL nulls. They have two very interesting properties, as
follows:
• The XPOL nulls maximize the backscattered signal (for a fixed antenna
polarisation), and thus maximize the detectability of the object in the
presence of noise.
• The XPOL nulls remain unchanged on scattering from the object, and
thus ‘carry’ information about any symmetry properties of the object.
Figure 1.21 shows a schematic representation of this invariance property of
the XPOL nulls. By definition, when P is an XPOL null then the scattered field
is also P polarised (in the BSA coordinates) as shown. For example, when the
scatterer has an axis of symmetry in the plane of polarisation then the XPOL
nulls are linear polarisations aligned and perpendicular to this symmetry axis. In
this way we can establish the orientation of an unknown symmetry axis by using
an SVD of the measured scattering matrix. But how does this diagonalization
process link with the idea of blind polarisations? To see this, we now seek inci-
dent polarisation states that are not invariant but rather transformed into their
orthogonal state on scattering. These states would then by definition constitute
the blind polarisations of the system. In distinction to the diagonalization pro-
cess, these states are called copolar nulls or simply COPOL nulls, and we shall
now show that there are always two such states for all [S] matrices, although
unlike XPOL nulls they are only orthogonal under special circumstances. The
trick is to solve for COPOL nulls starting from the diagonal form of the [S]
matrix, as follows. Recall that the XPOL nulls are always orthogonal (forming
a base we shall call ‘ab’) and so we can always write the scattering matrix in
this base in diagonal form, as shown in equation (1.164):
γ 0
[S] = a (1.164)
0 γb
where γa and γb are the complex singular values of [S]. We can represent
any polarisation state as a linear combination of these XPOL nulls, that is,
E = Ea a + Eb b, and hence for COPOL nulls we seek solutions of the form
58 Polarised electromagnetic waves
The second stage involves scattering of the wave by the target. This can be
represented by a 2 × 2 complex scattering matrix [S], which we express in the
sensor or BSA coordinate system as follows:
The third stage involves propagation back through the medium from B to A.
As we are remaining in the A sensor coordinate system then the reciprocity
theorem implies that [Mz (A,B)] = [Mz (B,A)]T (equation (1.146)), so that we
can write the following:
Stage III : E As = [Mz (B, A)] .E Bs = [Mz (A, B)]T . [SBSA ] . [Mz (A, B)] E A
This is our desired result. It shows that the observed scattering matrix for
a combined reciprocal propagation and scattering channel is, in the sensor
BSA coordinate system, again obtained as a congruent transformation of the
target scattering matrix by the propagation channel matrix. This represents a
generalization of the change of basis, which resulted in a unitary congruent
transformation.
Having now obtained these basic algebraic results for the complex amplitude
matrix, its form under change of base and its fusion with propagation, we now
turn to consider a geometrical interpretation of these transformation properties.
60 Polarised electromagnetic waves
Left Right
handed handed
θ
2θ
Cos2τ
π π
− ≤ 2τ ≤
2 2
extended linear regions. This distorts the information for points located away
from the equator. One advantage, however, is the clear separation of left and
right sense polarisations into upper and lower halves of the diagram. This per-
mits easy visualization of asymmetry due to calibration errors or scattering from
asymmetric objects such as helices.
To overcome the distortion limitation, use is often made of a polar projection
of the sphere, as shown in the centre of Figure 1.23. Here the pole is located at the
centre of the image, and the polar θ coordinate is represented in a more natural
way, without the need for any artificial discontinuities. However, one remaining
problem is that the double angle representation of the Poincaré sphere leads to
upper and lower hemispheres being mapped on top of each other. Hence there is
no discrimination in this diagram between left and right sense polarisations. One
compromise solution is to employ the representation shown on the right-hand
side of Figure 1.23. Here we maintain the polar representation with circular
polarisation at the centre, but now map θ in the range −π/2 to π/2. This frees
the second half of the polar plot for the opposite sense of polarisation. In this
way we can map both senses and all linear states onto a simple polar diagram
(Woodhouse, 2003).
Figure 1.24 shows examples of all three representations for a set of sam-
ple matrices. The first three are the backscattering matrices corresponding to
–80
–60
–40
2*Tau (degrees)
1 0 –20
[S] = 0
0 1 20
40
60
80
–150 –100 –50 0 50 100 150
2*Theta (degrees)
–80
–60
–40
1 0
2*Tau (degrees)
–20
[S] = 0
0 –1 20
40
60
80
–150 –100 –50 0 50 100 150
2*Theta (degrees)
–80
–60
–40
0 1
2*Tau (degrees)
–20
[S] = 0
1 0 20
40
60
80
–150 –100 –50 0 50 100 150
2*Theta (degrees)
–80
–60
–40
2*Tau (degrees)
–20
0
1 i
[S] = 20
40
i –1 60
80
–150 –100 –50 0 50 100 150
2*Theta (degrees)
–80
–60
–40
2*Tau (degrees)
–20
0
1 –1
[S] = 20
40
–1 1 60
80
–150 –100 –50 0 50
2*Theta (degrees)
100 150
Fig. 1.24 Copolarised backscatter power
signatures for canonical scattering matrices
62 Polarised electromagnetic waves
the Pauli set. The first of these is just the identity matrix, and we can see the
characteristic copolar nulls at left and right circular polarisations. Maximum
response is obtained for all linear polarisations, and there is no left/right depen-
dency of scattered power. This represents the most symmetric of scattering
systems, showing dependency on neither orientation nor sense of polarisation.
The second and third Pauli matrices demonstrate rotation dependence. Here the
copolar nulls are located at 45-degree and 0-degree linear polarisations respec-
tively. Maximum backscatter is now achieved for a set of elliptical polarisations
lying along a great circle of the sphere, passing through the poles correspond-
ing to left and right circular polarisation. Again, however, we note that there
is no left/right dependence on backscatter. These clearly show how the third
Pauli matrix can be obtained from the second by coordinate rotation through 45
degrees. The next scattering matrix shown is that for a helix. This is chosen to
illustrate an object that scatters one sense of circular polarisation in preference
to the other. The asymmetry in the copolarised plot is clearly seen in the first and
third diagrams. Finally we show the scattering matrix for a dipole oriented at 45
degrees to the horizontal. The orientation of the scatterer is clearly seen directly
in the third representation. These final two examples demonstrate the benefit of
the hybrid representation for visualising both rotation and sense preference in
scattering.
Of particular interest are the extreme values of these scattering functions.
These can be analytically evaluated with the help of a singular value decomposi-
tion of the [S] matrix. We showed in Section 1.4.3 that for an arbitrary scattering
matrix there exist two pairs of orthogonal states for which the scattered power is
maximized (XPOL nulls). We further showed that in backscatter these maxima
always correspond to copolar combinations and will therefore appear in the
polarisation charts of Figure 1.24. Also, there are two (non-orthogonal) zero
minimum points which for backscatter also correspond to zeros in the copolar
function. In backscatter, therefore, copolar and crosspolar plots are useful for
assessing the maximum variation of scattering with changes in polarisation.
However, in the general non-symmetric scattering matrix case, as arises in
bistatic scattering, such functions are not guaranteed to contain global maxima
and minima.
The angle between the singular vectors, χ , is called the skip angle. Note that
the matrix on the right-hand side has unit determinant. This will provide an
important link with the geometry of the Lorentz transformation in the next
section.
We note from this analysis that the four points P, Q, R, and S all lie in a
plane in the space of the Poincaré sphere. This structure is called the polari-
sation fork, and was first developed by Kennaugh (1952) and later studied by
several authors, particularly Huynen (1970, 1987). Consequently, the angu-
lar parameters of this plane and fork geometry are often termed the Huynen
parameters. Kennaugh also devised a simple set of geometrical rules for pre-
dicting the transformation of polarisation state by a scattering matrix. The rules
employ the point I shown as the intersection of the chord joining the COPOL
nulls and the diameter PQ in Figure 1.25. For a given scattering matrix, the
transformation of any incident polarisation can then be found by first invert-
ing the point through I onto the sphere. This new point is then rotated by π
about the diameter normal to PQ. The coordinates of the new point then cor-
respond to the polarisation of the scattered wave. Applying these rules to the
points P and Q themselves confirms that they remain unchanged on scatter-
ing, while R and S are both mapped into their antipodal orthogonal points as
expected.
Two special classes of scattering can be defined based on the α angle. When
the singular values are equal in amplitude, then αw = 45 degrees, the COPOL
nulls are orthogonal, and I is located at the origin. Then, due to the symme-
try, all polarisation states lying on a great circle formed by rotating the fork
about the RS axis are also XPOL nulls. This is the case for the three Pauli
matrices shown at the top of Figure 1.24, where the loci of maximum scat-
tered power are seen as bands of white across the diagrams. This arises as the
Pauli matrices have singular values that differ only in phase but are equal in
amplitude.
The second important class occurs when the scattering matrix is singular
(zero determinant) and consequently one of the singular values is zero. In this
case αw = 90 degrees, R,I,S collapse to the point Q, and all polarisation states
are transformed into the state P, regardless of their initial position on the sphere.
64 Polarised electromagnetic waves
Such a device is a polariser, in that it scatters the same polarisation state for all
input states. Important examples are shown in the lower two sections of Figure
1.24. The dipole and the helix both have singular scattering matrices, and in
both cases the fork collapses to a line.
Note that a similar construct can be used for the case of non-symmetric
scattering matrices. The complex singular values are still well defined, and
thus the fork angle is also uniquely defined. The axis PQ now corresponds
to the left singular vectors and is again a diameter of the sphere, as P and Q
remain orthogonal. Hence the fork geometry is still maintained. However, now
the scattered wave must be interpreted in a new P Q orthogonal base defined
by the right singular vectors. Hence we have a fork for transmit and a second
fork for reception, although the fork angle remains invariant.
amplitude information we need to extend the Pauli set by addition of the identity
matrix and obtain a modified spin matrix. To see how matrices of the form [Q]
change this spin matrix we form the triple matrix product shown in equation
(1.174):
1 g0 + g1 g2 − ig3
[R (χ, τ )] = [Q] · · [Q]∗T (1.174)
2 g2 + ig3 g0 − g1
If we set the skip angle χ to zero and determinant to unity in equation (1.174),
this transformation has the following simple canonical form:
g0new + g1new g2new − ig3new 1 eb 0 g0 + g1 g2 − ig3 eb 0
= . .
g2new + ig3new g0new − g1new 2 0 e−b g2 + ig3 g0 − g1 0 e−b
(1.175)
By direct expansion it follows that we can relate the Stokes vectors of the
incident and scattered field by a 4×4 matrix [M ], as shown in equation (1.176):
new
g0 m00 m01 m02 m03 g0
g new m10 m11 m12 m13 g1
1new =
g m20 m21 m22 m23 . g2
2
g3new m30 m31 m32 m33 g3
cosh(2b) sinh(2b) 0 0 g0
sinh(2b) cosh(2b) 0 0 g1
= . (1.176)
0 0 1 0 g2
0 0 0 1 g3
The left-hand side is then always itself Hermitian, and so the Pauli expansion
coefficients are real and hence all the elements of [M ] must be real. Finally, by
adding the unitary matrix component of the polar decomposition [U ] we obtain
a spin transformation of the form shown in equation (1.178):
1 g0 + g1 g2 − ig3
[R] = [U ] · [H ] · · [H ]∗T [U ]∗T
2 g2 + ig3 g0 − g1
1 0 1 0
⇒ [M ] = T · [M ] · T (1.178)
0 [O3 ] 0 [O3 ]T
However, we have already shown that the effect of a complex special unitary
transformation maps into a real 3 × 3 rotation matrix [O3 ] that acts on the
Cartesian components of the Stokes vector (equation (1.127)). Hence the uni-
tary matrix component transforms [M ], as shown in equation (1.178), where
66 Polarised electromagnetic waves
Q g0 P
g0 = –g1 new
g0 = g1
g0
d
new
g1
Fig. 1.26 Minkowski diagram for canonical
g1
transformation of the Stokes vector
1 − tan4 γ
tan δ = tanh 2b = (1.179)
1 + tan4 γ
1.6 The scattering vector formulation 67
Geometrically, therefore, the pure states P and Q remain unchanged (they are
the XPOL nulls of the scattering matrix and so by definition are invariant). This
corresponds directly to the invariance of the speed of light in special relativity.
However, the coordinates of other points do change, and we shall see in Chapter
2 that this has important physical interpretation in terms of depolarisation.
In summary we have seen that we can extend the idea of transformation of
polarisation basis by a unitary matrix to transformation of polarisation state by
scattering. In the unitary change of base we can write the general change of
base as a matrix exponential function (equation (1.180)), exposing as it does
the elementary geometry of the Poincaré sphere:
where σ are the set of three Pauli matrices, and n is a real three-vector defining
the axis for rotation through θ in R3—a real three-dimensional space of the
Stokes vector. The most general form of transformation of E that accounts
not only for change of base but also polarisation dependent scattering, can be
written in the form shown in equation (1.181):
where δ, θ are real, and p and n are three-vectors. For the important case of
symmetric scattering matrices (as occurs for backscatter in the sensor coordi-
nates, for example) it follows that n = (1, 0, 0), in which case the most general
scattering matrix may be written in geometrical form, as shown in equation
(1.182):
where the square-root factor is used to keep the total scattered power constant;
that is, |x|2 + |y|2 + |w|2 + |z|2 = |a|2 + |b|2 + |c|2 + |d |2 . Note that we can
relate the vectors in the two systems—lexicographic and Pauli—by a 4 × 4
unitary matrix transformation as shown in equation (1.185):
x 1 1 0 0 a
y 1 0 0 1 −i b
= √ (1.185)
w 2 0 0 1 i c
z 1 −1 0 0 d
and then representing the effect of using a basis set other than the Pauli
matrices by a 4 × 4 unitary matrix transformation of the Pauli coefficients
1.6 The scattering vector formulation 69
k = [U4 ] k (1.187)
Note, as a special case, that the effect of change of coordinates from wave
(FSA) to sensor (BSA) on the Pauli vector can be represented by such a unitary
matrix, as shown in equation (1.188):
x y x y
[S]FSA = ⇒ [S]BSA =
w z −w −z
0 1 0 0
1 0 0 0
⇒ k BSA =
0
k (1.188)
0 0 −i FSA
0 0 i 0
As we did for wave states using the Poincaré sphere, we now need to develop a
physical interpretation of the parameters involved in these higher dimensional
vectors. We begin by recognising that in general, any pair of sets of four basis
matrices for the expansion of [S] can be related in a smooth and continuous
way. To move from one set to another we employ unitary transformations (see
Appendix 2 for a discussion of general unitary transformations). Mathemati-
cally we can then relate the vector of coefficients in one basis k to those in
the second basis set k by a unitary matrix transformation, itself expressed as a
70 Polarised electromagnetic waves
Here we have used the notation P for the set of three Pauli matrices, G for the set
of eight Gell–Mann matrices, and D the sixteen Dirac matrices (see Appendix
2 for definitions of these matrix sets). These compact representations will be
useful when we turn to consider depolarisation effects in Chapter 2.
Depolarisation and
scattering entropy 2
This chapter is concerned with the process of wave depolarisation and its formal
description using the algebraic and geometrical tools developed in Chapter 1.
Depolarisation is inherently a stochastic process, and somewhat destroys the
transfer of vector information from source to far field. Methods for quantifying
this loss of information will be developed based on the concept of scattering
entropy, which we shall see provides a unifying concept of depolarisation across
many types of polarisation problems. InAppendix 3 we outline the basic features
of multivariate statistical signal analysis required for a description of coherence
and its relation to depolarisation in polarimetric studies. Readers not familiar
with these basic stochastic concepts should consult the Appendix for more
details.
The first important point is to draw a distinction between wave depolarisation
and crosspolarisation. These are often confused in the literature, and yet have
important and subtle distinguishing characteristics, as we now demonstrate.
The term ‘crosspolarisation’ refers to any process in wave propagation and
scattering that causes coupling between orthogonal states of polarisation. This
process can be deterministic or stochastic in nature. For example, backscatter
from a dipole oriented at θ degrees has a scattering matrix as shown in equa-
tion (2.1). Here we can see a level of crosspolarisation given by cos θ sin θ. As
a second example, consider wave propagation though a half-wave plate (that is,
a lossless birefringent plate with a thickness that generates a 180-degree phase
shift between its eigenpolarisations at the design wavelength λ). When such a
plate is oriented at θ degrees to the wave coordinates, the Jones matrix has the
Cross-polarising systems
Depolarising
systems
Here again we see a coupling of energy between modes, this time given by
sin 2θ . Such an element is widely used, for example, to rotate the polarisation
of a laser beam.
The key feature relating these two examples is that there exists some deter-
ministic matrix transformation that can be used to remove the crosspolarisation.
Hence neither of these systems depolarise the incident wave.
Depolarisation is therefore a coupling of energy from deterministic into
stochastic modes of the field. It is connected with reversibility of crosspo-
larisation processes. It can be considered a noise-generating process, although
as we shall show, information can still be extracted from the stochastic modes.
Hence as shown schematically in Figure 2.1, all depolarising systems cause
crosspolarisation but not vice versa. This chapter is concerned with systems
that lie in the shaded region of Figure 2.1.
Note that here we define a depolariser as a system that causes noise genera-
tion when illuminated by a pure polarised wave (waves of the form shown in
equation (1.133)). This avoids any of the semantic issues raised, for example,
in Hovenier (2004) concerning the definition of depolarisation when consider-
ing illumination by partially polarised waves. We have already seen in Section
1.5.3 that the Lorentz transformation makes it possible even for simple point
scatterers, characterized by a single amplitude matrix [S], to both increase and
decrease the coherence of partially polarised waves according to the geometry
of the Lorentz boost. However, the conservation of zero wave entropy means
that such systems are here characterized as polarising. We now turn to consider
systems where such a conservation law no longer applies.
This provides an important link with Lorentz geometry via the Minkowski
metric, as discussed in Chapter 1. Hence [J ] is a positive semidefinite (PSD)
Hermitian matrix; that is, its eigenvalues are both real and non-negative. It has
an associated real quadratic form defined as shown in equation (2.4):
w∗T [J ] w ≥ 0 (2.4)
where w is a two-element unitary complex wave vector. We can now use this
matrix to obtain a general expression for the coherence between signals in
arbitrary p and q polarisation channels. We start by defining the two desired
receiver channels on the Poincaré sphere from their unitary w vectors, as shown
in equation (2.5):
cos αP cos αQ
wP = w = (2.5)
sin αP eiδP Q sin αQ eiδQ
The received signals in these components are then obtained by projection onto
the incident E vector. The ‘PQ’ coherence can then be obtained as shown in
equation (2.6):
. + ,
SP = w∗TP E SP SQ∗
⇒ γPQ = + ,+ ,
SQ = w∗TQ E SP SP∗ SQ SQ∗
+ ,
∗T w
(2.6)
w∗T
P E.E Q w∗T
P [J ] w Q
+ , + , =
w∗T ∗T w .w ∗T E.E ∗T w w∗T ∗T
P E.E P Q Q P [J ]w P .w Q [J ]w Q
Hence [J ] contains all the information required to synthesize the coherence for
arbitrary polarisation channels p and q. For this reason it is termed the wave
coherency matrix, and forms the central focus for a quantitative treatment of
wave depolarisation. Note that if p = q then the coherence is always unity as
expected (we then compare a signal channel with itself). Often p and q will be
orthogonal, when for example we wish to measure the full polarisation state of
a wave. In this case we note that if we choose an orthogonal w pair such that
the following condition applies:
[J ] wQ = λwQ (2.7)
—that is, as an eigenvector of [J ]—then from equation (2.6) the coherence will
always be zero. Hence the eigenvalue problem of [J ] is linked to the coherence
74 Depolarisation and scattering entropy
properties of the wave. To pursue this idea further, we note that for a change of
polarisation base we have E 2 = [U2 ] E, and the new coherency matrix can then
be easily derived as shown in equation (2.8):
+ , + ,
[J ]2 = E 2 .E ∗2 = [U2 ] E.E ∗T [U2 ]∗T
+ ,
= [U2 ]. E.E ∗T [U2 ]−1 = [U2 ].[J ].[U2 ]−1 (2.8)
It follows from the properties of Hermitian matrices that we can then always
find a base such that [J ] is diagonal, as shown in equation (2.9):
∗T
λ 0 ∗T
[J ] = [U2 ] . [D] . [U2 ] = w w⊥ . 1 . w w⊥ λ1 ≥ λ2 ≥ 0
0 λ2
(2.9)
[C] = (E + L) + [E + L] (2.11)
where (..) are the polarised components, [..] the depolarising parameters, and
for each, E are those parameters associated with the eigenvectors and L the
eigenvalues. From the general structure of N × N coherency matrices we then
have the following constraints (see Appendix 2):
• L + (L) = N
• E + (E) = dim(SU(N)) − −rank(SU(N))
2.1 The wave coherency matrix 75
With this notation in place we can now write the wave coherency matrix in the
following compact form:
[J ] = (2 + 1) + [0 + 1] (2.12)
It follows from this that in order to consider the effects of wave depolarisation (N
= 2) we need only consider diagonal matrices, as the depolarisation parameter
lies entirely in the eigenvalue spectrum and not in the eigenvectors. We shall
see this is not true for higher dimensional problems. This special case of N = 2
then leads to a further important result known as the wave dichotomy.
In this case the coherence in equation (2.6) will be zero for arbitrary choice of
orthogonal p and q; that is, the signal is indistinguishable from noise, there is no
variation of coherence with polarisation, and depolarisation is complete. The
whole signal is characterized by only a single parameter: the noise variance, λ.
This represents the most extreme case of polarisation ‘memory loss’.
At the other extreme, if we consider a pure polarised wave then det([J ]) = 0
(see Section 1.3.4), and the matrix must then have one zero eigenvalue; that is,
[J ] is of the form shown in equation (2.14), where λ is in this case the squared
amplitude of the polarisation ellipse:
λ 0
[J ]polarised = (2.14)
0 0
In this case the coherence is unity for all choices of p and q, except when
p corresponds to the polarisation state itself. In this singular case the signal
component in the orthogonal channel is by definition zero and hence the coher-
ence function is no longer defined, and the numerator and denominator will
now both be zero. In this case the signal has zero depolarisation and is purely
deterministic in nature.
In general, of course, real signals lie between these two extremes, with
a coherence that depends on the choice of polarisation base. These examples
illustrate why the coherence is itself a bad way to characterize the depolarisation
properties of the signal. Not only does it vary with polarisation state but it
can also be singular for the case of pure polarised waves. We require instead
a basis invariant regular parameter to account for the level of depolarisation
in the signal. This can be defined in several ways—for example, from the
ratio of eigenvalues—by defining the degree of polarisation Dp as shown in
76 Depolarisation and scattering entropy
equation (2.15):
λ1 − λ2
Dp = 0 ≤ Dp ≤ 1 (2.15)
λ1 + λ 2
This function is by definition invariant to changes of polarisation base, and so
represents a fundamental property of the signal. Like coherence magnitude it
lies between 0 and 1; when Dp = 0 we have a noise signal, while Dp = 1 for
pure polarised waves.
This parameterization also allows us to derive a decomposition of the
coherency matrix into ‘polarised’ and ‘depolarised’ components by writing an
arbitrary [J ] as the sum of a polarised wave plus noise, as shown in the top row
of equation (2.16):
λ − λ2 0 1 0
Wave Decomposition 1 [J ] = [U ] . 1 . [U ]∗T + λ2
0 0 0 1
Wave Decomposition 2 [J ] = λ1 w.w∗T + λ2 w⊥ .w∗T
⊥
(2.16)
!
2
Hw = − pi log2 pi 0 ≤ Hw ≤ 1 (2.18)
i=1
When the entropy is zero we have zero uncertainty as to the state of polarisation;
that is, a coherence of one and a purely polarised wave. At the other extreme,
when the entropy is one then we have maximum uncertainty as to the state of
polarisation and a noise signal with a coherence of zero.
The solution to the wave dichotomy remains a matter largely of choice in
wave problems, but as we shall see for higher order coherency matrices the
2.1 The wave coherency matrix 77
+ , +
,
Ex Ex∗ Ex Ey∗ 1 g0 + g1 g2 − ig3
[J ] = + , + , =
Ey Ex∗ Ey Ey∗ 2 g2 + ig3 g0 − g 1
+ , + ,
Ex Ex∗ + Ey Ey∗
+ , + ,
Ex E ∗ − E y E ∗
x y
⇒g= + , +
Ex E ∗ + E y E ∗
,
(2.19)
y x
+ , + ,
i Ex Ey∗ − Ey Ex∗
where now
1 2
det([J ]) = λ1 λ2 = (g − (g12 + g22 + g32 )) ≥ 0
4 0 (2.20)
Trace([J ]) = λ1 + λ2 = g0 > 0
5
3
!
3
2
2
i=1 gi
gi2 = (λ1 − λ2 ) ⇒ Dp = (2.21)
g0
i=1
and so the degree of polarisation can be written as a ratio of two scalars derived
directly from the Stokes vector. The first—the g0 element—is just the total
wave intensity, while the second is the Euclidean norm of a three-vector in
the space of the Poincaré sphere. This suggests a parallel decomposition of the
Stokes vector into polarised and depolarised components, as shown in the top
row of equation (2.22). Here the first component represents a polarised wave
and a point on the Poincaré sphere. The second represents a noise signal, which
has zero everywhere except in the first element. This corresponds directly to
wave decomposition 1 in equation (2.16). The wave dichotomy means we can
also write an arbitrary Stokes vector as the sum of two orthogonal states. The
components of the two-state decomposition can be easily derived as shown in
78 Depolarisation and scattering entropy
⇒ g = g01 −g2
√5g32 2 + g02 √5 (2.22)
i=1 gi
3 2
i=1 gi
g
√533 2 −g3
√5 3
i=1 gi 2
i=1 gi
In summary, we have seen that there are several ways to characterize depolarised
waves. Firstly we note that while depolarisation is associated with a loss of
coherence, it is better to characterize it on the basis of the eigenvalue spectrum
of the wave coherency matrix rather than in terms of the coherence function
itself. Arising from this is the wave dichotomy, whereby the eigenvalues can
be used either as part of a signal + noise model which leads to the degree
of polarisation, or as a mixture of states model which leads to the definition
of wave entropy. Both have a simple representation in terms of the Stokes
vector.
Entropy can be viewed as an information destroying process, and with this in
mind it is interesting to characterize the various ways in which entropy can be
increased or decreased by scattering. The simplest way to increase entropy is to
consider the addition of thermal noise to a signal. Entropy can also be increased
at source, as is the case for passive remote sensing when the source is thermal
radiation. However, for active sensors the source usually has zero entropy by
design. Consequently, any change in entropy must be caused by environmental
scattering and propagation. We now turn to consider a detailed analysis of such
effects.
Clearly [S] will have the effect of changing the eigenvalues and eigenvectors of
[J ], and hence will change the degree of wave depolarisation. By expanding the
matrix product in equation (2.23) and collecting terms of [J ] to form elements
of the corresponding Stokes vectors g i and g s we can rewrite equation (2.23)
as a linear transformation between vectors, as shown in equation (2.24):
g s = [M ] .g i (2.24)
The matrix [M ] is called the Mueller matrix, after Hans Mueller, an MIT
physicist who first derived the properties of such matrices (but seems to have
published very little in the open literature). Parallel development of a similar
approach can be found in the contemporary work of Perrin (1942).
We have already developed a geometrical interpretation of equations of the
form shown in equations (2.23) and (2.24). In equation (1.177) we showed that
the elements of [M ] are all real, despite the fact that [S] is complex. This can
be confirmed by explicit derivation of the elements of [M ] from [S], as we now
demonstrate. We can reformulate equation (2.23) in terms of Stokes vectors by
using the Kronecker matrix product (see Appendix 1), since for any set of three
matrices we can write an expansion as shown in equation (2.25):
x11 x12 .... x1n x11
x21 x22 .... x2n x12
[X ] = .. ⇒x=
:
(2.26)
: : . :
xn1 xn2 .... xnn xnn
Using this result, we can now write the vectorized wave coherency matrix of
the scattered wave as shown in equation (2.27):
aa∗ ab∗ ba∗ bb∗ Ex Ex∗
ac∗ ad ∗ bc∗ bd ∗ ∗
[S] =
a b
⇒ xs = [Y ] xi = . Ex Ey∗
c d ca∗ cb∗ da∗ db ∗ Ey Ex
cc∗ cd ∗ dc∗ dd ∗ Ey Ey∗
(2.27)
Combining equations (2.27) and (2.28) we finally obtain an expression for the
matrix [M ] in terms of the elements of [S] and [Y ], as shown in equation (2.29):
m11 m12 m13 m14
m21 m22 m23 m24
[M ] = m31 m32 m33 m34 mij ∈
which can be expanded to yield an explicit mapping between any given [S]
matrix and the sixteen real elements of [M ], as shown in equation (2.30):
aa∗ + bb∗ + cc∗ + dd ∗ aa∗ − bb∗ + cc∗ − dd ∗ 2Re(ab∗ + cd ∗ ) 2Im(ab∗ + cd ∗ )
∗
1 aa + bb∗ − cc∗ − dd ∗ aa∗ − bb∗ − cc∗ + dd ∗ 2Re(ab∗ − cd ∗ ) 2Im(ab∗ − cd ∗ )
[M ] =
2
2Re(ac ∗ + bd ∗ ) 2Re(ac∗ − bd ∗ ) 2Re(ad ∗ + bc∗ )
2Im(ad ∗ − bc∗ )
−2Im(ac∗ + bd ∗ ) −2Im(ac∗ − bd ∗ ) −2Im(ad ∗ + bc∗ ) 2Re(ad ∗ − bc∗ )
(2.30)
0 0 0 −1
Mueller Property M8: [S]∗T → [M ]T
1 0 0 0
0 1 0 0
Mueller Property M9: [S]∗ → 4 [M ] 4 4 = 0 0 1 0
0 0 0 −1
Mueller Property M10:
a b a −b
[S] = → [M ] ⇒ → [34 ] [M ] [34 ]
c d −c d
1 0 0 0
0 1 0 0
[34 ] =
0 0 −1 0
0 0 0 −1
a b d b
[S] = → [M ] ⇒ → [2 ] [M ]T [2 ]
c d c a
1 0 0 0
0 −1 0 0
[2 ] =
0 0 1 0
0 0 0 1
−1 1 d −b 1
[S] = → 2 [234 ] [M ]T [234 ]
det([S]) −c a x
1 0 0 0
0 −1 0 0
[234 ] =
0 0 −1 0
0 0 0 −1
1
[M ]−1 = [234 ][M ]T [234 ]
x2
82 Depolarisation and scattering entropy
0 0 0 1
The relation M13 is important for application of the reciprocity theorem to scat-
tering problems (van de Hulst, 1981; Saxon, 1955). Property M12 in particular
leads to two important further results, summarized in equation (2.31):
8
1 [M ]T [234 ][M ] = x2 [234 ]
[M ]−1 = 2 [234 ] [M ]T [234 ] ⇒ (2.31)
x Tr([M ]T [234 ][M ]) = −2x2
These relations were first derived by Barakat (1981) and Simon (1987). Finally,
an important simple and widely used relationship between the elements of the
matrix [M ], called the Fry–Kattawar relation (Fry, 1981) is shown in equation
(2.32):
!
4 !
4
m2ij = 4m211 (2.32)
i=1 j=1
Although this relation is widely used in the literature, care must be taken in
its application. For example, there exist matrices that satisfy this relation but
do not have a corresponding single [S] matrix, and so for which the reverse
mapping does not exist. A simple example is diag(1,1,1,–1). Hence equation
(2.32) is a necessary but not sufficient condition for [M ] to correspond to a
single [S] (Hovenier, 1996, 2004).
All of these relations were developed with a view to establishing conditions
for the reversibility of the relation [S] = f −1 ([M ]). In general, [M ] = f ([S])
changes the degree of polarisation of the wave, but if the incident wave entropy
is zero (a purely polarised wave) then the scattered wave entropy is also zero.
This is termed the ‘conservation of zero wave entropy’, and is analogous to
the conservation of light speed in special relativity (see Section 1.5.3). This
property has to do with the question of reversibility of the mapping from [S]
to [M]; that is, [M ] = f ([S]) ⇒ [S] = f −1 ([M ])? We can use property M6
to find the conditions for the inversion to be unique, since z [S] → |z|2 [M ]
for all complex z, so [S] ⇒ [M ] ⇒ [SR ] is possible, but only where [SR ] is
the relative phase scattering matrix (that is, the scattering amplitude matrix can
be reconstructed from [M ], but only up to an arbitrary phase term). This helps
establish conditions for uniqueness, but leaves the thorny issue of existence
to be resolved. There is no guarantee that given a general 4 × 4 real matrix
it will have a corresponding single [S] matrix representation. Before dealing
with existence conditions, however, we turn to consider the special case of
backscatter and the implications of the BSA coordinate system for the Mueller
matrix.
2.2 The Mueller matrix 83
m11 m12 m13 m14
m12 m22 m23 m24
=
(2.33)
−m 13 −m 23 m33 m34
m14 m24 −m34 m44
∗
a b
[S] =
b −d
1 0 0 0
1 0 1 0 0
⇒ [M ] =
2 0 0 1 0
0 0 0 −1
aa∗ + 2bb∗ + dd ∗ aa∗ − dd ∗ 2Re(ab∗ − bd ∗ ) 2Im(ab∗ − bd ∗ )
aa∗ − dd ∗ aa∗ − 2bb∗ + dd ∗ 2Re(ab∗ + bd ∗ ) 2Im(ab∗ + bd ∗ )
·
2Re(ab∗ − bd ∗ ) 2Re(ab∗ + bd ∗ ) 2Re(bb∗ − ad ∗ ) ∗
2Im(−ad )
−2Im(ab∗ − bd ∗ ) −2Im(ab∗ + bd ∗ ) −2Im(−ad ∗ ) ∗
−2Re(ad + bc )∗
m11 m12 m13 m14
m12 m22 m23 m24
=
(2.34)
m13 m23 m33 m34
m14 m24 m34 m44
84 Depolarisation and scattering entropy
This matrix converts all Stokes vectors into a randomly polarised wave, hence its
name. However, there is no corresponding single [S] matrix. Clearly, therefore,
our analysis of depolarisation must go beyond relations like equation (2.32)
A and look more closely at the structure of Mueller matrices in general.
B The set of Mueller matrices [M ] form a subset of all possible 4×4 real matri-
C ces. Hence not all real matrices correspond to Mueller matrices and, because
of depolarisation like that shown in equation (2.35), of those that do there is
no guarantee that the inverse mapping to a single [S] will exist. This situation
is summarized graphically in Figure 2.2. It is of practical interest, therefore,
to determine two sets of conditions on a given matrix [M ]—firstly to test if it
A – the set of 4 × 4 real matrices corresponds to a Mueller matrix at all; and secondly, if it does, whether there is
B – the set of Mueller matrices
C – the set of Mueller matrices
a corresponding single amplitude matrix [S]. By doing this we may then estab-
for which [S] = f –1([M]) lish existence of the inverse mapping, and at the same time classify the different
possible types of depolarisation. A full analysis of this problem follows Section
Fig. 2.2 The set of Mueller matrices as a 2.5, but here we make some general observations about the related topic of the
subset of real 4 × 4 matrices
Stokes criterion.
We can, for example, enforce these by insisting that the following quadratic
form be positive semidefinite:
where [234 ] is defined in property M12. This set of conditions, first derived in
van der Mee (1992, 1993) and Givens (1993) in turn requires the eigenvalues of
the composite matrix [234 ] [M ]T [234 ] [M ] to be real and non-negative. This
can then be directly taken as a test of validity for candidate Mueller matrices.
2.3 The scattering coherency matrix formulation 85
There are, however, two alternative ways of forming a matrix of all possible
second-order products: the covariance and coherency matrices defined as outer
products of the corresponding scattering vectors in the lexicographic and Pauli
bases (see Section 1.6), as shown in equation (2.40) (Cloude, 1985, 1986).
These matrices have a useful symmetry as, unlike [Y ], they are complex Her-
mitian, having real elements along the diagonal and complex conjugate entries
in symmetric off-diagonal positions. This reordering of the product terms has
a more direct relationship to coherence between the elements of the k vector,
hence the name coherency matrix. For example, [C] can be used to generalize
the arguments about optimum polarisations treated for [S] matrices in Section
1.4.3 to depolarising scenarios (see Tragl (1990)). Furthermore, we can exploit
the Hermitian symmetry by noting that all the 2 × 2 principal minors of [C]
and [T ] are zero, so for example, in equation (2.40) we have aa*bb* – ab*ba*
= aa*cc* – ac*ca* = 0 and so on.
For a 4 × 4 matrix there are nine such minors, and for a 3 × 3 matrix there are
four. In fact these provide just the set of required constraint equations between
86 Depolarisation and scattering entropy
the Mueller matrix elements discussed around equation (2.36). Rather than
consider these in any more detail (to investigate further see Huynen (1970,
1987) and Hovenier (1994, 1996, 2004)) we proceed to the next key step and
consider the effect of depolarisation on the form of the coherency matrix.
a
a b b ∗
[S] = ⇒ [C] = k L k L =
∗T
. a b∗ c∗ d ∗
c d c
d
∗
aa ab∗ ac∗ ad ∗
ba∗ bb∗ bc∗ bd ∗
=ca∗ cb∗ cc∗ cd ∗ = [C]
∗T
The covariance and coherency matrices (and indeed the [Y ] matrix) are then
obtained as averages over the distribution, so, for example, over a population
of L independent samples we can obtain an estimate of the mean coherency
matrix as shown in equation (2.42):
!
L
[T ] = k i k ∗T
i (2.42)
i=1
In this case the principal minor relations, by the Schwarz inequality (see
Appendix 3), now have the general form of non-negative inequalities, auch
as aa∗ bb∗ − ab∗ ba∗ ≥ 0. The Mueller matrix [M ] corresponding to
[T ] is often called a sum-of-pure Mueller matrices (SPM) in the literature
(Hovenier, 2004).
where λi are the real eigenvalues, and ei the corresponding complex eigenvec-
tors. Note that [C] has the same eigenvalues as [T ] but its eigenvectors, eL ,
are related to those of [T ], eP by a 4 × 4 unitary matrix, as shown in equation
(2.44):
1 1 0 0
1 0 0 1 −i ∗T
eL = ULP4 eP = √
e ⇒ [C] = U
P [T ] ULP4
2 0 0 1 i LP4
1 −1 0 0
(2.44)
We have seen that for BSA reciprocal backscatter problems, the [S] matrix
is symmetric, in which case the transformation from [T ] to [C] is often
reformulated as a 3 × 3 matrix, as shown in equation (2.45):
√a a+d 1 1 √0
1 1
−→ 2b = [ULP3 ] √ a − d = √ 0 0 2 e P
b=c
d 2 2b 2 1 −1 0 (2.45)
[T ] = λ1 e1 e∗T ∗T ∗T ∗T
1 + λ2 e 2 e 2 + λ3 e 3 e 3 + λ4 e 4 e 4 (2.46)
‹[C]› - 4 × 4 Covariance
‹..›
[S]
‹..› ‹..›
t11 t12 t13 t14
t ∗ t22 t23 t24
[T ] = 12
t ∗
⇒ [M ] =
13
∗
t23 t33 t34
∗
t14 ∗
t24 ∗
t34 t44
∗ − i(t − t ∗ ) ∗ + i(t − t ∗ ) ∗ − i(t − t ∗ )
t11 + t22 + t33 + t44 t12 + t12 34 34 t13 + t13 24 24 t14 + t14 23 23
1 ∗ ∗
t12 + t12 + i(t34 − t34 ) t11 + t22 − t33 − t44 ∗ + i(t − t ∗ )
t23 + t23 14 14
∗ − i(t − t ∗ )
t24 + t24 13 13
2 t13 + t13
∗ − i(t − t ∗ )
24 24
∗ ∗ )
t23 + t23 − i(t14 − t14 t11 − t22 + t33 − t44 t34 + t + i(t12 − t ∗ )
∗
34 12
∗ + i(t − t ∗ )
t14 + t14 ∗ + i(t − t ∗ )
t24 + t24 ∗ − i(t − t ∗ )
t34 + t34 t11 − t22 − t33 + t44
23 23 13 13 12 12
(2.48)
m11 m12 m13 m14
m m24
m22 m23
[M ] = 21 ⇒ [T ] =
m31 m32 m33 m34
m41 m42 m43 m44
m11 + m22 + m33 + m44 m12 + m21 − i(m34 − m43 ) m13 + m31 + i(m24 − m42 ) m14 + m41 − i(m23 − m32 )
1m12 + m21 + i(m34 − m43 ) m11 + m22 − m33 − m44 m23 + m32 + i(m14 − m41 ) m24 + m42 − i(m13 − m31 )
2 m13 + m31 − i(m24 − m42 ) m23 + m32 − i(m14 − m41 ) m11 − m22 + m33 − m44 m34 + m43 + i(m12 − m21 )
m14 + m41 + i(m23 − m32 ) m24 + m42 + i(m13 − m31 ) m34 + m43 − i(m12 − m21 ) m11 − m22 − m33 + m44
(2.49)
These represent the most general scattering case in the wave or FSA coordinate
system. Note that [T ] in the sensor or BSA coordinates has the same eigenvalues
as in the wave system, but the eigenvectors are related by a unitary matrix
2.3 The scattering coherency matrix formulation 89
0 1 0 0
0 1 1 0 0 0
eFSA = U4C eBSA =
0
e (2.50)
0 0 −i BSA
0 0 i 0
t11 t12 t13 0
∗
t12 t22 t23 0
[T ]BSA =
t ∗
13
∗
t23 t33 0
0 0 0 0
t11 + t22 + t33 ∗
t12 + t12 ∗
t13 + t13 ∗ )
−i(t23 − t23
∗ ∗ ∗ )
1 t12 + t12 t11 + t22 − t33 t23 + t23 −i(t13 − t13
⇒ [M ]BSA = (2.51)
2
∗
t13 + t13 ∗
t23 + t23 t11 − t22 + t33 ∗ )
i(t12 − t12
∗ )
−i(t23 − t23 ∗ )
−i(t13 − t13 ∗ )
i(t12 − t12 −t11 + t22 + t33
m11 m12 m13 m14
m12 m22 m23 m24
[M ]BSA =
m
13 m23 m33 m34
m14 m24 m34 m44
m11 + m22 + m33 − m44 m12 − im34 m13 + im24
1
⇒ [T ]BSA = m12 + im34 m11 + m22 − m33 + m44 m23 + im14
2
m13 − im24 m23 − im14 m11 − m22 + m33 − m44
(2.52)
t11 t12 0 t14
∗
t12 t22 0 t24
[T ]FSA =
0
0 0 0
∗
t14 ∗
t24 0 t44
t11 + t22 + t44 ∗
t12 + t12 ∗ )
i(t24 − t24 ∗
t14 + t14
∗ ∗ ) ∗
1 t12 + t12 t11 + t22 − t44 i(t14 − t14 t24 + t24
⇒ [M ]FSA = ∗ ) ∗ ) ∗ )
(2.53)
2 −i(t24 − t24 −i(t14 − t14 t11 − t22 − t44 i(t12 − t12
∗
t14 + t14 ∗
t24 + t24 ∗ )
−i(t12 − t12 t11 − t22 + t44
90 Depolarisation and scattering entropy
m11 m12 m13 m14
m12 m22 m23 m24
[M ]FSA =
−m13 −m23 m33 m34
m14 m24 −m34 m44
m11 + m22 + m33 + m44 m12 − im34 m14 − im23
1
⇒ [T ]FSA = m12 + im34 m11 + m22 − m33 − m44 m24 − im13
2
m14 + im23 m24 + im13 m11 − m22 − m33 + m44
(2.54)
There are then four important results derived directly from the eigenvector
expansion of [T ]:
1. The conservation of zero wave entropy, introduced after equation (2.32),
implies that [T ] has only one non-zero eigenvalue, and so we can state
the following important result (Cloude, 1989):
[M ] → [T ] ⇒ λi ≥ 0, i = 1, 2, 3, 4 (2.56)
equations (2.51) and (2.53)). This leads us directly to the trace condi-
tion on backscatter Mueller matrices discussed in equation (2.38). As
a special case we consider the possibility of designing wave depolaris-
ers for use in backscatter. We may, for example, consider using a very
rough surface or multiple scattering cavity with backscatter geometry.
We define the following general form for an isotropic depolariser with
0 ≤ δ ≤ 1:
1 0 0 0
0 δ 0 0
[M ] =
0 0
(2.58)
δ 0
0 0 0 δ
Reciprocity
/
m11 − m22 − m33 − m44 = 1 − 3δ = 0 (sensor coordinates)
⇒
m11 − m22 + m33 − m44 = 1 − δ = 0 (wave coordinates)
(2.59)
1 0T 1 0T
[U−1 ] [UN ] = = (2.60)
0 UN −1 0 exp(iHN −1 )
!
M
HN −1 = hk k ⇒ h = depolarisation state vector
k=1
This operator acts to convert the first column of the coherency matrix to the
identity. The next stage is to apply such an operator to the matrix of ranked
eigenvectors ei of the coherency matrix
[UN ] = e1 e2 . . . eN (2.61)
and to identify the submatrix U N −1 with the depolarising aspects of the scatter-
ing process. In this way U N −1 involves continuous smooth transformation away
from the polarised reference state (the dominant eigenvector). The submatrix
U N −1 may be further parameterized in terms of an N − 1 × N − 1 Hermitian
matrix, related to the unitary transformation by a matrix exponential, and itself
conveniently expanded in terms of a set of scalar parameters being the basis
elements of the underlying algebra. This vector of coefficients we then term the
depolarisation state vector (see equation (2.60)), as it determines the nature of
the depolarisation, with different state vectors representing different types and
degrees of depolarised scattering.
This then leads us to the following notation to characterize the number
of parameters involved in polarising and depolarising components of the
decomposition of a general N × N coherency matrix T N :
T N = (E + L) + [E + L] (2.62)
where [..] are the depolarising parameters and (..) the polarising terms, and
for each, E are those parameters associated with the eigenvectors and L the
eigenvalues. From the general structure of N × N coherency matrices we then
have the following constraints:
• [L] + (L) = N
• [E] + (E) = dim(SU(N)) − rank(SU(N)) = N(N − 1)
Note that the total number of eigenvector parameters is given by the sum [E]
+ (E) = dim(SU(N)) − r(SU(N)), where dim() = N2 − 1 is the dimension of
the group and r = N − 1 is the rank of the Cartan sub-algebra or the number of
mutually commuting generators (see Appendix 2). For example, for N = 1 there
are no useful eigenvector parameters, and for N = 2 we only have two, while
for N = 4 (the most general bistatic scattering case) we have twelve parameters
available. We now look at special cases to illustrate how this classification
scheme works, and in particular how general scattering symmetries restrict the
number of parameters.
2.4 General theory of scattering entropy 93
T2 = (2 + 1) + [0 + 1] (2.65)
Note that there are no depolarisation parameters in the eigenvectors, and that
all depolarisation effects lie in the eigenvalue spectrum. This leads to the
wave dichotomy and degree of polarisation ideas discussed in Section 2.1.1.
More interesting possibilities exist for higher-dimensional matrices, as we now
consider.
In this case [T ]/[M ] have a maximum of nine parameters, and SU(3) is the gov-
erning unitary group for the eigenvectors of [T ]. However, the unitary reduction
operator means that all depolarisation effects are controlled by SU(2), which has
dimension 3 and rank 1. Hence in N = 3 backscatter the polarised/depolarised
decomposition can be written in compact form as follows:
T3 = (4 + 1) + [2 + 2] (2.68)
This shows that the eigenvectors now contain two depolarising parameters,
complemented by two real eigenvalues. The polarised component has four
eigenvector parameters and one eigenvalue (its amplitude). Note that this
decomposition makes no assumptions about scattering symmetry (other than
reciprocity). We now turn to consider the restrictions placed by scattering
symmetries on the form of the coherency and Mueller matrices in backscatter.
k0 k0
k P = k1 ⇒ k Q = k1
k2 −k2
t11 t12 t13 t11 t12 −t13
⇒ [T ] = [TP ] + TQ = t12∗ t22 t23 + t12
∗ t22 −t23
∗
t13 ∗
t23 t33 ∗
−t13 ∗
−t23 t33
t11 t12 0
= t12
∗ t22 0 (2.69)
0 0 t33
2.4 General theory of scattering entropy 95
In this case we see that one consequence of reflection symmetry + is that ,the
∗
coherence
+ , between co- and crosspolarised channels will be zero: S HH S HV =
∗
SVV SVH = 0, and so on. Reintroducing the idea that θ may not be zero
produces the following general form for a reflection symmetric medium:
1 0 0 t11 t12 0 1 0 0
[Tθ ] = 0 cos 2θ − sin 2θ t12
∗ t22 0 0 cos 2θ sin 2θ
0 sin 2θ cos 2θ 0 0 t33 0 − sin 2θ cos 2θ
(2.70)
Note that this will introduce apparent correlations between co- and cross-
channels due to crosspolarisation (but not depolarisation) by the axis mis-
alignment.
The second type of symmetry to consider is rotation symmetry. Here we
consider the form of [T ] that remains invariant under arbitrary rotations in the
plane of polarisation. By definition, and as an extension of the discussion around
equation (2.69), [T ] must then be formed from linear combinations of the outer
product of eigenvectors of the plane rotation matrix (see Section 4.1.1). In this
case we can therefore predict the general form of a rotation symmetric [T ] as
shown in equation (2.71):
a 0 0
[T ] = av1 v ∗T ∗T ∗T
1 + bv 2 v 2 + cv 3 v 3 = 0 b+c i(b − c) (2.71)
0 −i(b − c) b+c
Finally we consider the most general symmetry, formed by insisting that every
axis AA’, for all θ, has the PQ reflection symmetry. This is termed azimuthal
symmetry, and the form of [T ] can be found by adding reflection symmetric
versions of the rotation matrix of equation (2.71) to obtain equation (2.72):
a 0 0 a 0 0
[T ] = 0 b+c i(b − c) + 0 b+c −i(b − c)
0 −i(b − c) b+c 0 i(b − c) b+c
a 0 0
= 2 0 b + c 0 (2.72)
0 0 b+c
Here we have obtained a very simple diagonal coherency matrix, with two
equal diagonal values. We now turn to consider the polarised/depolarised
decomposition for each of these symmetry cases.
equation (2.73):
m 0 0
[T3 ] = 0 mκ 0
0 0 mκ
1+κ 0 1−κ
m
⇔ [C3 ] = 0 2κ 0
2 1−κ 0 (2.73)
1+κ
1 + 2κ 0 0 0
m 0 1 0 0
⇔ [M ] =
2 0 0 1 0
0 0 0 2κ − 1
In this symmetry assumption the polarising/depolarising decomposition there-
fore reduces to equation (2.74):
T3az = (0 + 1) + [0 + 1] (2.74)
In this case only two parameters are required to characterize the medium (such
as VV power and HH/VV coherence or HV, but not both HV and HH/VV
coherence, as these two are now related). A convenient parameterization for
the most usual situation in remote sensing applications (a > b + c in equation
(2.72)) is provided in terms of the polarised power (m) and depolarisation or
noise parameter (κ) shown in equation (2.73).
Note that there are two important special cases. When κ = 0 we have zero
depolarisation and backscatter from a sphere symmetric scatterer (or random
collection of such), and at the other extreme κ = 1, and [T ] reduces to the 3 × 3
identity matrix. This represents the most depolarising case. Note, however, that
the Mueller matrix for this case still does not equal a noise source ideal depo-
lariser (equation (2.35)); in fact, it corresponds to a 33% isotropic depolariser.
As mentioned in the discussion around equation (2.59), this is due to the fact
that HV and VH remain correlated for all backscatter problems, and hence an
identity coherency matrix corresponds not to equal noise power in all channels,
but with the cross-channel having only half the power of the copolarised chan-
nels. This indicates that a proper analysis of polarised signals in the presence
of noise requires a full 4 × 4 coherency matrix analysis. This residual structure
implies that even in the most depolarising backscatter case there will be some
residual information on the scatterer properties, even just that following from
the reciprocity theorem.
[T ]ref = (3 + 1) + [0 + 2] (2.76)
This result shows formally that for backscatter from random media with reflec-
tion symmetry [E] = 0 and [L] = 2, so the eigenvectors contain no information
about depolarisation, only about the polarising influence of the dominant eigen-
vector. This results because the 3 × 3 matrix of eigenvectors for such media can
be written as a function of a 2 × 2 unitary matrix, as shown in equation (2.77):
[U2 ] 0 1 0
[U3 ] = ⇒ [U−1 ] [U3 ] = (2.77)
0 1 0 1
All the depolarisation information for such media is therefore contained in the
two minor eigenvalues of [T ]. These eigenvalues are not usually used them-
selves as measures of depolarisation in such media. Instead, widespread use is
made of two functions of the eigenvalues λi : scattering entropy and anisotropy
(Pottier, 1997), defined as in equations (2.78) and (2.79):
!
3
λi
H =− Pi log3 Pi Pi = 5 0≤H ≤1 (2.78)
λ
i=1
λ2 − λ3
A= 0≤A≤1 (2.79)
λ2 + λ 3
The entropy is zero for zero depolarisation; that is, when [T ] has only one
non-zero eigenvalue. At the other extreme, H = 1 when [T ] is diagonal and
maximum depolarisation occurs. The anisotropy is a useful parameter for
assessing the type of symmetry. For azimuthal symmetry the minor eigenvalues
are equal, and so A = 0. On the other hand, if only the smallest eigenvalue falls
to zero then A = 1, and we have significant departure from azimuthal symmetry.
This situation occurs, for example, for in-plane scattering from one-dimensional
rough surfaces and in scattering by chiral particles (see Chapter 3).
The spread of total scattered power across the eigenvalues is evidently a good
indicator of depolarisation. This can be turned into a quantitative measure by
first normalizing the eigenvalues to unit sum, so we define three parameter Pi ,
as shown in equation (2.81). These can then be interpreted as probabilities of
the statistically independent ‘polarised’states, given by the eigenvectors ei . The
spread of probabilities can then be represented by a single scalar, the entropy
being defined as shown in equation (2.81):
λi ! 3
0 ≤ Pi = 5 ≤ 1 ⇒ H = − Pi log3 Pi 0≤H ≤1 (2.81)
λ
i=1
This then provides a suitable depolarising parameter; which is zero for zero
depolarisation (while still allowing crosspolarisation), and one for maximum
depolarisation. But what of the polarised component? From each eigenvector
we can select the scattering mechanism α as a suitable basic invariant polarised
parameter, as shown in equation (2.82).
e1 cos α
e = e2 = eiχ sin α cos ψeiφ1 ⇒ α = cos−1 (|e1 |) 0 ≤ α ≤ 90◦
e3 sin α sin ψeiφ2
(2.82)
α = P1 α1 + P2 α2 + P3 α3 (2.84)
Note that the form of the unitary matrix of eigenvectors shown in equation
(2.83) is overparameterized. The unitary reduction operator provides a much
more rigorous approach (for example, the columns of the unitary matrix are
not independent, as implied in equation (2.83), as they are all mutually orthog-
onal). However, the approach in equation (2.83) has a strong physical appeal
in terms of the probabilistic interpretation, and provides a useful parameteriza-
tion of the mean scattering mechanism. We have therefore derived a suitable
pair of parameters: entropy H and scattering mechanism α. To aid interpre-
tation of different types of polariser/depolariser, use can then be made of the
two-dimensional H/α or entropy/alpha plane, as shown in Figure 2.5 (Cloude,
1996).
2.4 General theory of scattering entropy 99
70
60
Alpha (degrees)
50 Feasible region
40
30
line
etry
20 mm
hal sy
mut Curve I
10 Azi
0
0 0.2 0.4 0.6 0.8 1 Fig. 2.5 Entropy/alpha diagram for N = 3
Entropy H backscatter
1 0 0
CURVE I [T ]I = 0 m 0 0 ≤ m ≤ 1
0 0 m
0 0 0
[T ]II = 0 1 0 ≤ m ≤ 0.5 (2.85)
0 0 2m
CURVE II
2m − 1 0 0
0 1 0 0.5 ≤ m ≤ 1
II =
[T ]
0 0 1
With entropy and anisotropy now defined, several combinations of these two
have been proposed as suitable for the enhancement of important special cases of
depolarisation. These are shown as a1 to a4 in equation (2.86). These parameters
have large values for the following cases:
Again we can define an entropy/alpha plane as shown in Figure 2.6. In this case
the upper and lower bounds can be found by considering the parameterized
2.4 General theory of scattering entropy 101
80
70
60
Alpha (degrees)
50
Feasible region
40
30
20
10
0
0 0.2 0.4 0.6 0.8 1 Fig. 2.6 Entropy/alpha diagram for N = 2
Entropy H depolarisation
In this case the lower and upper curves show a symmetry about α = 45 degrees.
This can be traced to the two-dimensional eigenvector space, one of which has
α = 0 and the other π/2. In the N = 3 case we saw a 2:1 bias in favour of the
π /2 eigenvector subspace, which tends to lift the alpha range with increasing
entropy (Figure 2.5).
In this N = 2 case we also have a simple geometrical interpretation of α on
the Poincaré sphere (see Figure 1.12). In the special case that x and y in equation
(2.87) represent polarisation wave states (see Section 1.33), the angles α and δ
are then related to the orientation and ellipticity angles θ, τ of the polarisation
ellipse via a spherical triangle construction on the Poincaré sphere. However,
here we have generalized this idea so that x and y can now be arbitrary channels
of data; that is, two elements of the [S] matrix, anticipating the ideas of compact
polarimetry (discussed more fully in Section 9.3.4).
The averaging implied by the entropy/alpha approach does not pick
out the state P corresponding to the maximum eigenvalue as in the
polarised/depolarised decomposition, but instead forms an average based on
a probabilistic interpretation of making measurements on the wave and obtain-
ing polarisation X and Y with probabilities P1 and P2 respectively. Hence the
average polarisation state would have a corresponding alpha value given by α.
For example, when the coherency matrix approaches the identity (noise), then
α = π4 , being a mixture of the state X (α = 0) with its antipodal orthogonal
state Y (α= π /2).
In radar there are two important special cases when the N = 2 formalism
becomes important. It is often advantageous in radar design (from a cost, data
rate and coverage point of view) to employ a single transmitted polarisation
102 Depolarisation and scattering entropy
(2.91)
2.4 General theory of scattering entropy 103
It follows that the coherency matrices for reflection symmetric random media
take the form shown in equation (2.92), from which we note the presence of
complex off-diagonal terms.
9 :
[JH ] = k H k ∗T
H
; <
1 |k0 |2 + |k1 |2 cos2 2θ + 2Re(k0 k1∗ ) cos 2θ |k1 |2 cos 2θ sin 2θ + k0 k1∗ sin 2θ
=
2 |k1 |2 cos 2θ sin 2θ + k1 k0∗ sin 2θ |k1 |2 sin2 2θ
9 :
[JV ] = k V k ∗T
V
; <
1 |k0 |2 + |k1 |2 cos2 2θ − 2Re(k0 k1∗ ) cos 2θ |k1 |2 cos 2θ sin 2θ − k0 k1∗ sin 2θ
=
2 |k1 |2 cos 2θ sin 2θ − k1 k0∗ sin 2θ |k1 |2 sin2 2θ
(2.92)
Only for zero tilt (θ = 0) or uniformly random θ do these diagonal terms dis-
appear. This suggests that the phase terms will be important for identifying
coherent point scatterers. Using a coherent (zero entropy) assumption we can
also relate the dual polarisation alpha parameter to the scattering matrix ele-
ments for a symmetric point (SXY = 0; see equation (4.12)) as shown in
equation (2.93):
(SXX − SYY ) sin 2θ
tan α2 =
(SXX + SYY ) + (SXX − SYY ) cos 2θ
(2.93)
(SXX − SYY ) sin 2θ
δ2 = arg
(SXX + SYY ) + (SXX − SYY ) cos 2θ
In the more general case of a non-symmetric point scatterer these relations take
the form shown in equation (2.94):
2SXY cos 2θ + (SXX − SYY ) sin 2θ
tan α2 =
(SXX + SYY ) + (SXX − SYY ) cos 2θ − 2SXY sin 2θ
(2.94)
2SXY cos 2θ + (SXX − SYY ) sin 2θ
δ2 = arg
(SXX + SYY ) + (SXX − SYY ) cos 2θ − 2SXY sin 2θ
[T ]bistatic = (6 + 1) + [6 + 3] (2.95)
.
E AS = S β i β s .P A
⇒ P B .E As = P A .E Bs
E BS = S −β i −β s .P B
Reciprocity Theorem
T
⇒ S β i β s = S −β i −β s (2.96)
In equation (2.97) we show explicitly how the elements of the [S] matrices are
related:
0 1 T 1 0
[S] = S −β i −β s
−1 0 0 −1
a b a −c
⇒S= ⇒ Sα = (2.97)
c d −b d
Theseare just two elements of the scattering matrix group identified in equation
(1.145). As shown there, this can be interpreted in terms of the scatterer as a
180-degree rotation about the bisectrix b, as shown in Figure 2.7. This we define
as the scatterer being in its reciprocal position. Finally, we show the effect on
Scatterer
[S]
x
x
bi bs
A B
y
y
Incident Bisectrix Scattered
Fig. 2.7 Reciprocity interpreted as a rotation wave wave
about the bisectrix b
2.4 General theory of scattering entropy 105
a+d a+d
a−d
k= →k = a−d (2.98)
b+c α −(b + c)
i(b − c) i(b − c)
t11 t12 0 t14
9 : t ∗ t 0 t24 9 :
= ⇔ M reciprocity
reciprocity 12 22
TFSA 0 0 t33 0 FSA
∗
t14 ∗
t24 0 t44
m11 m12 m13 m14
m12 m22 m23 m24
=
−m13 −m23
(2.99)
m33 m34
m14 m24 −m34 m44
[U2 ] 0 0
[U3 ] 0
[U4 ] = ⇒ [U−1 ] [U4 ] = 0 1 0 (2.100)
0 1
0 0 1
However, there is now one extra depolarising eigenvalue to consider, and so the
polarised/depolarised can be written in terms of a reduced set of ten parameters,
as shown in equation (2.101):
recip
T4 = (4 + 1) + [2 + 3] (2.101)
Following this logic there are two other symmetry combinations to consider
from the scattering matrix group. In the first we combine a scatterer with its
mirror image in the plane of scattering, defined from the incident and scattered
wave vectors k, as shown in Figure 1.14. The scattering vectors and averaged
coherency and Mueller matrices for this case then have the form shown in
106 Depolarisation and scattering entropy
equation (2.102):
a+d a+d
a−d a−d
k=
b + c → k β = −(b + c)
i(b − c) −i(b − c)
t11 t12 0 0 m11 m12 0 0
t ∗ t22 0 0 m21 m22 0 0
→ T = 0
12 ⇔ M =
0 t33 t34 0 0 m33 m34
0 ∗
0 t34 t44 0 0 m43 m44
(2.102)
This case is again governed by SU(4), but the 4 × 4 unitary matrix can be
factored in terms of 2 × 2 unitary matrices, as shown in equation (2.103):
[U2 ] [0] [I2 ] [0]
[U4 ] = ⇒ [U−1 ] [U4 ] = (2.103)
[0] [U2 ] [0] [U2 ]
where [I2 ] is the 2 × 2 identity matrix, and [0] the 2 × 2 null matrix. Again
SU(2) governs the depolarisation subspace, and the polarised/depolarised
decomposition has eight parameters organized as shown in equation (2.104):
plane
T4 = (2 + 1) + [2 + 3] (2.104)
i(b − c) −i(b − c)
t11 t12 t13 0 m11 m12 m13 m14
t ∗ t22 t23 0 m12 m22 m23 m24
→ T =
t ∗ t ∗ t33 0 ⇔ M = m13
12
13 23 m23 m33 m34
0 0 0 t44 −m14 −m24 −m43 m44
(2.105)
Again the SU(4) eigenvector dependence can be represented as a 3 × 3 unitary
matrix as shown in equation (2.106), so again SU(2) controls the depolarisation
subspace.
[U2 ] 0 0
[U3 ] 0
[U4 ] = ⇒ [U−1 ] [U4 ] = 0 1 0 (2.106)
0 1
0 0 1
The polarised/depolarised decomposition then has ten parameters arranged as
shown in equation (2.107):
In the most general symmetry case, when all the above symmetries apply—as,
for example, for a random cloud of spheroidal particles—we obtain the averaged
bistatic coherency matrix by adding coherency matrices for all elements of the
scattering matrix group, as shown in equation (2.108):
a+d a+d a+d a+d
a−d a−d a−d a−d
k= → kα = → kβ = → ky =
b+c −(b + c) −(b + c) b+c
i(b − c) i(b − c) −i(b − c) −i(b − c)
t11 t12 0 0 m11 m12 0 0
+ symmetric , ∗
t12 t22 0 0
m
12 m22 0 0
⇒ Twave = ⇔ M =
0 0 t33 0 0 0 m33 m34
0 0 0 t44 0 0 −m34 m44
(2.108)
bistatic+symmetry
[T ] −−−−−−−−−→(2 + 1) + [0 + 3] (2.110)
2α
cos α
(2 + 1) → λmax (2.111)
sin αeiδ
!
4
λi
H =− Pi log4 Pi 0 ≤ H ≤ 1, Pi = 5
λ
i=1 (2.112)
λi − λj
Aij = 0 ≤ Aij ≤ 1, i>j
λi + λj
All vary between 0 and 1, with entropy being a general measure of total depolar-
isation. It is zero for non-depolarising systems, and 1 for the ideal depolariser.
The anisotropies then provide information about the variation of depolarisation
with changes in polarisation: if they are zero then there is no variation, while
as they approach 1 there exists subspaces in the polarisation domain where
depolarisation can be small. This then leads us naturally back to the idea of the
entropy/alpha decomposition.
!
4 !
4
H =− Pi log4 Pi 0 ≤ H ≤ 1, α= Pi αi 0 ≤ α ≤ 90◦ (2.114)
i=1 i=1
Again there are bounding curves in this plane. The upper and lower values for
α as a function of increasing entropy can be calculated as shown in equation
(2.115). Figure 2.9 shows how the valid region in the H/α plane for N = 4
bistatic scattering differs from that used in the N = 3 backscatter and N = 2
dual polarisation cases.
At low entropies there is little difference between the N = 2 and N = 3
cases, but at higher entropy values the alpha range for N = 4 is shifted to
higher values. This is a consequence of the addition of a new eigenvector, itself
having an alpha value of π/2. This acts to lift the mean alpha value for any
given entropy. In the limit of H = 1, the alpha range is reduced to a single point
at 3π/8. Again the utility of this diagram is not so much in its high fidelity
parameterization of depolarisation in N = 4 cases, but in the convenience of
representing the most general bistatic scattering problem by one polarising and
one depolarising parameter, in such a way as to maintain invariance to changes
80
70
60
50
Alpha
40
30
20
10
0
0 0.2 0.4 0.6 0.8 1 Fig. 2.9 Summary of entropy/alpha dia-
Entropy grams for N = 2, 3, and 4 scattering
110 Depolarisation and scattering entropy
of polarisation base.
1 0 0 0 3mπ
α=
0 m 0 0 2(1 + 3m)
TI =
0
⇒
0 m 0
1 m3m
H = − log4
0 0 0 m 1 + 3m (1 + 3m)3m−1
π
0 0 0 0
α=
0 1 0 0 2
0 ≤ m ≤ 0.5 ⇒
0 0 2m 0 1 2m4m
H = − log4
0 0 0 2m (1 + 4m) (1 + 4m)4m+1
2m − 1 0 0 0
3π
0 α=
1 0 0 0.5 ≤ m ≤ 1 ⇒ 4m +4
0 0 1 0 1 (2m − 1)2m−1
H = − log4
0 0 0 1 (2 + 2m) (2 + 2m)2+2m
(2.115)
The most extreme example is the isotropic depolariser, with a Mueller matrix
[MI ] of the form shown on the left-hand side of equation (2.117):
1 0 0 0 1 0 0 0
0 0 0 0 0 δ 0 0
[MI ] =
0 0 0 0 ⇒ [MII ] = 0 0 δ 0
0 0 0 0 0 0 0 δ
1 0T
[MIII ] =
0 [O]3 [][O]T3
⇒ δ1 0 0 (2.117)
[] = 0 δ2 0
0 0 δ3
This matrix converts all Stokes vectors into a randomly polarised wave, but
there is no corresponding single [S] matrix. This form leads to a standard
2.5 Characterization of depolarising systems 111
1 0 0 0
0 |ε| ≤ 1
δ 0 0
[M ] =
0
⇒ |δ| ≤ 1 (2.118)
0 δ 0
2δ − ε ≤ 1
0 0 0 ε
Again we note important physical restrictions on the diagonal elements from the
forward scattering symmetry. Given the need to always add additional constraint
equations, we can ask if all triplets δ1 , δ2 , δ3 are physically consistent, and
whether the forms in equation (2.117) exhaust all possibilities.
In this regard we have shown how we can use the coherency matrix concept
to classify depolarisers. For example, by mapping the general depolariser
of equation (2.117) into [T ] we see that the real diagonal elements δ1 , δ2 and
δ3 are constrained by the four inequalities shown in equation (2.119). We can
interpret this geometrically by first considering a real three-dimensional space
formed by triplets δ1 , δ2 , δ3 . If we normalize the Mueller matrix to have m11 =
1, the unit cube in this space then represents the set of candidate depolarisers,
as shown on the left-hand side of Figure 2.10. Equation (2.119) then represents
the equation of four planes in this space, and all physical depolarisers must be
inside the volume bounded by these planes. The shape of this region is shown
on the right in Figure 2.10. To illustrate the sometimes subtle implications of
these constraints, we show, in equation (2.120), three candidate depolarisers,
all of which seem to be reasonable suggestions:
1 1 1 1 1
1 1 −1 −1 δ1
1 −1 1 −1 . δ2 ≥ 0 (2.119)
1 −1 −1 1 δ3
1 0 0 0 1 0 0 0
0 0.5 0 0 0 0.5 0 0
[M1 ] =
0
[M2 ] =
0 0.4 0 0 0 0.4 0
0 0 0 0.3 0 0 0 −0.3
1 0 0 0
0 0.5 0 0
[M3 ] =
0
(2.120)
0 −0.4 0
0 0 0 −0.3
However, using the constraints of equation (2.119) we see that [M2 ] is not a
physical depolariser. Its corresponding coherency matrix has a negative eigen-
value, and hence no physical system can generate such a depolarisation process.
Note, however, that [M1 ] and [M3 ] are both located inside the feasible volume
and therefore correspond to realizable depolarisers.
The constraints on forward scattering and backscattering matrices in equa-
tions (2.116) and (2.118) can now also be seen as special cases of this geometry.
In forward scattering, rotation symmetry constrains δ1 = δ2 = δ, and the
following special case is thus obtained:
1 1 1 1 1 1 + 2δ + ε ≥ 0
1 1 −1 −1 δ
. ≥ 0 ⇒ 1 − ε ≥ 0 (2.121)
1 −1 1 −1 δ 1−ε ≥0
1 −1 −1 1 ε 1 − 2δ + ε ≥ 0
We see that these inequalities give rise to the constraints shown in equation
(2.118). Equally in backscatter, the reciprocity theorem reduces [S] by relating
the crosspolarised elements, and this has the effect of reducing the rank of [T ]
from 4 to 3. Hence one eigenvalue (λ3 ) of [T ] must always be zero, even in
the presence of depolarisation. This then corresponds directly to the constraint
equation shown in equation (2.116). We see that the geometry of Figure 2.10 is
important in classifying depolarisers, and underlines the more fundamental role
played by [T ] rather than [M ] in understanding processes of wave scattering
and depolarisation.
We can further extend this analysis to include the case of a ‘general’ depo-
lariser proposed on the far right in equation (2.117), and obtained by a rotation
of the Poincaré sphere (Lu, 1996). The transformation of [T ] under such a
rotation can be derived using the mapping from rotations in a real three-
dimensional space to unitary transformations in a two-dimensional complex
space: the SU(2)–O3 homomorphism, as shown in equation (2.122).
SU (2) − O+
1 0
g −−−−−→[S] = [U2 ] [S] [U2 ]∗T
3
g =
0 [O3 ] (2.122)
∗T ∗
⇒ [T ] = [U4B ] [T ] [U4B ] [U4B ] = [U2 ] ⊗ [U2 ]
We see now, however, that this represents only a subset of possible depolarisers,
and a more general classification should be based on the full 4×4 unitary matrix
transformations of [T ] = [U4 ] [T ] [U4 ]∗T . This underlines the importance of
general unitary transformations in the analysis of polarisation phenomena.
2.6 Relating the Stokes/Mueller and coherency matrix formulations 113
0 0 0 1 0 0 −i 0
(2.126)
0 0 1 0 0 0 0 1
0 0 0 −i 0 0 i 0
Q3 =
1 0 0 0 Q4 = 0 −i 0 0
0 i 0 0 1 0 0 0
114 Depolarisation and scattering entropy
For example, we can now relate the scattered intensity in the Stokes/Mueller
formulation to that in the coherency matrix form by equating expressions for
scattered intensity and using [Q], as shown in equation (2.127):
w∗T [T ]w = g Tr [M ] g t ⇒ w∗T ⊗ wT t = g Tr ⊗ g Tt g
⇒ w∗T ⊗ wT [Q] = g Tr ⊗ g Tt (2.127)
⇒ w∗T ⊗ wT = g Tr ⊗ g Tt [Q]−1
Hence any choice of transmit and receive Stokes vectors g t and g r can be
transformed into an equivalent w formulation and vice versa. Given a complex w
we can convert it into a weight vector for the elements of [M ], a subset of which
can then be further decomposed into a direct product of Stokes vectors. Note,
however, that from a physical point of view the most important relationship
is equation (2.124). This relationship is not so evident in the Stokes/Mueller
formulation, and so we can in some sense consider the [T ] formulation to be
more fundamental.
In the previous chapters we considered in detail an algebraic formulation
of vector wave propagation and scattering, and have used this to establish a
detailed parameterization of depolarisation phenomena. It is now time to con-
sider examples of the application of this theory to physical models for surface
and volume scattering.
Depolarisation in
surface and volume
scattering
3
In Chapter 1 we introduced the scattering amplitude matrix [S] and its trans-
formation properties, and in Chapter 2 we saw how to extend these concepts
to include the presence of depolarisation. Here we derive examples of the
explicit form of the amplitude matrix and its coherency structure for two impor-
tant canonical problems: surface scattering and volume scattering. Details of
the underlying derivations are widely available in the existing literature, with
several excellent books covering the topics required (Born and Wolf, 1989;
Ishimaru, 1991; Jackson, 1999; Kong, 1985; Lakhtakia, 1989; Tsang, 1985;
Ulaby, 1986; van de Hulst, 1981; Mishchenko, 2000). Here our purpose is
mainly to concentrate on the polarimetric properties of these solutions, and so
we proceed by first summarising the key formulae and then move on to high-
light their important polarimetric structure in terms of polarised and depolarised
component decompositions. We also concentrate on those models that a) pro-
vide a fully polarimetric solution, including phase and amplitude effects, and b)
lead to simple analytical results without the need for numerical methods such
as finite difference and finite element. This will enable us to establish a few
suitable benchmarks for later comparison with numerical experiments or more
advanced modelling techniques. Inevitably this means that we concentrate on
low-frequency solutions, where analysis is tractable. Hence we start by con-
sidering scattering by electrically smooth surfaces and by volumes composed
of electrically small particles. However, these not only have direct relevance
to low-frequency radar applications (particularly at L and P bands), but also
provide a framework by which to judge more complicated high-frequency
solutions.
In order to calculate a general [S] matrix we must solve Maxwell’s differential
equations together with appropriate boundary conditions applied on the surface
of the object. From the former we postulate polarised plane wave solutions
and, by enforcing the latter, obtain a set of equations for the unknown co- and
crosspolarised scattering coefficients in a chosen coordinate system. These can
then be solved for two orthogonal incident polarisation states to obtain the four
complex elements of the scattering amplitude matrix [S]. Key to this process
is an understanding of the behaviour of vector EM fields across a boundary
E 1 H1 D1 B1
between two materials. n
There are four sets of boundary conditions arising from the set of four n ×E1 = n × E2 n. D1 = n. D2
Maxwell equations (see equation (1.1)). At an interface between two media, E 2 H2 D2 B2
n ×H1 = n × H2 n B1 = n B2
the two curl equations require continuity of the tangential components of E and
H , while the divergence equations require continuity of the normal components
of D and B. For a general boundary we then define the local normal vector n, as Fig. 3.1 Vector boundary conditions in elec-
shown in Figure 3.1. Note that of the four sets of boundary conditions only two tromagnetic scattering
116 Depolarisation in surface and volume scattering
From equation (1.145) we can then generate the S β daughter problem (without
yet calculating the detailed form of the matrix), which from the symmetry of
the plane must be equal to the original mother matrix S. The only solution
to this constraint is that b = c = 0, or the matrix has zero crosspolarisation.
Hence a smooth infinite plane interface causes zero crosspolarisation. This
leaves only the two diagonal copolar elements of the amplitude matrix. Note
that this is true for any scattering angle θ (not just specular) and for arbitrary
surface roughness, as long as the roughness profile is one-dimensional (like a
corrugated rough surface).
For each polarisation combination, such as VV, we have two unknowns:
the reflected and transmitted electric field components. However, by matching
Ei n Er n
Ei Er
Hi Hr
θi θr θi θr
ε1 θr ε1 θt
x
ε2 ε2
Et
Et
Ht
Fig. 3.2 Surface reflection and transmission z
for incident V (parallel, ‘p’ or TM) and H
(perpendicular, ‘s’ or TE) polarisations V or parallel (p) incident polarisation H or perpendicular (s) incident polarisation
3.1 Introduction to surface scattering 117
Hrs cos θr − His cos θi = −Hts cos θt ⇒ (Hrs − His ) cos θi = −Hts cos θt
8 s
cµo
Ei + Ers = Ets (3.2)
E= n H⇒
n1 (Ei − Ers ) cos θi = n2 Ets cos θt
s
The first equation follows from the continuity of tangential electric field for ‘s’
polarisation. The second requires some manipulations, projecting the H field
onto the x axis and using the relationship between E and H components of a
plane wave from Maxwell’s equations. Note that c is the velocity of light in
free space.
Similarly, for the V or ‘p’ polarised problem, when the E field is polarised in
the plane of incidence, we can obtain two equations by enforcing continuity of
tangential E and H , as shown in equation (3.3):
8 p p p
n1 (Ei + Er ) = n2 Et
p p p (3.3)
(Ei − Er ) cos θi = Et cos θt
Both solutions for reflection and transmission coefficients can then be combined
into the reflection and transmission matrices, as shown in equation (3.4). Note
that for simplicity of notation we assume non-magnetic materials where the
relative permeability µr = 1.
This latter form is more commonly used in radar scattering problems. Further-
more, the incident wave is often in free space, and so we can set ε1 = 1 to
further simplify the equations. Note that these equations apply even for com-
plex permittivity ε2 , such as occurs for microwave reflection from wet soils and
sea ice, for example. In this case the wave is changed in both amplitude and
phase on reflection, but equation (3.5) is still valid. We now turn to consider
the properties of such lossy material in more detail.
This shows that the field inside the material decays exponentially from the
surface (and to first order the decay rate is proportional to frequency, so increas-
ing at higher frequencies) with coefficient κz , and that the equiamplitude and
equiphase contours are no longer the same. This is called an inhomogeneous
plane wave, and is characteristic of propagation into lossy material (those with
a loss tangent greater than zero). This idea of projecting the wave vector β o
onto surface components is also an important one, and will be used extensively
in our analysis of interferometry in Chapter 5.
Note that wave attenuation is more often expressed in terms of a power
attenuation coefficient σe with units of decibels/meter (dB/m). Equation (3.11)
can then be used for the conversion from κz .
20
σedB = κz = 8.686κz dB/m
ln(10)
(3.11)
ln(10) dB
κz = σ = 0.115σedB m−1
20 e
A key concept in such material is the penetration or skin depth, defined as the
depth at which the signal is attenuated to exp(−1) (−8.686 dB). This can be
considered a typical penetration depth into the material. In general, therefore,
the penetration depth is given in terms of the complex refractive index, as shown
in equation (3.12):
√
1 2
δ= = " 0.5 (3.12)
κz ε 2
β0 ε − sin2 θi 1+ −1
(ε − sin2 θi )2
This is the most general expression, but in the special case of very good con-
ductors with conductivity σ this expression takes the more familiar form of the
skin depth shown in equation (3.13):
"
2
δ≈ (3.13)
ωµ0 σ
As an important example of a lossy dielectric in microwave remote sensing, and
one that illustrates the underlying complexity of modelling dielectric constant,
we now summarize the main dielectric properties of soil.
120 Depolarisation in surface and volume scattering
ρb x y x
εsoil
x
=1+ εss − 1 + mv εfw − mv (3.14)
ρss
The imaginary component arises from the final term, defined from a Debye
model of water absorption as follows:
74.1 iσδ
εfw = 4.9 + − (3.15)
1 + i ff0 2π ε0 f
where σδ is the soil conductivity, and f0 the relaxation frequency (which varies
from 9 GHz at 0◦ C to 19 GHz at 25◦ C). Note that one of the main contributions to
soil conductivity is its salinity ‘sa ’ in ppthou with a typical derived relationship
3.1 Introduction to surface scattering 121
f (z) = e −κ z
z
Reflection/transmission coefficients
0.6 T VV
0.4
0.2
0
–0.2
–0.4
–0.6
–0.8
–1
Fig. 3.4 Fresnel reflection and transmission 0 10 20 30 40 50 60 70 80 90
coefficients for εr = 4 Angle of incidence (degrees)
in the dipole radiation pattern. Using Snell’s law we can then obtain an
expression for this Brewster angle θB , as shown in equation (3.18):
π εr2
θr + θ t = ⇒ n1 sin θi = n2 cos θi ⇒ tan θB = (3.18)
2 εr1
Finally, we note that at normal incidence (θ = 0◦ ) both HH and VV have
the same magnitude but a 180-degree phase difference (see below for a
discussion of phase and coordinates), while at grazing incidence (θ =
90◦ ) both |HH| and |VV| have unit magnitude and zero phase difference.
Hence there must be a switch in phase for one of the polarisation channels
at some angle of incidence.
• The phase of the HH reflected signal is always 180 degrees, for ε2 > ε1 .
From Figure 3.2 we see that this means that the perpendicular component
(H) undergoes a 180-degree phase shift on reflection, as discussed in the
limit of perfect conductors. This is a direct consequence of the boundary
conditions and coordinate system employed. The VV channel phase has
a more complicated structure. We first note that the VV channel has zero
phase for angles of incidence less than the Brewster angle and a 180-
degree phase shift thereafter. This can be explained by inspection of the
coordinate systems employed in Figure 3.2, as follows. From Figure 3.2
we see that the ‘zero’ VV phase is actually a consequence of the same
180◦ phase change on surface reflection that is seen in the HH component
V
(consider the case as θi tends to zero). The reason it appears as a zero
H phase is that the surface coordinate definition of in-phase for polarisation
V components parallel to the plane of incidence is that their z components
are parallel.
H To further illustrate this coordinate-dependent phase notation, we show in
Incident wave Reflected wave
Figure 3.5 the coordinates used for a sensor-oriented (BSA) description of
Fig. 3.5 Sensor (BSA) coordinate system normal incidence reflection. Note that by using the same logic as above, the
for describing normal incidence Fresnel HH reflection coefficient will now be positive (with the 180-degree surface
reflection phase shift now matching the coordinate shift of the BSA system), while VV
3.1 Introduction to surface scattering 123
Dihedral backscatter
0.8
0.6
0.4
Reflection coefficients
0.2
–0.2
–0.4 HH for n = 5
VV for n = 5
HH for n = 2
–0.6 VV for n = 2
Making the usual coordinate corrections, the scattering matrix in the sensor
(BSA) system now has the form shown in equation (3.24):
S 0 R R 0
[S]BSA = ⊥ = A HHB HHB (3.24)
0 S|| 0 −RVVB RVVA
We see now that the polarimetric phase behaviour becomes more complicated
when we start to consider scattering by dielectric media, and this will motivate
the development of the ‘alpha parameter’ description of scattering in Chapter
4. First, however, we turn to another important surface scattering model, where
polarisation behaviour can be fully quantified.
ks = 0 ks < 0.3
smooth slightly rough
is limited, so that the rms roughness of the surface s is small compared to the
wavelength β = 2π/λ; that is, βs is a small quantity. In practice, βs < 0.3 is a
suitable condition, and so this constitutes a good low-frequency approximation,
typically valid at L (1.2 GHz) and P (450 MHz) bands for remote sensing of nat-
ural surfaces. Figure 3.9 shows a schematic representation of this small-scale
roughness. This is also called the vector small perturbation or Bragg scattering
model (after William Lawrence Bragg (1890–1971), who first formulated the
boundary condition required for such scattering). From our point of view the
most important observation is that the perturbed boundary conditions lead to
different polarisation coefficients from those obtained for the Fresnel case. The
scattered field from an arbitrary rough surface characterized by a height func-
tion z(x,y), illuminated by a wavelength long enough for this small perturbation
assumption to apply, is given as shown in equation (3.26) (Ulaby, 1986):
s
Epq = i 2β cos θ Bpq Ẑ (βX + β sin θ, βY )
(1 − εr ) cos φs
BHH =
cos θs + εr − sin2 θs cos θ + εr − sin2 θ
(1 − εr ) εr sin θ sin θs − εr − sin2 θs εr − sin2 θ cos φs
BVV =
εr cos θs + εr − sin2 θs εr cos θ + εr − sin2 θ
⇒
− (1 − εr ) εr − sin2 θ sin φs
BHV =
θ + ε − 2
θ εr cos θ + εr − sin2 θ
cos s r sin s
(1 − εr ) εr − sin2 θs sin φs
BVH =
ε cos θ + ε − sin2 θ cos θ + ε − sin2 θ
r s r s r
(3.26)
where
1
Ẑ(βx , βy ) = z(x, y)e−i(βx x+βy y) dxdy
2π
∞
For the special case of backscatter (θs = θ and φs = π ) this yields the following
important simplification:
s
Epq = i 2β cos θ Bpq Ẑ (2β sin θ)
cos θ − εr − sin2 θ
B = B⊥ = = RHH
HH
θ + ε − 2
θ
cos r sin
0 1
⇒ (εr − 1) sin2 θ − εr (1 + sin2 θ ) (3.28)
BVV = B|| = = RVV
2
ε cos θ + ε − sin 2
θ
r r
BHV = BVH = 0
where Ẑ is the Fourier transform of the surface profile. For backscattered power
this becomes the Fourier transform of the surface correlation coefficient or the
normalized roughness spectrum. Hence, to first order the backscatter depends
only a particular frequency component of the surface roughness spectrum
(Bragg scattering). But what of the polarisation properties of such scattering?
To proceed we make the following observations.
• The HH component equals the HH Fresnel equation, but the VV com-
ponent is very different from that predicted for Fresnel reflection. In
particular there is no longer a Brewster angle effect.
• The BSA scattering matrix for Bragg surface backscatter then has the
following form:
S⊥ 0 BHH 0
[S]BSA = = −i2β cos θ Ẑ (3.29)
0 S|| 0 BVV
Figure 3.10 shows the variation of the polarisation coefficients BHH and BVV
with angle of incidence for backscatter from a surface with εr = 4. Compare the
results with those for Fresnel reflection in Figure 3.4. Note that in this case the
B coefficients are not reflection coefficients. The actual level of backscattered
5
Bragg coefficients
0
0 10 20 30 40 50 60 70 80 90 Fig. 3.10 Bragg coefficients for backscatter
Angle of incidence (degrees) from a rough surface with for εr = 4
128 Depolarisation in surface and volume scattering
signal depends on the component of the Fourier transform of the surface (satis-
fying the Bragg condition). Note especially the absence of a Brewster angle in
the VV channel, and also note that the magnitude of VV is always greater than
or equal to HH—the opposite trend to that observed for Fresnel reflection.
As a final observation we note that the effect of surface roughness is contained
in the Fourier transform component Ẑ, which is common to all polarisation
channels. Therefore, if we take a ratio of scattered signals, such as HH/VV, then
the effects of roughness will cancel, and the ratio depends only on the angle
of incidence and dielectric constant of the surface. This provides an impor-
tant stimulus to using polarisation diversity for surface parameter retrieval in
remote sensing. For example, we note that as the angle of incidence increases
so the difference between HH and VV increases, and that this ratio is a func-
tion of dielectric constant and hence an indicator of soil moisture or salinity.
One way to represent these variations is by using the scattering alpha parame-
ter, discussed in Chapter 4. According to this representation we can represent
the scattering ratio in a rotation invariant manner by forming the following
function:
BHH − BVV π
αb = tan −1 0≤α≤ (3.30)
B + B 2
HH VV
Figure 3.11 shows how alpha changes with angle of incidence for three cases
of dielectric constant, εr = 2, εr = 9, and εr = 81. We note that the dynamic
range in alpha available to distinguish between wet and dry surfaces increases
with angle of incidence. By taking the difference between the curves we obtain
an estimate of this dynamic range, as shown in Figure 3.12. This can be used,
for example, to assess the polarimetric calibration requirements of a sensor in
order to be able to distinguish wet and dry surfaces. We note that for angles
of incidence less than 25 degrees (typical of spaceborne radar systems) the
available dynamic range is limited to less than 5 degrees of alpha variation.
40 n = 1.41
n=3
n=9
35
30
25
Alpha
20
15
10
20
15
10
cos θ − εA − sin2 θ
0
F = e−2βs cos θ → [R] = F cos θ + εA − sin2 θ
εA cos θ − εA − sin2 θ
0
εA cos θ + εA − sin2 θ
(3.31)
Here s is the surface rms roughness and β the free space wavenumber 2π/λ.
Figure 3.14 illustrates how the power attenuation factor varies for typical rough-
ness variations (2, 4 and 6 cm rms) at L-band frequency (1.3 GHz). There are
130 Depolarisation in surface and volume scattering
–15
–20
–25
80
70
75
60
80
Dielectric constant A
65 60
50
70
40
75
30
20 75
65
70 75
60
10
70 70
55 65 65 65
Fig. 3.15 Dihedral alpha parameter for a 60 60 60
22.5-degree angle of incidence and dielectric 10 20 30 40 50 60 70 80
variations in the two surfaces Dielectric constant B
three important consequences of this result for polarimetric studies. The first is
that in the presence of roughness the dihedral return, even though it is based
on a strong specular scattering effect, can be attenuated to a level where the
direct surface return can be dominant. Thus in practice we must consider the
possibility that both phenomena may be present at the same time. This will lead
us to develop so-called decomposition theorems for the interpretation of radar
scattering (see Chapter 4).
The second key idea is that any ratio of elements of the scattering matrix
is unchanged by the presence of surface roughness. As in the case of Bragg
scattering, we can then calculate the scattering alpha parameter (from equation
(3.30)) as a function of angle of incidence and dielectric constants of the two
surfaces involved. Figures 3.15 and 3.16 show examples of the dynamic range
of alpha for a shallow angle θ = 22.5◦ and a steeper angle θ = 45◦ . We note
the following key points:
3.1 Introduction to surface scattering 131
75
35
40 30
70
50
55
45
60
70
65
60
Dielectric constant A
50
30
40 35
70
50
55
40
45
60
65
30
20
30
40 35
70
10
50
55
45
60
65
70 Fig. 3.16 Dihedral alpha parameter for a
10 20 30 40 50 60 70 80 45-degree angle of incidence and dielectric
Dielectric constant B variations in the two surfaces
1. The shallow angle alpha shows little dependence on the first dielectric
component (from the horizontal surface) but is directly related to the
second (from the vertical), due to its effective higher angle of incidence.
2. The steeper angle alpha shows more dependence on both dielectric con-
stants (horizontal and vertical surface reflections) and hence, for any
given alpha value, we can only ever obtain an estimate of the product of
the two dielectric constants.
3. The dynamic range in alpha is much larger, even for the shallow angle,
than that obtained for direct Bragg backscattering (Figure 3.12).
4. The alpha angles are nearly always greater than π/4, as expected for
dihedral scattering, but can go lower than this for very dry materials and
angles of incidence where the Brewster angle effect causes a switch in
the scattered phase.
Finally, we note that one other advantage of the alpha parameter formulation
is its invariance to rotations in the plane of polarisation. Such rotations can
occur, for example, when we consider scattering from sloped surfaces. In such
cases, although the alpha parameter can remain unchanged, crosspolarisation
is introduced to the scattering vector. We now turn to consider slope effects on
the polarisation properties of surface scattered waves.
V z
H
β n = n1 x + n2 y + n3 z
Incident wave f
coordinates
n×β
t= (3.32)
n × β
With this local coordinate system in place we must now modify the scattering
matrix in two ways:
• The angle of incidence θ used for evaluation of the Bragg or Fresnel
coefficients is no longer simply φ̄ in Figure 3.17, but is now defined from
the inner product between the surface normal and incident wave vector,
as shown in equation (3.33):
The combined effect of these two angles is to modify the scattering matrix of
the facet to that shown in equation (3.35), which we see leads to the generation
of crosspolarisation from the facet:
cos χ sin χ S (θ) 0 cos χ − sin χ
S n = . ⊥ . (3.35)
− sin χ cos χ 0 S|| (θ) sin χ cos χ
3.2 Surface depolarisation 133
s
Epq = i 2β cos θ Bpq Ẑ (2β sin θ)
cos θ − εr − sin2 θ
BHH = B⊥ =
cos θ − + εr − sin2 θ
0 1
⇒ (εr − 1) sin2 θ − εr (1 + sin2 θ )
BVV = B|| =
2
εr cos θ + εr − sin2 θ
BHV = BVH = 0
(3.36)
a b 0
βs1
−→ [TS ] = ms b∗ c 0
0 0 0
λ1 0 0 ∗T
= u1 u2 u3 . 0 0 0 . u 1 u2 u3
0 0 0
⇒ Hs = 0
One way to do this is to assume that the major perturbation to equation (3.36)
arises from micro-variations in surface slope as roughness increases. There is a
single parameter in the scattering vector that is influenced strongly by changes
in slope: χ , as shown in equation (3.38) (see also equation (3.35)). In this case
we can therefore propose that depolarisation is primarily caused by integration
over a distribution of χ . Even if the exact details of this distribution (p(χ )) are
3.2 Surface depolarisation 135
unknown, the consequences for the form of the resulting coherency matrix are
important, as shown in equation (3.38):
1 0 0 a b 0 1 0 0
[TS ] = ms 0 cos 2χ sin 2χ b∗ c 0 . 0 cos 2χ − sin 2χ
0 − sin 2χcos 2χ 0 0 0 0 sin 2χ cos 2χ
τ= cos 2χ.p (χ) d χ
a bτ 0
→ [TS ] = ms b∗ τ cδ 0 ⇒ Hs > 0
δ= cos 2χ.p (χ) d χ
2 0 0 c(1 − δ)
(3.38)
+ , + ,
t22 − t33 |SHH − SVV |2 − 4 |SHV |2
R (δ) = =+ , + , = 2δ − 1 0≤R≤1
t22 + t33 |SHH − SVV |2 + 4 |SHV |2
(3.39)
+ , + ,
t22 + t33 |SHH − SVV |2 + 4 |SHV |2
M (θ, εr ) = = + , = tan2 αb (3.40)
t11 |SHH + SVV |2
+ ∗
,
sLL sRR
γLLRR = + ,+ ,
∗
sLL sLL ∗
sRR sRR
+ ,
(p1 + ip2 ) (−p1 + ip2 )∗
= + ,+ ,
(p1 + ip2 ) (p1 + ip2 )∗ (−p1 + ip2 ) (−p1 + ip2 )∗
8 (3.41)
p1 = SHH − SVV
p2 = 2SHV
+ 2, + 2,
|p2 | − |p1 |
−−−−−−−−−−−→ γLLRR = + 2 , + 2 , = −R
reflction symmetry
|p1 | + |p2 |
Note that in the presence of a surface with mean azimuth slope χ , the X-Bragg
coherency matrix has the modified form shown in equation (3.42):
1 0 0 a bτ 0 1 0 0
[TS ] = ms 0 cos 2χ sin 2χ b τ cδ
∗ 0 . 0 cos 2χ − sin 2χ
0 − sin 2χ cos 2χ 0 0 c(1 − δ) 0 sin 2χ cos 2χ
(3.42)
from which it follows that the revised diagonal elements of [T ] have the form
shown in equation (3.43):
t11 = a
t22 = c(δ cos 4χ + sin2 2χ ) (3.43)
t33 = c(cos 2χ − δ cos 4χ )
2
which implies that while the material index M remains invariant to azimuth
slope, the roughness indicator R varies as cos 4χ ; therefore, as mean slope
increases so the apparent roughness increases. We note, however, that by
estimating the LLRR coherence in both amplitude and phase we can com-
pensate for such slope effects (χ from the mean phase of the coherence, see
Chapter 4) and obtain a basis invariant estimate of the roughness (R from
the coherence amplitude, corrected for χ ). In this way we see that circu-
lar polarisations and their coherence are of great utility in surface scattering
studies.
As this approach is based on the low-frequency Bragg model but provides
a link to high-frequency depolarisation effects it is termed the extended or X-
Bragg surface scattering model. We can formulate this model for a geometrical
representation of its depolarisation properties in the backscatter entropy/alpha
plane as follows. If we assume for simplicity a uniform distribution of slopes
for p(χ ) with width , then the X-Bragg model takes on the explicit form
shown in equation (3.44), where the coefficients a, b and c are given in terms
3.2 Surface depolarisation 137
cos θ − εr − sin2 θ
BHH =
a = |BHH + BVV |2
cos θ +
εr − sin2 θ
∗
b = (BHH + BVV ) (BHH − BVV ) ⇒ 0 1
(ε − 1) sin 2
θ − ε (1 + sin 2
θ )
c = |BHH − BVV |2
r r
BVV =
2
εr cos θ + εr − sin2 θ
We can now consider the loci of H/α points generated when varies from 0
to π for a fixed angle of incidence and varying dielectric constant εr . Such loci
are shown in Figure 3.18, for θ varying from 20 to 50 degrees. We have chosen
a high dielectric constant (εr = 75), so these loci represent the upper bounds on
α for each angle of incidence. Again we note that for each angle of incidence
the alpha angle changes only little with roughness changes, as implied by the M
parameter in equation (3.40). We note that as the angle of incidence increases,
X-Bragg model
90
80
70
60
50
Alpha
40
30
20
u = 50º
10
u = 20º u = 40º
u = 30º
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Fig. 3.18 Geometrical representation of the
Entropy X-Bragg model in the entropy/alpha plane
138 Depolarisation in surface and volume scattering
z Reflected beam
θr dωr
Incident beam
dωi θi
φr y
φi
Note that the cos θr factor appears because the BRDF is defined per-unit pro-
jected area. As a consequence we note that f (θi , θr , φ) is related to σs , the
scattering cross-section per-unit illuminated area as shown in equation (3.47):
σs
f = (3.47)
cos θi cos θr
As an extreme case, a Lambertian surface is defined so that the outgoing power
per unit solid angle, per-unit projected area is independent of angle. Hence
f = 1/π for a Lambertian surface; and it also acts as an ideal depolariser
with scattering entropy H = 1, so that it has a coherency matrix of the form
shown in equation (3.48). Natural surfaces, however, do not show this type of
depolarising behaviour, and generally have entropy less than 1 (see Chapter 9).
1 0 0 0
0 1 0 0
[Tlambertian ] =
0 0 1 0
(3.48)
0 0 0 1
Having established these basic properties of the BRDF, we now turn to consider
the microfacet model, the physics of which is rooted in geometrical optics. In
its basic form it postulates scattering from a collection of randomly oriented
microfacets (see Figure 3.6) comprising the rough surface. The geometrical
distribution of these facets, together with a local reflection law—usually the
Fresnel equations and Snell’s law—then defines the structure of the model.
One key feature of the microfacet model for our purposes is that it reduces the
complexity of the angular dependencies to a familiar similarity transformation
of the single facet scattering matrix, as we now demonstrate.
140 Depolarisation in surface and volume scattering
The starting point is to convert the three primary scattering angles into four
secondary geometrical parameters. The first is an angle , which is the local
angle of incidence (and angle of reflection) onto the facet, and is important in
determining the physical properties of scattering by a facet. This angle is defined
from the inner product of local surface normal unit vector n and direction vector
of the incident wave r i , as shown in equation (3.50):
cos = n.r i ⇒ cos 2β = cos θi cos θr + sin θr sin θr cos (φi − φr ) (3.50)
From the local reflection law we can also relate to the reflected wave vector
r r , as shown in equation (3.51):
r i + r r = 2 cos n (3.51)
This leads us to the second important angle, θ, being the polar angle of the local
normal, as shown in equation (3.52):
cos θi + cos θr
cos θ = (3.52)
2 cos
Each microfacet is then characterized by its normal n, which is further assumed
to be symmetrically distributed about the z axis according to some given prob-
ability function. A commonly used example is the Gaussian surface height
distribution, which in normalized form can be written in terms of the local
surface slope (tan θ ) and the slope variance s2 , as shown in equation (3.53):
1 − tan2 θ
p(θ )d ω = 1 ⇒p(θ) =
2
exp (3.53)
2π s2 cos3 θ 2s2
When used in the scalar microfacet model this leads to the following form for
the function f :
1 − tan2 θ
f = exp R() (3.54)
8π s2 cos4 θ cos θi cos θr 2s2
where R() is a Fresnel reflectivity function for the selected scalar channel.
Note that this form of the BRDF satisfies both the normalization constraint in
equation (3.46) and the reciprocity relation in equation (3.49). Note, impor-
tantly, that this is a deterministic model. The parameters of the right-hand side
are all fixed for a given geometry and surface roughness. The geometry and
roughness effects then act as a multiplier to the scattering coefficient—itself a
deterministic quantity. Importantly this enables us to extend this model to a fully
polarimetric version by replacing R(β) by a vector formulation, as follows.
The main complication in formulating a vector BRDF is in matching the local
coordinates. There are four such systems to consider. The first is the system
defined by the incident wave direction r i and the mean surface normal in the z
direction. The second is then defined by the local surface normal n and incident
wave direction. This second is rotated about the incident direction by an angle
ηI , as shown in equation (3.55). The third coordinate system is defined by the
local surface normal n and the reflected wave direction. This has finally to be
related to the fourth: namely, that formed by the reflected wave direction and
the mean normal z. This is related to the third by a rotation about the reflected
direction by an angle ηr , as shown in equation (3.55).
3.2 Surface depolarisation 141
Having determined the form of the [S] matrix, we can now vectorize to obtain
the corresponding scattering vector k, which can be expressed compactly as a
unitary transformation, as shown in equation (3.56):
cos ηi − ηr 0 0 sin ηi − ηr
√ 0 cos ηi + ηr sin ηi + ηr 0
kM = ap
0 − sin ηi + ηr cos ηi + ηr 0
− sin ηi − ηr 0 0 cos η i − ηr
Rss () + Rpp ()
Rss () − Rpp ()
×
(3.56)
0
0
The most important consequence of this model is that all the surface rough-
ness effects are contained in the scalar multiplier ap . Hence, like the Bragg
model for backscatter, there is zero depolarisation under this model—one of
the most commonly used for bistatic surface scattering. Consequently the scat-
tering entropy is zero and the scattering coherency matrix is rank 1, as shown
in equation (3.57):
[T ] = k M .k ∗T
M (3.57)
There is again a clear discrepancy between the model predictions and observed
depolarisation behaviour of real surfaces (see Chapter 9). In particular, as we
found with the backscatter Bragg model, there is no way to parameterize smooth
change from a zero entropy polariser to Lambertian depolariser. To overcome
this a heuristic compromise is often sought by adding a Lambertian component
to the microfacet model with parameter ad , often determined by fitting the
model to experimental data. In this case, depolarisation under bistatic surface
142 Depolarisation in surface and volume scattering
We see that the penetration can be expressed as a scale factor times the free
space wavelength, and for volume scattering to be possible we require this
factor to be much greater than unity. We see that the factor depends only on the
imaginary part of the square root of the effective dielectric constant, but often
a further simplification is made, since for most materials in microwave remote
sensing the imaginary part is much smaller than the real. We therefore have the
following common approximation:
√
ε λo ε
< 0.1 → δp ≈ (3.60)
ε 2π ε
To illustrate this relation, we now consider five important application areas in
microwave remote sensing: namely, scattering from water surfaces, ice, snow,
soil and vegetation (Ulaby, 1982, 1986; Dobson, 1985).
3.3 Introduction to volume scattering 143
Water
Pure water has a high dielectric constant (in both real and imaginary parts;
see equation (3.15)) across the microwave spectrum, and this severely limits
the penetration depth of microwaves. Radar scattering by water can therefore
be considered a surface-dominated problem, albeit often a dynamic one with
time varying and spatially inhomogeneous roughness due to interacting wind
and current systems. Note, however, that breaking waves pose an important
example where volume scattering can occur in the water context, as these
generate a turbulent air/water mixture with a relatively low effective dielectric
constant and where wave penetration can be significant and volume scattering
effects be observed.
Snow
Dry snow is a mixture of ice and air—both very low-loss materials—and so
penetration can occur to hundreds of wavelengths, with a small overall dielec-
tric constant lying in the range 1.5–2. The imaginary part of the dielectric
constant is, however, very sensitive to changes in water content, and so wet
snow (which has dielectric constant always less than 4) shows a greater variety
of penetration depths. In general, therefore, snow can behave as a surface or
volume scatterer (or both), depending on environmental conditions, viewing
geometry, and operating frequency.
Soils
As discussed in equation (3.14), soil is a composite material, the dielectric
constant of which depends on a large number of parameters. It is therefore
difficult to derive general conclusions about soil penetration. However, across
a large part of the microwave spectrum, penetration depth depends primarily
on soil moisture content, and for most mid-latitude soil types penetration is
therefore seldom more than a wavelength or so. An important exception must
be made for hyper arid sand/soil environments, where penetrations of many
wavelengths can occur. Soil is therefore similar to snow in that both surface
and volume scattering can occur, although in most mid-latitude agricultural
areas the assumption is often made of pure surface scattering.
Vegetation
Vegetation scattering is perhaps the most complex of all, and simple models for
penetration depth are often of limited use. Indeed, vegetation scattering pro-
vides an important focus for the development of polarimetric interferometry in
Chapter 6, which allows us, amongst other things, to estimate the penetration
144 Depolarisation in surface and volume scattering
depth into vegetation. It is sufficient here to point out that vegetation is often a
strong volume scattering environment, combining multiple wavelengths of pen-
etration with the presence of strong dielectric discontinuities (moist branches,
twigs, leaves, and so on), which are often larger than a wavelength, and act to
scatter the waves in a complex manner.
Surface and volume scattering effects are often discriminated on the basis
of their dependence on angle of incidence. Surface scattering generally shows
a fall-off with increasing angle, while volume effects vary only slightly with
change of angle. However, in this text we are concerned primarily with the
polarisation properties of the two phenomena. In preparation for a discussion
of the polarisation properties of volume effects, we now consider scattering by
an individual particle. In the next section we use these results with averaging
to analyse the depolarisation behaviour of random volume scattering.
z
x1 Particle
frame
x2 = x3
can now model variations in shape from prolate (needle-like) through spherical
and into oblate (disc-like). This provides a more realistic variation of shape
for modelling in many natural media. Again the solution can be established
from Maxwell’s equations by expanding fields in terms of vector spheroidal
wave functions and matching coefficients across the boundary. However, this
solution is not nearly as convenient as the Mie series, for the following reasons
(Mishchenko, 2000, 2006, 2007):
• The computation of vector spheroidal functions is a complicated math-
ematical and numerical problem in itself, especially for absorbing
particles.
• The spheroidal functions are not orthogonal on the surface of the scatterer,
and the unknown expansion coefficients must therefore be solved using
an infinite set of algebraic equations, which in practice must be truncated
and solved numerically. Unfortunately, for large particles this system
becomes ill-conditioned, and care is needed in the numerical computation
of results.
As a result of these difficulties, several alternative numerical methods have
been developed based, for example, on finite element, finite difference, and T-
matrix approximations. These provide more robust general-purpose numerical
solutions, but are too complicated in structure for simple analytical studies.
To avoid these problems we illustrate the important polarisation physics by
considering the case of Rayleigh scattering by spheroids. Lord Rayleigh (1842–
1919) first derived an approximation for scattering in the small particle limit
when βa << 1 by assuming that the fields in and near the particle are the
same as those derived from electrostatics, and that the internal field inside the
particle is homogeneous. This approximation leads to a fully analytic solution,
even for chiral particles, which as we have seen in Chapter 1, case III, is a
case of special importance in polarimetric studies. We now consider the main
polarimetric features of this solution method.
With reference to Figure 3.20, we now assume that the particle dimensions
are small compared to the wavelength of the incident radiation so that the field
inside may be considered constant, and we can ignore any time delays for fields
to propagate across the particle. In quantitative terms this requires the following
condition:
|n| βa 1 (3.61)
where α ij are 3 × 3 polarizability tensors, which are all diagonal in the par-
ticle frame. The field radiated by elementary electric (p) and magnetic (m)
dipole moments can be obtained directly from Maxwell’s equations, as shown
in equations (1.10) and (1.11), and are given explicitly in equation (3.63):
0 1 exp (−iβ r)
o
E = ω2 µo (I − rr).p − ωβo r × m (3.63)
4π r
Here r is a unit vector in the direction of the scattered wave. Before we can use
this relationship with equation (3.62) we first must transform the coordinate
system from the particle into the laboratory frame. The latter can be defined as
shown in Figure 3.21. Without loss of generality, we consider the wave incident
along the +z direction and scattering in the y-z plane in the direction r, which
makes an angle with the z axis. We then wish to find the 2 × 2 scattering
amplitude matrix [S] for radiation polarised parallel (P) or perpendicular (S) to
the scattering plane, as shown in Figure 3.21. Using the r vector as defined in
Figure 3.21 together with equation (3.63), we can express the radiated field by
the particle as shown in equation (3.64):
2
1 1 β exp(−iβo r)
ES = p1 − √ (m3 sin ψ − m2 cos ψ) o
εo εo µo 4π r
2 (3.64)
1 1 β exp(−iβo r)
EP = (p2 cos ψ − p3 sin ψ) − √ m1 o
εo εo µo 4π r
r
Ψ
ES r inc = [0 0 1]
z r = [0 sin Ψ cosψ ]
EP
y
E = [1 0 0] E S = [1 0 0]
rinc ⇒ S
EP = [0 1 0] E P = [0 cosΨ -sinψ ]
EP
cos τ 0 − sin τ cos θ sin θ 0
[R] = 0 1 0 . − sin θ cos θ 0
sin τ 0 cos τ 0 0 1
cos τ cos θ cos τ sin θ − sin τ
= − sin θ cos θ 0 (3.65)
sin τ cos θ sin τ sin θ cos τ
Using this rotation matrix we can then express equation (3.62) in the laboratory
frame, as shown in equation (3.66):
inc
plab R−1 α ee R R−1 α em R E
= . (3.66)
mlab −R−1 α em R R−1 α mm R H inc
where now we can make use of the incident electric and magnetic fields for a
TEM wave in the laboratory frame directly as
T εo
E inc = ES EP 0 ⇒ H inc = −EP ES 0 (3.67)
µo
Evaluation of the elements of one block of equation (3.66) will illustrate the
general form of the solution, as shown in equation (3.68):
cos τ cos θ − sin θ sin τ cos θ ρ1 0 0 cos τ cos θ cos τ sin θ − sin τ
R−1 αR =
cos τ sin θ cos θ sin τ sin θ
. 0 ρ2 0
. − sin θ cos θ 0
− sin τ 0 cos τ 0 0 ρ2 sin τ cos θ sin τ sin θ cos τ
ρ2 + (ρ1 − ρ2 ) cos2 τ cos2 θ (ρ1 − ρ2 ) cos2 τ sin θ cos θ − (ρ1 − ρ2 ) cos τ sin τ cos θ
=
(ρ1 − ρ2 ) cos τ sin θ cos θ ρ2 + (ρ1 − ρ2 ) cos τ sin θ − (ρ1 − ρ2 ) cos τ sin τ sin θ
2 2 2
− (ρ1 − ρ2 ) cos τ sin τ cos θ − (ρ1 − ρ2 ) cos τ sin τ sin θ ρ1 sin τ + ρ2 cos τ
2 2
(3.68)
We now have all the elements we require for a solution. We start by setting the
incident field as S or P polarised in equation (3.67), and then use equation (3.66)
with (3.68) for each block element to obtain the p and m vectors. Finally we
use equation (3.64) to derive the S and P components of the scattered field. In
this way we can calculate the scattering matrix as a function of the three angles
, θ, τ and the four pairs of polarizability constants for the particle.
Rather than calculate the full set of equations, we consider instead two
important special cases that permit further simplification of the above equations.
148 Depolarisation in surface and volume scattering
From these equations the elements of the scattering matrix can then be directly
obtained as shown in equation (3.70):
SPP SSP
[S] =
SPS SSS
SPP = (AP − 1) sin θ cos τ (sin θ cos τ cos − sin τ sin ) + cos
SSP = (AP − 1) cos θ cos τ (sin θ cos τ cos − sin τ sin )
⇒
SPS = (AP − 1) sin θ cos θ cos2 τ
SSS = (AP − 1) cos2 θ cos2 τ +1
(3.70)
where we have defined the particle anisotropy Ap as a ratio of principal
polarizabilities, so that
ρ1
Ap = (3.71)
ρ2
For simplicity we have cancelled all factors common to the elements of [S]. Of
particular interest is the case of backscatter ( = 180◦ ), when the equations
further simplify to those shown in equation (3.72):
SPP = − (AP − 1) sin2 θ cos2 τ − 1
S SSP SSP = − (AP − 1) sin θ cos θ cos2 τ
[S] = PP ⇒ (3.72)
SPS SSS
S = (AP − 1) sin θ cos θ cos2 τ
PS
SSS = (AP − 1) cos2 θ cos2 τ +1
This scattering matrix is expressed in the wave coordinates, and if we convert
to the sensor (BSA) system, commonly used in radar studies, we obtain the
following solution:
SPP = (AP − 1) sin2 θ cos2 τ + 1
S SSP SSP = (AP − 1) sin θ cos θ cos2 τ
[S]BSA = PP ⇒ (3.73)
SPS SSS
S = (AP − 1) sin θ cos θ cos2 τ
PS
SSS = (AP − 1) cos2 θ cos2 τ +1
which we see is complex symmetric, as expected. Note that this matrix has a
straightforward eigenvalue expansion, as shown in equation (3.74):
cos θ sin θ 1 0 cos θ − sin θ
[S]BSA = . .
− sin θ cos θ 0 sin2 τ + AP cos2 τ sin θ cos θ
(3.74)
3.3 Introduction to volume scattering 149
εo (εr − 1) V L2 + 1
εr −1
ρi = ⇒ AP = (3.75)
1 + (εr − 1) Li L1 + 1
εr −1
1 − e2 1 1+e x22
e2 −1 + ln x1 > x2 = x3 e2 = 1 − (prolate)
2e 1 − e x12
L1 =
1+f2 1 x22
2 1 − arctan f x1 < x2 = x3 f2 = − 1 (oblate)
f f x12
1
L2 = L3 = (1 − L1 ) (3.76)
2
Figure 3.22 shows a plot of the relationship between particle shape r = x1 /x2
and the polarizability ratio for low and high values of dielectric constant. Hence
we see that Ap is a general indicator of particle shape, although it is bounded in
value by the dielectric constant to lie between limits given by equation (3.77)
15 er = 5
er = 50
10Log(Ap), Ap = Polarisability ratio
Ap = r
10
–5
–10
–15
(Ablitt, 2000):
1 εr + 1
< AP < (3.77)
εr 2
Note also that to a good approximation LL12 ≈ xx21 (the error is less than 0.04
in all cases), and hence for large εr equation (3.75) yields AP ≈ xx12 , which
then becomes a direct indicator of particle shape. This again demonstrates the
importance of SVD in the analysis of the scattering matrix. In this case the
singular vectors lead to estimation of the particle orientation angle and the ratio
of singular values to the projected shape and dielectric constant of the particle.
We can now evaluate this explicitly. The result is shown in equation (3.81):
SPP SSP
[S] =
SPS SSS
SPP = (AP − 1) sin θ cos τ (sin θ cos τ cos + sin τ sin ) + cos
+iκ (AP − 1) cos θ cos τ (sin θ cos τ − cos τ sin θ cos − sin τ sin )
SSP = (AP − 1) cos θ cos τ (sin θ cos τ cos + sin τ sin )
+iκ[(AP − 1) cos τ cos2 θ cos τ + cos τ sin2 θ cos + sin τ sin θ sin +(1+ cos )]
⇒
SPS = (AP − 1) sin θ cos θ cos2 τ
−iκ[(AP − 1) cos τ cos2 θ cos τ + cos τ sin2 θ cos − sin τ sin θ sin +(1+ cos )]
SSS = (AP − 1) cos2 θ cos2 τ +1
+iκ (AP − 1) cos θ cos τ (sin θ cos τ − cos τ sin θ cos − sin τ sin )
(3.81)
This result shows that the presence of a small amount of chirality in the particle
leads to a decomposition of the backscattering matrix into the sum of two
152 Depolarisation in surface and volume scattering
two complex functions for arbitrary sphere size and material. The solution is
expressed as an infinite series, with so-called Mie coefficients an and bn , as
shown in equation (3.85) (Hovenier, 2004; Ishimaru, 1991):
SHH () 0
[S] =
0 SVV ()
5
∞ 2n + 1
SVV () = {an πn (cos ) + bn τn (cos )}
n=1 n(n + 1)
⇒ (3.85)
5∞ 2n + 1
SHH () =
{bn πn (cos ) + an τn (cos )}
n=1 n(n + 1)
dPn (cos )
πn (cos ) =
d cos
(3.86)
d πn (cos )
τn (cos ) = cos πn (cos ) − sin2
d cos
where Pn (..) are the Legendre polynomials (see equation (5.49)). The Mie
coefficients an and bn depend on the (generally complex) refractive index and
size parameter x = βr, where r is the radius of the sphere. The computation of
these coefficients can be quite laborious, especially for large particles, and there
are now available several public domain codes for their numerical calculation.
An example dataset (courtesy of Michael Mishchenko of NASA Goddard) for
a sphere of size βr = 4 and refractive index n = 1.32 (equivalent to water/ice
at optical wavelengths) is shown in Figures 3.23–3.25. Figure 3.23 shows the
phase function—a rather ambiguous term, as it contains no relation to phase
between complex numbers. Rather, it is a plot of the 1,1 element of the Mueller
matrix [M ]. From equation (2.48) we see that this is just the total scattered power
(trace of coherency matrix). Figure 3.23 therefore shows how the total scattered
power varies with scattering angle. We note the enhanced scattering in the
101
M11
100
10–1
10–2
0 20 40 60 80 100 120 140 160 180 Fig. 3.23 Phase function for scattering by a
Scattering angle (degrees) large dielectric sphere (βr = 4, n = 1.32)
154 Depolarisation in surface and volume scattering
Amplitude (dB)
–1
–2
–3
–4
90 1
120 60
0.8
0.6
150 30
0.4
0.2
180 0
210 330
We have shown by example that even spherical scatterers can generate inter-
esting polarisation behaviour in bistatic scattering. When we extend this to
consider non-spherical particles such as spheroids and even more complicated
geometries, then the polarisation response changes even more (in particular for
backscatter, which is useful for radar studies). However, there is one issue we
have yet to consider: that such particles are usually not viewed alone, but in
clouds with some distribution over size, shape, orientation, and so on. It is then
of interest to see how much of this structure survives the averaging inherent in
such depolarisation processes. We now turn to consider such effects in volume
scattering.
(3.88)
We can then average this matrix over all possible angles θ and τ . The probability
distributions p(..) for a random distribution are defined as shown in equation
(3.89):
dθ cos τ d τ π π
p (θ) = −π ≤θ <π p (τ ) = − ≤τ < (3.89)
2π 2 2 2
where the eigenvalues take the explicit form shown in equation (3.91) (Boerner,
1992, Chapters I–4; Cloude, 1999; Ablitt, 2000).
π
1 2 π
λ1 = t11 = π
(4 + 4X cos2 τ + X 2 cos4 τ ) cos τ d θd τ
4π −2 −π
4 4
= 2 + X + X2
3 15
π (3.91)
1 2 π 2 2
λ2 = t22 = π
X cos τ cos 2θ cos τ d θd τ =
2 4 2
X
4π −2 −π 15
π
1 2 π 2 2
λ3 = t33 = π
X 2 cos4 τ sin2 2θ cos τ d θd τ = X
4π −2 −π 15
and hence the ratio of eigenvalues yields information about the shape and mate-
rial composition of the particles in the volume through the parameter Ap , where
Ap ≥ 0 (see equation (3.77)), as shown in equation (3.93):
√
λ2 A2 − 2AP + 1 1 (1 + 3Rλ ) ± 5Rλ (3 − Rλ )
0 ≤ Rλ = = P2 ≤ ⇒ AP =
λ1 2AP + 6AP + 7 2 1 − 2Rλ
(3.93)
Figure 3.26 shows how the eigenvalue ratio Rλ changes with change in shape
of parameter Ap . We notice three important points in the function. As Ap tends
to zero, so the ratio tends to 1/7 = 0.1429, being the limit for a cloud of oblate
spheroids. For Ap = 1 the ratio tends to zero and there is zero depolarisation
(a cloud of spheres). As Ap tends to infinity we see that the ratio tends to a
maximum value of 0.5 (dipole cloud), this being the most depolarising case.
If we seek to invert this relationship by measuring the eigenvalue ratio in
order to estimate mean particle shape, then we face an ambiguity issue as
shown in Figure 3.27. Here we see that for ratios less than 0.1429 we have both
a prolate and oblate solution for the same ratio. For ratios greater than 0.1429
and less than 0.5, we obtain a unique prolate solution. For ratios between 0.5
and 1 we violate the assumptions of this model, and must consider other types
of depolarisation or noise in the data.
A similar treatment can be used for scattering by a random cloud of chiral
(or handed) particles. In this case the scattering vector for an individual particle
3.4 Depolarisation in volume scattering 157
0.4
0.35
Eigenvalue ratio
0.3
0.25
0.2
0.15
0.1
0.05
0
–20 –15 –10 –5 0 5 10 15 20
Fig. 3.26 Eigenvalue ratio of coherency
10*log(Ap) matrix versus particle shape parameter Ap
104
103
102
101
Ap
100
10–1
–2
10
can be obtained from the scattering matrix (derived in equation (3.82)), and
written for rotation θ, as shown in equation (3.94):
1 0 0 2 + X cos2 τ
k = 0 cos 2θ − sin 2θ X cos2 τ
0 sin 2θ cos 2θ −2iκX cos2 τ
2 + X cos2 τ
= X cos2 τ (cos 2θ + 2iκ sin 2θ ) (3.94)
X cos2 τ (sin 2θ − 2iκ cos 2θ )
Working through a similar procedure as used for the non-chiral case, we find
the following result for the eigen-decomposition of the coherency matrix for a
158 Depolarisation in surface and volume scattering
2
λ1 = (2A2P + 6AP + 7)
15
2
λ2 = (1 + 4κ) (AP − 1)2
15
2
λ3 = (1 − 4κ) (AP − 1)2 (3.96)
15
There are two important points raised by this result. Firstly, the eigenvectors
are now complex; that is, they are no longer just the default Pauli scattering
mechanisms. This highlights the importance of using the eigenvectors for a
full characterization of the scattering process. For example, expanding equa-
tion (3.95) and using (3.96) we obtain the following non-diagonal form of the
coherency matrix for volume scattering from a random cloud of chiral particles:
2
2A + 6AP + 7 0 0
2 P
[T ] = 0 (AP − 1)2 −i4κ(AP − 1)2 (3.97)
15
0 i4κ(AP − 1)2 (AP − 1)2
Secondly, the presence of chirality has caused a split in the smaller eigenvalues.
Indeed, we note that the chirality parameter is directly related to the scattering
anisotropy (see equation (2.79)), defined as the normalized difference of minor
eigenvalues, as shown in equation (3.98):
λ2 − λ3
A= = 4κ (3.98)
λ2 + λ 3
dθ cos τ d τ π π
p (θ) = −≤θ < p (τ ) = − ≤τ < (3.99)
2 2 2 2
80
70
60
50
Alpha
40
30
20
10
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Fig. 3.28 Variation of entropy/alpha for a
Entropy cloud of dipoles
160 Depolarisation in surface and volume scattering
80
70
60
50 Ap = 20dB
Alpha
P
40
30 Ap = 6dB
20
Ap = 2dB Q
10
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
Fig. 3.29 Entropy alpha variation for clouds
of prolate and oblate particles Entropy
the prolate spheroids are much more sensitive to orientation effects, and display
a much wider range of depolarisers. We identify two points of interest. The point
P represents the most extreme depolariser: a random cloud of dipoles with
H = 0.9464, α = 45◦ . This point is often taken as a model for branch/needle-
dominated forest scattering in radar remote sensing applications. For example,
it is embedded in the Freeman–Durden decomposition approach (see Chapter
4). The point Q corresponds to a random cloud of oblate spheroids (H = 0.62,
α = 20◦ ). This is more typical of leaf-dominated vegetation scattering, and
shows less depolarisation than the prolate case.
t11 = z(1 − cos ψ) + (1 + cos ψ)2
t11 t12 0 0
t12 = (cos2 ψ − 1)
t ∗ t22 0 0
[T ] = f 12 ⇒ t22 = z(1 + cos ψ) + (1 + cos ψ)2
0 0 t33 0
t = z(1 + cos ψ)
0 0 0 t44 33
t44 = z(1 − cos ψ)
2A2P + 6AP + 7 2(AP − 1)2
f = z= (3.101)
30 2A2P + 6AP + 7
3.4 Depolarisation in volume scattering 161
λ4 = fz(1 + cos ψ)
(Note that the rank ordering of the eigenvalues depends on scattering angle.)
For backscatter we note that λ4 = 0 and the eigenvalues reduce to those shown
in equation (3.96). We can use these equations to highlight two important exam-
ples. In the first we consider bistatic scattering from a cloud of spheres (Ap = 1).
In this case, f = 0.5, z = 0, and the eigenvalues have the form shown in
equation (3.103):
90 2
120 60
1.5
150 1 30
180 0
210 330
240 300
Fig. 3.30 Bistatic scattering diagram for a
270 cloud of small spheres
162 Depolarisation in surface and volume scattering
0.5
0.2
0.1
0
0 20 40 60 80 100 120 140 160 180
Fig. 3.31 Bistatic variation of scattering
eigenvalues for a dipole cloud Scattering angle (degrees)
A2p
λ1,2 = 2 + cos2 ψ ± 1 − cos2 ψ + cos4 ψ
15
A2p
λ3 = (1 − cos ψ) (3.104)
15
A2p
λ4 = (1 + cos ψ)
15
Figure 3.31 shows how the eigenvalues (normalized so that λ = 1) vary with
scattering angle. Note how for backscatter and forward scattering the coherency
matrix has rank 3, as expected from reciprocity. The other interesting point
relates to 90-degree or lateral scattering. Here we see that the minor eigenvalues
are all equal, corresponding to a noise subspace of the coherency matrix. Hence
only at this angle can we give the scattering interpretation as a single [S] matrix
(the dominant eigenvector) with a noise background. At all other angles we see
anisotropy in the eigenvalue spectrum and hence structure to the depolarisation
process.
It is useful to plot the scattering entropy variation corresponding to
Figure 3.31. Figure 3.32 shows how bistatic entropy H varies with scatter-
ing angle. We note that nowhere is depolarisation complete (H = 1), and that
for forward scatter and backscatter the entropy is smallest and hence depolar-
isation at a minimum. The reason for this is the loss of one eigenvalue to the
reciprocity theorem, which, as we noted earlier, limits the level of depolarisation
achievable in backscatter. The maximum depolarisation occurs for the lateral
(90-degree) scattering case. This then represents the strongest depolariser so
far encountered. However, we have considered only single scattering depo-
larisation. When waves encounter multiple scattering—when we account for
3.4 Depolarisation in volume scattering 163
0.95
0.9
0.85
Scattering entropy
0.8
0.75
0.7
0.65
0.6
0.55
0.5
0 20 40 60 80 100 120 140 160 180
Fig. 3.32 Bistatic variation of scattering
Scattering angle (degrees) entropy for a dipole cloud
2 2
1 3 1 3
n n
Hence the coherency matrix for such a scenario is not simply given by the
sum of coherency matrices for the n individual particles, but involves mutual
interactions. These interactions give rise to additional sources of depolarisation,
as we now demonstrate.
where the subscripted S is for single scattering, L for the ladder terms, and C
for the cyclical terms. Concentrating on the new polarisation effects caused
by multiple scattering and their effect on the eigenvalue spectrum of [T ], we
further rewrite the coherency matrix in the form shown in equation (3.106),
where ‘MS’ now contains all the multiple scattering contributions:
We can shed some light on the structure of these new depolarising terms by
considering the multiple paths in Figure 3.33 and writing the multiple scattering
terms as a function of their scattering vectors, as shown in equation (3.107):
9 ∗T :
[TMS ] = k (1,n) + k (n,1) k (1,n) + k (n,1)
9 :
= k (1,n) k ∗T
(1,n) + k (n,1) k ∗T
(n,1) + k (1,n) k ∗T
(n,1) + k (n,1) k ∗T
(1,n)
9 : 9 :
= k (1,n) k ∗T ∗T ∗T ∗T
(1,n) + k (n,1) k (n,1) + k (1,n) k (n,1) + k (n,1) k (1,n)
We see that the ladder terms are formed from averages over conventional
coherency matrix factors. However, the cyclical terms involve a mixture of
3.4 Depolarisation in volume scattering 165
path and time-reversed path contributions. The calculation of these latter terms
is particularly difficult, but there is an interesting relationship between ladder
and cyclical terms in the exact backscatter direction. This link arises because
of the reciprocity theorem, which relates time-reversed paths as an apparent
exchange of transmitter and receiver positions. We have already seen in equa-
tion (1.147) that the two coherent [S] matrices and corresponding scattering
vectors for the two paths are related as shown in equation (3.108):
1 0 0 0
1 0 T 1 0 0 1 0 0
S(1,n) = S(n,1) ⇒ k (1,n) =
0 0 −1 0 k (n,1)
0 −1 0 −1
0 0 0 1
(3.108)
Note that this means that the ladder terms by themselves violate the reciprocity
theorem, and care must be exercised when using predictions from VRT for
exact backscatter calculations from random media. It is only when we add the
‘L’ and ‘C’ terms that we obtain a rank-3 coherency matrix in equation (3.106).
Turning now to an example of depolarisation by multiple scattering in a
direction other than backscatter, we consider forward incoherent scattering
by a random cloud of particles. From symmetry arguments, the Mueller and
coherency matrices for such a scenario must have the normalized form shown
in equation (3.110):
1 0 0 0 1 + ε + 2δ 0 0 0
0 δ 0 0 0 1−ε 0 0
[M ] =
0 0 δ 0 ⇔ [T ] =
0 0 1−ε 0
0 0 0 ε 0 0 0 1 + ε − 2δ
|ε| ≤ 1
⇒ |δ| ≤ 1 (3.110)
2δ − ε ≤ 1
We see that the diagonal elements must obey some restrictions in order that the
eigenvalue spectrum of [T ] be non-negative, and that the rotational subspace
eigenvalues be equal. We saw above that single scattering from such a cloud
would yield a rank-1 coherency matrix and zero depolarisation (ε = δ = 1).
However, conditions change when we allow multiple scattering. This problem,
when formulated for a cloud of spheres, can be determined analytically using
VRT (at least for the ladder terms contributions; and here we are assuming
that the coherent forward scattered wave is completely attenuated). It can be
expressed as a function of the order of multiple scattering n, with details in
166 Depolarisation in surface and volume scattering
Brosseau, 1994), and here we state the main result, shown in equation (3.111).
We show both the Mueller and corresponding coherency matrix forms of the
result as a function of n. (Note that in this case there is no dependence on
bistatic scattering angle, as we are restricting attention to forward scattering.)
For example, for n = 0 (single scattering) we obtain a Mueller matrix with the
form of the 4 × 4 identity matrix, and a rank-1 coherency matrix with three
zero eigenvalues. These correspond to forward scattering by a sphere matrix.
n
1 7
1+ 0 0 0
2 10 n
3 7
0 0 0
2 10 n
[M ] = ⇒ [T ]
3 7
0 0 0
2 10 n
3 1
0 0 0
2 2
2 n n
7 7 5
+ + 0 0 0
3 10 3 7 n n
2 7 1 5
3 0 + − 0 0
3 10 3 7 n n
=
2 2 7 1 5
0 0 + − 0
3 10 3 7 n n
2 7 5 5
0 0 0 + −
3 10 7 3
(3.111)
Figure 3.34 shows how the eigenvalues of this coherency matrix vary with
scattering order n. Note that for n = 0 (single scattering) the matrix has only
one non-zero eigenvalue, as expected, while as n increases so the normalized
eigenvalues all tend to 0.25; that is, to equality and maximum depolarisation
(H = 1). Figure 3.35 shows the corresponding multiple scattering entropy as
a function of increasing order. Note that the entropy approaches 1—a perfect
depolariser for orders n > 5 (corresponding to ε = δ = 0 in equation (3.110)).
A second interesting observation concerns the anisotropy of the eigenvalues.
Two of the minor eigenvalues are always equal, as expected from symmetry,
0.9
lambda 1
lambda 2
0.8
lambda 3
lambda 4
Relative eigen value level
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 2 4 6 8 10
Fig. 3.34 Variation of forward scattering
eigenvalues with scattering order n Scattering order
3.5 Simple physical models for volume scattering and propagation 167
0.9
0.8
Relative eigen value level
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 2 4 6 8 10
Fig. 3.35 Forward scattering entropy and
Scattering order anisotropy versus scattering order n
and so there is only one anisotropy, A34 , to consider. The variation of this
anisotropy with order is also shown in Figure 3.35.
Note how the anisotropy is high for low orders of multiple scattering (n > 0)
and decreases steadily with increasing order. This implies that although the
spheres depolarise, they do not do so equally for all incident polarisations. We
see from the form of the Mueller matrix that this anisotropy is due to ε < δ
for all scattering orders. This means that the multiple scattering depolarises
incident circular polarisations, with a Stokes vector of the form (1, 0, 0, ±1)
more than linear, with Stokes vector (1, ±1, 0, 0), with the difference between
the two decreasing with increasing scattering order.
!
Nv !
Nv
σv = σPi σe = QPi (3.112)
i=1 i=1
Here, σp and Qp are the single particle backscattering and extinction cross-
sections respectively, defined as the ratios of scattered and absorbed powers to
the incident power density, and both having units of m2 . For Rayleigh scattering,
when the particles are small compared to a wavelength these parameters can
be written explicitly in terms of the particle size and material composition, as
shown in equation (3.113):
λ2 1 − εr 2λ2 εr − 1
QP = QA + QS = Im (βb r)3 + (βb r)6
π εr + 2 3π εr + 2
8π 2 1 − εr 3
≈ QA = Im r (3.113)
λo εr + 2
64π 5 εr − 1 6
σP = r
λ4o εr + 2
Note that the assumption of identical particle size can be relaxed by including a
new function—the particle size distribution function p(r)—so that the expres-
sion for total extinction (and other parameters) takes on an integral form as
170 Depolarisation in surface and volume scattering
r2
While no simple forms exist for the size distribution in vegetation scattering, for
example, results are often taken from atmospheric cloud physics, where drop
size distributions are better characterized (Ulaby, 1986). The main consequence
of allowing a particle size distribution is that the backscatter σv becomes a
quadratic function of mv . Thus, in this more complicated version of the WCM,
we can combine this with the linear relation between extinction and mv to
postulate a linear relation between the ratio and mv , as shown in equation
(3.118):
σv
≈ C3 mv (3.118)
2σe
Finally, by combining all of these ideas we can then rewrite the expression for
backscatter from a volume in terms of the water content and angle of incidence,
as shown in equation (3.119), for both the uniform particle assumption (3.119a)
and the distributed particle size case (3.119b):
This example illustrates how we can often formulate volume scattering prob-
lems in terms of a small number of physical parameters, and then use scattering
models to simplify the expressions. However, we note that from a polarisa-
tion point of view the predicted behaviour of the WCM is trivial. With the
assumption of independent identical spherical particles and single scattering,
the scattering matrix for each is just the 2 × 2 identity, and hence there is zero
cross- or depolarisation prediction and equal scattering (in amplitude and phase)
for the copolar channels. In the next section we shall see how to extend this
model to account for non-spherical particles. We shall also make use of this
model again in a coherent form when considering polarimetric interferometry
in Chapter 7. In order to illustrate how polarisation effects can be included
in volume scattering we turn now to consider a more complicated version of
the WCM.
2c
z h z h z h
Fig. 3.37 Three examples of volume scatter-
x x x ing: prolate (left), spheres (centre), and oblate
(right)
Here the parameter Lu is the particle shape factor of the ellipsoid along its u
axis, as already encountered in equation (3.76). We now consider three types of
volume of interest: one with embedded needles or prolate spheroids as shown
on the left side of Figure 3.37; one with spheres, at the centre of the figure;
and one with discs (oblate spheroids) on the right of the figure. In all cases
the positions of the particles are assumed random, but their orientations are the
same. For each of these we can write the following special forms of equation
(3.121) for the effective dielectric constant of the medium.
a) Case 1: A volume of needles aligned along the z axis (La = Lb =
0.5, Lc = 0). In this case we have different dielectric constants for dif-
ferent polarisations. For the electric field polarised in the x, y and z
directions we have the following values:
(εi − εh )
εrx = εh 1 + 2vi
(εi + εh )
y (3.122)
εr = εrx
εrz = εh + vi (εi − εh )
εrx = εh + vi (εi − εh )
y
εr = εrx
(3.124)
εh
εrz = εh 1 + vi 1−
εi
The above examples are all for so-called oriented volumes, where the parti-
cles are aligned to form a crystal-type structure. However, in many practical
applications the particles show no particular alignment (forming a random vol-
ume). For random orientation the dielectric constant no longer depends on
polarisation, and the scalar mixture formula takes the following form:
vi ! 1
εr = εh + (εi − εh ) (3.125)
3 1 + Au ( εεhi − 1)
u=a,b,c
It can be seen that this involves an average over the shape functions along
the three principle particle axes. For a cloud of spheres such averaging, of
course, makes no difference, and the effective medium permittivity is the same
as equation (3.123). For the two other cases, however, we obtain the following
modified results.
d) Case 4 : For a random cloud of oblate spheroids (discs):
vi εh
εr = εh + (εi − εh ) 2 + (3.126)
3 εi
e) Case 5 : For a random cloud of prolate spheroids (needles):
vi (εi − εh )(5εh + εi )
εr = εh + (3.127)
3(εh + εi )
V We can now make use of these results to estimate the polarisation dependent one-
way loss factor L(w, θ ) for a slab of vegetation. Figure 3.38 shows a schematic
H o
Ez of the problem. A wave is incident at θo from the normal, and traverses a slab
of thickness h. In particular, we consider the slab to be constructed of three
h phases of material, the first—a host background—being free space, so that
εh = 1. The second phase is a random collection of discs, modelling the leafy
Fig. 3.38 Geometry of wave propagation or fine-scale structure of the vegetation. Each leaf has a complex dielectric
through a slab of vegetation constant εL . Critically, for this phase the dielectric constant is independent
of polarisation. Finally, we model the stalks or trunks as a vertically oriented
array of prolate spheroids with complex particle dielectric constant εs , as shown
schematically in Figure 3.39. These will have a higher dielectric constant for
vertical polarisation than for horizontal.
The loss factor for the composite medium can then be written as the product
Leaf of polarisation independent (for the random component) and dependent (for the
oriented volume) terms as shown in equation (3.128):
Stalk
L w = LL .LS w (3.128)
Fig. 3.39 Composite vegetation layer com- The random component is given by an exponential propagation factor, as shown
posed of leaf and stalk contributions in equation (3.129), where n = n − in is the complex refractive index of the
3.5 Simple physical models for volume scattering and propagation 173
− cos2πθh λ nL
LL = e 0 (3.129)
For small volume fractions this refractive index can be related to the dielectric
constant as shown in equation (3.130):
√ ε − π hεr
nL = Im( εr ≈ r ⇒ LL = e cos θ0 λ (3.130)
2
2vi εL − 2π hv ε
εr ≈ ⇒ LL = e 3λ cos θ0 L L (3.131)
3
hvL − 2π t LAI εL
LAI = ⇒ LL = e 3λ cos θ0 L (3.132)
tL
The oriented stem loss factors cause a differential extinction; that is, a difference
between the eigenpolarisations, which in this case are H and V polarisation. The
HH polarisation is the easiest to consider first (polarisation perpendicular to the
plane of the page in Figure 3.38). This has a dielectric constant expressed
solely in terms of the stalk volume fraction (again using the approximation that
|εs | 1) as shown in equation (3.133):
) *
y (εs − 1) 4vs εs
εr = 1 + 2vs = (1 + 2vs ) − i 2 2 (3.133)
(εs + 1) εs + ε s
This shows that the H polarised wave is only weakly influenced by the presence
of the stalk components, and the extinction in this channel will be dominated by
the leafy component given in equation (3.126) (which affects all polarisations
equally).
The VV polarisation can be obtained by projecting V onto the x and z axes
and then applying the appropriate component from Case 1 (equation 3.122). For
example, Figure 3.38 shows how to calculate the z component of the electric
field from a geometrical factor sin θo . Likewise. the x component is given by
cos θo and this component ‘sees’the same dielectric constant as HH. The vertical
component sees a different extinction given from Case 1, as shown in equation
(3.134):
εrz ≈ 1 + vs εs − 1 − ivs εs (3.134)
174 Depolarisation in surface and volume scattering
In this way the loss factors for HH and VV polarisation due to the stalk com-
ponents can be written as shown in equation (3.135). Note that for normal
incidence the effect of the vertical stalks is minimal and the extinction is
isotropic (the same in HH and VV). Only as the angle of incidence increases,
so the differential effect increases, with the stalks having maximum influence
at 90◦ or grazing incidence.
4π nHH h
HH
L ≈ exp − nHH = Im
y
εr
λ cos θ
o
Ls (w) ⇒
4π cos2 θo nHH + sin2 θo nVV h
L ≈ exp −
n = Im εrz
VV λ cos θo VV
(3.135)
This model shows how extinction propagation through oriented volumes can
be related to simple physical parameters such as volume fractions and dielectric
constants. The above model, however, is restricted to low-frequency scattering
and in particular to low albedo problems (when the ratio of scattering loss to
total extinction is small).
where F is the frequency in MHz and a and b are regression parameters obtained
from a global fit across all datasets. The mean values of these two parameters
for 20, 50 and 80% of data are shown in Table 3.1. Figure 3.40 shows plots of
extinction at the 20, 50 and 80% levels versus frequency for normal incidence
(θ = 0◦ ) over the range 100 MHz to 1.3 GHz. Figure 3.41 shows the accom-
panying multiplicative scale factor for increasing angle of incidence. To find
the total two-way extinction at frequency F for a given angle of incidence θ,
Forest extinction
12
10
80%
2-Way extinction (dB)
8
50%
20%
4
2.8
2.6
2.4
Extinction scale factor
2.2
1.8
1.6
1.4
1.2
take the value from Figure 3.40 and multiply its dB value by the scale factor
obtained from Figure 3.41. We note the following general points:
1 a + c + 2Re(b) 0
[JH ]Bragg =
2 0 0
(3.137)
1 a + c − 2Re(b) 0
[JV ]Bragg =
2 0 0
Note that these matrices imply zero scattering entropy and α2 = 0 for all
angles of incidence and dielectric constants. We can extend this to include
depolarisation caused by surface roughness by considering the X-Bragg model.
This takes the projected form shown in equation (3.138):
c sin 4 sin 4
1 a + 1 + + 2Re(b) 0
[JH ]XBragg = 2 4 4
2 0
c
1−
sin 4
2 4
c sin 4 sin 4
1 a + 2 1 + 4 − 2Re(b) 0
[JV ]XBragg =
4
c
sin 4
2 0 1−
2 4
(3.138)
Here we see non-zero scattering entropy, but note a mixture of roughness and
dielectric constant dependence in the terms. Contrast this with the full coherency
matrix formalism where we are able to separate roughness and moisture effects.
Still, the level of depolarisation in equation (3.138) is small, and to see how
high the entropy can be we turn again to the case of volume scattering from a
random cloud of anisotropic particles. Using the notation of equation (3.92) we
see that the dual polarisation coherency matrices can be expressed in terms of
the particle anisotropy Ap , as shown in equation (3.139):
1 3A2p + 4Ap + 8 0
[JH ]Vol = [JV ]Vol = (3.139)
15 0 (Ap − 1)2
For dipole scatterers, Ap tends to infinity and we have the strongest depolariser.
For oblate particles, on the other hand, Ap = 0 and the depolarisation is weaker.
These two cases may be distinguished in dual polarised systems using their
3.5 Simple physical models for volume scattering and propagation 177
We see that the maximum dual polarised (N = 2) entropy from such a cloud
is 0.811, and note that while dual polarised systems offer some potential for
the separation of different types of volume scattering based on their levels of
depolarisation, they remain inferior to full [S] matrix systems.
Decomposition
4 theorems
By expanding the matrices on either side in terms of the Pauli spin matrices
we can write this as a unitary vector transformation as shown in equation (4.3)
(verified by expansion and use of standard trigonometric identities):
1 0 0 0 SHH + SVV
1 0 cos 2θ − sin 2θ 0
SHH − SVV
k (θ) = √ (4.3)
2 0 sin 2θ cos 2θ 0 SHV + SVH
0 0 0 1 SHV − SVH
Here we note that the complex sums Shh + Svv and Shv − Svh are invariant to
rotations, which already gives them a special physical significance. Of particular
importance is the form of this transformation for backscatter problems, when
the transmit and receive coordinates coincide and θ corresponds to rotation of
the object about the line of sight. In this case, Shv = Svh in the BSA system,
and so for this case the transformation reduces to three dimensions, as shown
in equation (4.4):
1 0 0 SHH + SVV
1
k (θ) = √ 0 cos 2θ − sin 2θ . SHH − SVV = [R]k (4.4)
2 0 sin 2θ cos 2θ 2SHV
Hence we see that while 4 × 4 unitary matrices [U4 ] are required to represent
the most general bistatic scattering case, 3 × 3 unitary matrices [U3 ] are of
importance for the special case of backscatter. In equation (4.3) we noted that
Shh + Svv is an invariant to rotation, and it is interesting to ask if there are
any other complex combinations of the [S] matrix that remain unchanged by
rotations. To see this we now consider the use of eigenvalue decompositions.
Importantly, these three eigenvectors form a complete set; that is, any k vector
can be expanded uniquely in terms of this mutually orthogonal set. Hence we
can use the matrices corresponding to these vectors as a basis for the expansion
of an arbitrary [S] matrix to obtain equation (4.6):
1 0 0 1 0 0
k = x 0 + y 1 + z 1 = f 0 + ge−i2θ 1 + hei2θ 1
0 i −i 0 i −i
(4.6)
180 Decomposition theorems
where the complex scalars f , g and h remain invariant to rotations of the object
about the line of sight. Furthermore, we may give each vector a physical inter-
pretation in terms of a canonical scattering mechanism by converting the vectors
back into their corresponding [S] matrices, as shown in equation (4.7).
1 0 1 i 1 −i
[S] = f + ge−i2θ + hei2θ
0 1 i −1 −i −1
f + ge−i2θ + hei2θ i(ge−i2θ − hei2θ )
= (4.7)
i(ge−i2θ − hei2θ ) f − ge−i2θ − hei2θ
The first of these corresponds to the sphere symmetric specular scattering pro-
cess, while the remaining two correspond to scattering from a helix, the first
(g) with right and the second (h) with left sense. Both scatter all incident
waves into circular polarisation, and hence are not found generically as paired
polarising objects in scattering by natural media. For this reason an alternative
factorization is commonly used, designed to maintain invariance to rotations
but employing more generic scattering mechanisms.
The remainder term is the same in both cases, and has the form of a canonical
scattering process: dihedral scattering at an angle θ. This expansion was first
proposed in Krogager (1992, 1993), and has the vector form shown in equation
(4.9). Here we see that while the sphere symmetric term has been maintained,
the left and right helical coefficients have been combined into a single domi-
nant helix scattering term and the remainder used to generate a new scattering
mechanism: a dihedral scatterer oriented at angle θ/2.
0 0
(h − g) ei2θ + 2g cos 2θ if |h| > |g|
1
−iei2θ sin 2θ
k = f 0 + (4.9)
0
0
0
(g − h) e−i2θ + 2h cos 2θ if |g| > |h|
ie−i2θ sin 2θ
4.1 Coherent decomposition theorems 181
As well as providing the three real parameters of the SDH decomposition, this
technique can also yield an efficient algorithm for estimation of the orientation
of the scatterer θ (Lee, 2002; Schuler, 2002). From the form of the circular S
matrix (equation (4.8)) we see that θ can be directly related to the product of
SLL and SRR as shown in equation (4.11). In equation (4.11a) we show how the
phase of the LL/RR product is related to θ and to the phase of hg ∗ .
∗
SLL SRR = −hg ∗ ei4θ
∗
(4.11a)
arg(SLL SRR ) = π + arg(hg ∗ ) + 4θ
∗ ∗ ∗ ∗
−SLL SRR = − (SHH − SVV + 2iSHV ) SVV − SHH − 2iSHV
∗
= |SHH − SVV |2 − 4 |SHV |2 + i4Re((SHH − SVV ) SHV )
(4.11b)
4Re((SHH − SVV ) SHV∗ )
⇒ tan−1 = arg(hg ∗ ) + 4θ
|SHH − SVV | − 4 |SHV |2
2
+ ∗
,
+ ∗
, −1 4Re (SHH − SVV )SHV
arg −SLL SRR = tan + , + , = 4θ
|SHH − SVV |2 − 4 |SHV |2
+ ∗
,
1 −1 4Re (SHH − SVV )SHV 1 2Re(t23 )
⇒ θ = tan + , + , = tan−1
4 |SHH − SVV | − 4 |SHV |
2 2 4 t22 − t33
(4.11c)
In equation (4.11b) we show how the phase of −SLL SRR ∗ can then be related to
the sum of 4θ and a scattering phase between h and g. Importantly, the product
hg ∗ is often real, and so this phase is zero. We can see from equation (4.11b)
that this will occur, for example, when SHV equals zero. This always occurs for
some θ for so-called symmetric scatterers, as discussed in the Cameron decom-
position below. More generally, however, in the presence of depolarisation we
can use averaging to estimate the average phase, as shown in equation (4.11c).
Here we can see that the average scattering phase will be zero when we are
aligned to the dominant axis in reflection symmetric depolarisers (see Section
2.4.2.3), since SHV is then uncorrelated with both SHH and SVV . In this case we
see from the equation that θ may then be easily related to the elements of the
coherency matrix [T ]. Note that while this is a robust and widely used algorithm
for estimation of θs its main drawback is the limited range of θ available, due
to the factor of 14 used in the equation.
This discussion about orientation leads us to consider an important class
of objects termed symmetric scatterers (Cameron, 1996; Touzi, 2007). These
182 Decomposition theorems
J V
Symmetry axis
H
a b a –b a 0
Jref [S] = + ⇒
Fig. 4.1 Definition of symmetric scatterers c d –c d 0 d
Here we see that it is only possible to diagonalize [S] by a rotation if the phase
of the complex HV term equals that of the HH-VV term, which again gives
special significance to the Pauli matrix expansion of [S]. This result provides us
with an alternative algorithm for estimating the orientation angle of symmetric
scatterers from the scattering vector, as shown in equation (4.13):
2SHV π π
tan 2θ = − ≤θ ≤ (4.13)
SHH − SVV 2 2
While this appears to be a better approach to the LL/RR phase (it provides
a wider range of angle estimates, for example) it is formed as the ratio of
complex entities and hence is sensitive to fluctuations or departures from the
symmetric assumption (and is unstable when SHH = SVV ; that is, for specular
surface scattering). However, there are two important ideas springing from this
observation, both of which are concerned with devising strategies for dealing
with scatterers that do not obey the symmetry constraint.
The first is called the Cameron decomposition (Cameron, 1996). The idea
here is to filter the scattering vector before applying equation (4.13). The filter-
ing is performed so as to generate the maximally symmetric component of the
complex scattering vector—to keep as much of the original vector as possible,
but to enforce the symmetry assumptions that the phase of HV and HH-VV are
equal and that the amplitude ratio gives tanθ . The former requires that HV and
HH-VV share a common complex factor ε, as shown in equation (4.14). To find
4.1 Coherent decomposition theorems 183
The Cameron algorithm can then be presented in two stages, as shown in equa-
tion (4.15). The first is to generate an estimate of θ from the ratio of real
quantities, and then to use this value to combine the complex numbers b and c
into a single term ε.
a a
k = b −−
−−−−−−−−−−−−−→ sym = ε cos 2θ
max symmetic component k
c ε sin 2θ (4.15)
bc∗ + b∗ c
tan 4θ = ε = b cos 2θ + c sin 2θ
bb∗ − cc∗
Note that the orientation estimate is then formally equivalent to the LL/RR
coherence discussed in equation (4.11). One other important parameter deriving
from this filtering idea is the degree of symmetry Dsym of a scattering vector,
which expresses the ratio of power in the symmetric component to the total
power, as shown in equation (4.l6).
|a|2 + |ε|2
Dsym = 0 ≤ Dsym ≤ 1 (4.16)
|a| + |b|2 + |c|2
2
When we vectorize this expression using the Pauli basis we obtain the following
canonical form:
(λ1 + λ2 ) cos 2τ cos 2τ 0 i sin 2τ λ1 + λ2
k= (λ1 − λ2 ) = 0 1 0 λ1 − λ2 (4.18)
i (λ1 + λ2 ) sin 2τ i sin 2τ 0 cos 2τ 0
184 Decomposition theorems
While this is formally correct, again it does not allow extension to the idea of
coherence and so is susceptible to complex noise fluctuations. For this reason the
most robust algorithm for orientation estimation remains the LL/RR coherence
(equation (4.11)). The only exception to this is in low noise (low entropy)
environments or when the restricted range of the LL/RR estimate is too limiting
for the application.
We see that α changes the copolarised terms of the scattering matrix as required,
although it remains invariant to rotations of the measurement coordinates. The
angle is conveniently defined over the range 0 ≤ α ≤ 90◦ , and represents a
smooth change of scattering mechanism. To see this we start with the sphere
= 0º = 45º = 90º
symmetric specular mechanism with target vector k = [1,0,0,0]T and corre-
sponding 2 × 2 identity matrix for [S]. This we define as the α = 0◦ boundary.
Sphere Symmetry Dipole Dihedral/Helix
Changes in α then correspond to departures in scattering from this reference
mechanism. These variations can be summarized as shown in Figure 4.2. In
the range 0◦ ≤ α ≤ 45◦ the copolarised scattering terms differ in amplitude
but are equal in phase, reaching, as an extreme point, α = 45◦ , which corre-
Fig. 4.2 Physical interpretation of the sponds to an anisotropic scatterer with only one non-zero copolarised term. A
scattering alpha angle simple example of this is a linear dipole scatterer, which has strong copolarised
4.1 Coherent decomposition theorems 185
scatter for polarisations parallel to its axis and zero for the orthogonal case.
In the range 45◦ < α ≤ 90◦ we move into a region where the copolarised
scattering coefficients are 180 degrees out of phase. This can occur for mul-
tiple scattering (dihedrals) or asymmetric scattering such as helices. In the
extreme case of α = 90◦ these scatterers have equal amplitude but maintain
the 180-degree phase difference (metallic dihedrals or helices, for example).
The alpha angle for any [S] matrix can then be directly estimated as shown in
equation (4.21):
|SHH + SVV | π
cos α = √ 0≤α≤ (4.21)
2 |SHH |2 + 2 |SHV |2 + |SVV |2 2
One key advantage of using this angle in place of simple ratios such as SHH /SVV
is that it is invariant to rotations of the object. For example, a dipole changes
its [S] matrix elements with rotation θ (see equation (1.161)), but both numer-
ator and denominator of equation (4.21) remain invariant to θ, and hence α is
unchanged. The same is true for rotation of a dihedral or any other rotationally
dependent scatterer. Hence we can identify the scattering mechanism without
needing to know the orientation. There remains an interesting question of how
best to estimate the mean α angle in the presence of noise and fluctuations. This
cannot be answered using a simple coherence estimate as it was for the orienta-
tion θ, and requires application of the entropy/alpha decompositions presented
in Chapter 2.
To illustrate this we highlight an important special case of this alpha trans-
formation. We can, for example, use it to represent the set of all diagonal [S]
matrices with complex diagonal values λ1 and λ2 as SU(2) transformations of
the identity (corresponding to a sphere symmetric scattering mechanism), as
shown in equation (4.22):
λ1 + λ 2 cos α − sin αe−iδ 0 1
1 λ1 − λ2 = e sin αe
√
iφ iδ cos α 0 0
2 |λ1 |2 + |λ2 |2 0 0 0 1 0
|λ1 + λ2 | λ1 − λ2
⇒ cos α = √ , δ = arg (4.22)
2 |λ1 |2 + |λ2 |2 λ1 + λ 2
We can now combine the above results to propose a ‘point reduction theorem’,
as follows. According to this theorem we can express an arbitrary scattering
mechanism w (and here we restrict attention to the backscatter reciprocal three-
element case) as transformations of the sphere symmetric Pauli scatterer, as
shown in equation (4.23):
cos α
w = eiφ iφ1
sin α cos ψe
sin α sin ψeiφ2
(4.23)
1 0 0 cos 2τ 0 i sin 2τ cos αd − sin αd e−iδ 0 1
= eiφ
0 cos 2θ − sin 2θ 0 1 0
sin αd e
iδ cos αd 0
0
sin 2θ cos 2θ i sin 2τ 0 cos 2τ 0 0 1 0
186 Decomposition theorems
The power in this projection can then be computed as shown in equation (4.26):
It is also of interest to consider the match between w and the scattering mecha-
nism embedded in k. This can be found from the normalized inner product, as
shown in equation (4.27):
w∗T k
d= 0 ≤ |d | ≤ 1 (4.27)
k
and vertical dipoles, we can also count spheres and 45◦ and 0◦ rotated dihedrals
as orthogonal mechanisms. The eigenvectors of the rotation matrix [R] used
in equation (4.5) also form an orthogonal set. However, the SDH components
themselves do not form an orthogonal set (see equation (4.9)). The helix and
dihedral do not represent orthogonal processes.
This orthogonality is a desirable property of coherent decomposition theorem
(CDT) schemes, as it provides optimum separability in classification schemes
and forms the basis for an expansion of second-order averages in the coherency
matrix, as discussed in Section 4.2. To illustrate this we consider a sphere
mapping based on diagonal [S] matrices, as developed in Chapter 3. In this case
we consider classification of scattering vectors of the reduced form, shown in
equation (4.28):
cos α sin α
w1 = sin αeiδ ⇒ w⊥ = − cos αeiδ (4.28)
0 0
The smaller Q, the better the isolation between the mechanisms. Small Q better
wB wD enables separation of the mechanisms when both are present in the same scat-
B tering problem, as they then emerge as separate eigenvectors of the coherency
matrix [T ]. The scenario we are considering is an arbitrary combination of sur-
face and dihedral scattering, as shown schematically in Figure 4.3. Note that
A
the relative amplitude of the mechanisms is not so important in this context
Fig. 4.3 Schematic of combined direct and as is their relative polarimetric properties. Figures 4.4 and 4.5 show the factor
specular dihedral surface returns Q for the same two angles of incidence used in Figures 3.15 and 3.16. Note
that Q is smaller for the larger angle of incidence. For the shallower angle, Q
can still be small (−10 dB or so), but for dry materials (with a small εr < 20)
–14
–12
.4
.6
.8
.2
70
10
60
Dielectric constant A
50
–14
–12
.6
.4
.8
.2
40
30
10
20 –14
–12
.4
.6
.8
.2
10
10 –12
10
10 20 30 40 50 60 70 80
Fig. 4.4 Magnitude of Q (equation (4.23))
for 22.5-degree angle of incidence Dielectric constant B
15
70
60
Dielectric constant A
50
20
15
25
40
2520
30 15
15
15
20 20
30 26
20 20
26 25
10 20 25
25
–15 20
–15 2025
–10
10 20 30 40 50 60 70 80
Fig. 4.5 Magnitude of Q (equation (4.23))
for 45-degree angle of incidence Dielectric constant B
4.2 Incoherent decomposition theorems 189
at shallow angles we note that the separation is the poorest at around −6 dB.
Again we note, for the shallow angle, that it is the dielectric constant of the
second surface B that causes the largest variation in Q.
The larger angle of incidence has a wide range of combinations of dielectric
constants where Q is better than −20 dB. Hence we note that separability of the
two mechanisms wD and wB , based on orthogonality, is largely independent of
dielectric constant variations and improves with increasing angle of incidence.
This discussion of orthogonality, of course, has the largest impact in depo-
larisation problems, when the orthogonal mechanisms can be easily separated
as eigenvectors of the Hermitian coherency or covariance matrix. This brings
us to consider the set of issues raised by incoherent decompositions.
!
R !
R
[T ] = [Ti ] ⇔ [M ] = [Mi ] (4.30)
i=1 i=1
This can then be used as a method of noise filtering, for example, by inverting
equation (4.31):
In this section we look at generalizations and extensions of this idea. Such rep-
resentations are called incoherent decomposition (ICDs), because the addition
of coherency matrices implies addition of power, which in turn implies there is
no phase coherence between the elements. When this is not the case we must
employ coherent decomposition theorems (CDTs)—a description of which was
given in the previous section. There are two key elements to any ICD:
1. The choice of order R in equation (4.30). In principle we are free to
choose any order, but in practice there are several constraining factors.
The first is recognition of the fact that each composite matrix must have
at least one free parameter to describe its form and structure (n for the
noise element in equation (4.31), for example). On the other hand we
have seen that a general coherency or Mueller matrix has at most sixteen
free parameters, and so this means that R ≤ 16. Further restrictions apply
from various symmetry constraints. The most important of these is the
reciprocity theorem in backscatter, for which the coherency matrix has
rank 3 and only nine free parameters. In most backscatter applications,
therefore, R ≤ 9 (a notable exception being the treatment of backscat-
ter Faraday rotation, which is nonreciprocal, and has R ≤ 10). Other
symmetries may apply on a case-by-case basis: for example, backscat-
ter reflection symmetry (a common case in microwave remote sensing),
for which R ≤ 6, and the most severe case, being backscatter azimuth
symmetry with R ≤ 2.
2. The nature of the component elements of the expansion Ti . This leads to
a further general partitioning of the decomposition as follows. The idea is
often to force one or more components of the ICD to correspond to a rank-
1 coherency matrix (or equivalently a Mueller matrix that corresponds to
a single amplitude [S] matrix). Consequently, we can write the general
ICD in the form shown in equation (4.34), where P ≤ R is the number
of rank-1 components:
!
P !
R
!
P !
R
[T ] = [Ti ] + Tj ⇔ [M ] = [Mi ] + Mj (4.34)
i=1 j=Q i=1 j=Q
0 0 0 1 0 0 0 λ2 0 0 0 1 0
1 0 0 0
0 1 0 0
= (λ1 − λ2 ) k D k D + λ2
∗T
0 0
(4.36)
1 0
0 0 0 1
1 0 0 0
1 1 0 1 0 0
= (2A2P + 6AP + 7)k D k ∗T
D + (AP − 1)2
15 15 0 0 1 0
0 0 0 1
It follows that there are still three choices for the null space, given by the three
eigenvectors of the change of basis matrix. Huynen selected the first on physical
grounds and suggested the following general form, which now bears his name:
the Huynen decomposition:
t11 t12 t13 0 0 0
[T ] = [T1 ] + [TN ] = t12
∗ t22 − n22 t23 − n23 + 0 n22 n23
∗
t13 ∗ − n∗
t23 t33 − n33 0 n∗23 n33
23
(4.41)
We clearly see the null space in the remainder term corresponding to the first
eigenvector in equation (4.40), and the elements nij are selected so as to force
the first matrix to have rank 1. However, by choosing one of the other eigenvec-
tors in equation (4.40) for the null space we would produce a different rank 1
matrix in the expansion, and hence we see that although this approach does con-
strain selection of the rank-1 component it does not provide a unique solution.
This approach does, however, raise the issue of orthogonal spaces and their
importance in ICDs. We now turn to consider an alternative formulation, which
solves the uniqueness problem by considering a general eigenvalue expansion.
Cloude and Pottier first proposed this idea in Cloude (1996), and the decom-
position therefore bears their name. Proof of the uniqueness of this expansion
follows directly from the Hermitian nature of the coherency matrix, which can
be expanded in terms of its orthogonal eigenvectors and real eigenvalues, as
shown in equation (4.43):
8
!
4 λ1 ≥ λ2 ≥ λ3 ≥ λ4 ≥ 0 ∈
[T ] = λi ei e∗T
i (4.43)
i=1 e∗T
i ej = 0
!
4
[T ] = λ1 [T1 ] + [TN ] [TN ] = λi ei e∗T
i (4.44)
i=2
In this case the null space of [TN ] is defined by the dominant eigenvector e1 ,
and by definition this remains invariant to all unitary transformations of the
problem, and not just the rotation invariance required in the Huynen approach.
We note that this approach also has the advantage of automatically scaling with
varying rank of the coherency matrix. In backscatter reciprocity, for example,
the smallest eigenvalue is always zero, and the most general expansion is there-
fore provided in terms of three components, as shown in equation (4.45). The
extension to rank 2 and rank 1 depolarising systems follows immediately.
In this case, therefore, the best single mechanism to isolate in each matrix
is w, and not the individual eigenvectors ei . In this case we make use of the
following ICD:
volume-scattering elements that return the reflected signal back to the observer.
While this combination can often be complex and a source of wave depolari-
sation, it is often assumed dominated by a double specular reflection process
or simple dihedral reflection from dielectrics, rather than a reflection/scattering
combination. For this to occur the volume needs to be populated not only by
Dihedral/double bounce
scattering
small particles (leaves, branches, and so on) but also by some electrically large
scatterers (vertical tree-trunks, for example), in which case the reflection from
the surface (modelled by the Fresnel equations, modified by the Rayleigh coef-
ficient to account for surface roughness; see equation (3.31)) is followed by
a second Fresnel reflection from the surface of the large scatterer, as shown
Fig. 4.7 Geometry of dihedral backscattering schematically in Figure 4.7. This reflection can be written in terms of the angle
mechanism of incidence using the constraint that the final angle of reflection must be 180◦
for backscatter, shown in equation (4.52):
cos θ − εr1 − sin2 θ
RH 1 =
cos θ + εr1 − sin2 θ
εr1 cos θ − εr1 − sin2 θ
RV 1 =
RH 1 RH 2 0 εr1 cos θ + εr1 − sin2 θ
[SD ] ∝ → (4.52)
0 −RV 1 RV 2
sin θ − εr2 − cos2 θ
RH 2 =
sin θ + εr2 − cos2 θ
εr2 sin θ − εr2 − cos2 θ
RV 2 =
εr2 sin θ + εr2 − cos2 θ
measurements. There are, of course, many ways to do this, but several common
assumptions emerge in practice.
As mentioned, the main reason for adopting this simplified form is to enable
inversion of the model directly from data, as follows. The first step is to use
the HV channel to directly estimate the volume scattering component mv , as
shown in equation 4.54:
mv = t33 (4.54)
Also shown in equation (4.55) are equations used to estimate the scattering
power terms ms and md . Finally, the last parameter, the scattering phase, can
be estimated as shown in equation (4.56):
π
φd = arg(t12 ), φs = 0 if αF >
4
(4.56)
φs = arg(t12 ), φd = 0 π
if αF <
4
Note, however, that the model inversion does not guarantee non-negative
estimates of power. In the presence of noise or violations of the underlying
model assumptions, some of the power components could be estimated as
negative—an undesirable feature of such models.
Note also that this approach provides a means for avoiding negative powers by
keeping only the non-negative eigenvalue spectrum of TSD . Also, if we assume
one of ms , md is always zero then we can estimate the volume parameter Fp .
We can also use this model to estimate the ratio of surface-to-volume scattering
components, termed µ, which, as we shall see in Chapter 7, is an important
parameter in the development of polarimetric interferometry. According to the
model, when we select the crosspolarisation channel HV, the surface-to-volume
scattering ratio µ is by definition zero, as the surface components are all con-
strained into the rank-2 upper portion of the coherency matrix. On the other
hand, to find the surface-to-volume ratio for the largest surface component we
can proceed as follows. First we select the maximum ‘surface’ component from
the maximum of the pair ms and md as defined in equation (4.59). (Note that we
are including both direct surface and dihedral returns in the ‘surface’ compo-
nents to distinguish them from the pure volume scattering component.) From
this we can then calculate the alpha parameter of the maximum component, as
shown in equation (4.60):
" 1
2 − 2
mmax = max(md , ms ) ⇒ αmax = cos−1 1 +
t12
t22 − t33 − mmax
(4.60)
200 Decomposition theorems
Notice that this ratio is not the maximum surface-to-volume ratio. This can
always be obtained by using a w calculated as a generalized eigenvector of
the matrix [TV ]−1 [TSD ] (see equation (4.46)). The general solution to this is
complicated, but if we make the assumption that there is one dominant surface
component (either ms or md ), then we can obtain a direct estimate of the maxi-
mum surface-to-volume ratio by finding the largest eigenvalue of [TV ]−1 [Tmax ],
as shown in equation (4.62):
cos2 αmax cos αmax sin αmax
mmax − µ
Fp Fp =0
mv cos α sin αmax − µ
max sin αmax
2
(4.62)
mmax 1
⇒ µmax = sin2 αmax + cos2 αmax
mv Fp
It is also interesting to note that we can also define another small value of
surface-to-volume scattering (in addition to the HV channel) by selecting the w
vector orthogonal to the maximum surface component. In this case we obtain
a µ value, given as shown in equation (4.63):
− sin αmax
w = cos αmax ⇒ w∗T [T ] w = mmin + mv (1 + (Fp − 1) sin2 αmax )
0
surface mmin
⇒ µmin = = (4.63)
volume mv (1 + (Fp − 1) sin2 αmax )
assumption that the volume is random and hence acts as a scalar attenuation of
all polarisations equally, which has no effect on the alpha parameters of surface
and dihedral components. However, in some cases this assumption is not true,
and leads us to yet another class of decompositions—but this time based on a
multiplicative expansion. We now turn to consider such ideas.
Note that the crosspolarised channels are influenced by the mean propagation
constant, while the copolar channels are also influenced by the differential
propagation. We can combine these effects into a single 4 × 4 matrix [PL ] as
202 Decomposition theorems
shown. The final step is then to express the matrix [PL ] in the Pauli base, so
we can finally relate propagation to the coherency matrix formulation. This is
shown in equation (4.67), where we see we can express the matrix in terms of
elementary hyperbolic functions.
1 0 0 1 1 1 0 0
1 1 0 0 −1 0 0 1 −i
[PP ] =
[PL ]
2 0 1 1 0 0 0 1 i
0 i −i 0 1 −1 0 0
cosh νz sinh νz 0 0
sinh νz cosh νz 0 0
= e−i(βa +βb )z
0
(4.67)
0 1 0
0 0 0 1
where z is now the slant range coordinate, related to z , the vertical coordinate,
by θ o , the angle of incidence (see Figure 3.38). We note again the attenuation
of amplitude by mean wave extinction, but also note that since P is not gen-
erally diagonal then there results some distortion to the apparent polarimetric
parameters of [T ].
To illustrate this we consider the effects on the apparent scattering param-
eter α for a scatterer viewed through a medium with zero retardence but total
4.2 Incoherent decomposition theorems 203
80
70
Observed alpha angle
60
50
9 bB
40
30 6 dB
3 dB
0 dB
20
10
cos α −s cosh A sinh A cos α
=e .
sin α sinh A cosh A sin α
(4.70)
sin α + cos α. tanh(0.1151 ∗ )
⇒ tan α =
cos α + sin α. tanh(0.1151 ∗ )
Figure 4.8 shows how this distortion behaves for varying differential extinction.
We see that for = 0 dB we have zero distortion as expected (this is the
basic assumption of the Freeman decomposition). However, as increases
we see that the range of apparent alpha becomes compressed until, when the
differential extinction becomes very large, we obtain an apparent alpha of 45◦
for all scatterers. This corresponds to all scattering being filtered by a polariser
with an orientation given by the dominant eigenpolarisation.
Finally, we consider the more general case when the eigenpolarisations are no
longer matched to the backscattering matrix basis. In this case we must modify
equation (4.63) to include a unitary congruent transformation by a 2 × 2 unitary
matrix, as shown in equation (4.71):
−iβ z
e−iβa z 0 Saa Sab e a 0
[S]observed = [U2 ] . [U2 ]T .
0 e−iβb z Sba Sbb 0 e−iβb z
/
βa = βo na − iκa
(4.71)
βb = βo nb − iκb
204 Decomposition theorems
Now converting to the Pauli basis we obtain the following form for the unitary
matrix, as shown in equation (4.74):
1 0 0 1 1 1 0 0
0 1 1 1 0 0 −1 0 1 0 0 1 −i
L
U4P = U
2 0 1 1 0 4 0 0 1 i
0 i −i 0 1 −1 0 0
2
cos αw + sin αw cos 2δ
2
i sin2 αw sin 2δ 2i sin αw cos αw sin δ 0
−i 2
α 2 α − 2
α −2 sin αw cos αw cos δ 0
sin w sin 2δ cos w sin w cos 2δ
=
2i sin αw cos αw sin δ 2 sin αw cos αw cos δ cos2 αw − sin2 αw 0
0 0 0 1
(4.74)
in equation (4.76):
∗T
[T (z)] = e−(κb +κa )z [P(τ )] U4P (αw , δ) [T ] U4P (αw , δ) P(τ ∗ )
z (4.76)
τ = νz = (κb − κa ) − iβo na − nb
cos θo
cos ψ sin ψ SHH SHV cos ψ sin ψ
Sψ = (4.77)
− sin ψ cos ψ SVH SVV − sin ψ cos ψ
By expansion and vectorization in terms of the Pauli matrices this can be written
as a unitary transformation of the scattering vector, as shown in equation (4.78):
cos 2ψ 0 0 − sin 2ψ SHH + SVV
1 0 1 0 0 SHH − SVV
k (ψ) = √
SHV + SVH (4.78)
2 0 0 1 0
sin 2ψ 0 0 cos 2ψ SHV − SVH
SHV − SVH
tan 2ψ = (4.79)
SHH + SVV
One problem with practical use of equation (4.79), however, is that it involves
the ratio of two complex quantities on the right and a real quantity on the left,
and hence it is susceptible to data fluctuations and residual calibration errors.
It would be more robust to devise an algorithm that involves the ratio of real
quantities so that averaging may be used. We can use the scattering vector
formulation combined with a 4 × 4 unitary transformation into the circular
basis to solve this as follows.
206 Decomposition theorems
This can be derived from the circular change of base transformation in equation
(1.162). This can then be used to transform equation (4.80) into the circular basis
by a unitary similarity transformation, as shown in equation (4.81):
SLL (ψ) cos 2ψ 0 0 − sin 2ψ SLL
SLR (ψ) 0 1 0 0 SLR
[U4circ ]∗T
SRL (ψ) = [U4circ ] 0 0 1 0 SRL
SRR (ψ) sin 2ψ 0 0 cos 2ψ SRR
1 0 0 0 SLL
0 exp (−i2ψ) 0 0
= . SLR (4.81)
0 0 exp (i2ψ) 0 SRL
0 0 0 1 SRR
Note, significantly, that the copolarised terms (LL and RR) are not distorted by
Faraday rotation. If again we assume that the underlying scatterer is reciprocal
(SLR = SRL on the right-hand side of equation (4.81)), as in backscatter BSA
problems, then the Faraday rotation angle can simply be found from the phase
of the product of crosspolarised terms, as shown in equation (4.82):
1 ∗
ψ= arg(SRL SLR ) (4.82)
4
such phase information and hence reduce our ability to exploit coherence. There
is, however, a second way to generate phase information from active systems:
by the use of interferometry. Here we shall see that new possibilities arise for
the control of phase, even in random media problems, and this, when combined
with polarimetry, will lead us finally to a general formulation of how best to
exploit the ‘orientation memory’ effect in scattering from random media.
Introduction to radar
5 interferometry
Here we have further decomposed the phase of the signal into two parts—the
first a propagation phase that depends on the distance between the radar and
the scattering point (R1 ), and the second a scattering phase that depends on
the detailed nature of the scattering process. In polarimetry we concentrated
on looking at the changes of scattering phase by combining two signals with
different polarisations but collected at the same point in space, so cancelling
the range phase term, as shown in equation (5.2):
.
s1 = a1 e−i( λ R1 +φS1 )
4π
particles this depolarisation can be very high (especially in the case of a random
volume of extremely prolate spheroids), leading to a noisy polarimetric phase.
It would be useful to be able to counter this depolarisation in some way. Radar
interferometry provides such an option, as we now show.
S1
In radar interferometry we begin again with the reference signal s1 of equation
B
(5.1). However, rather than diversify polarisation for a fixed position, we now db
S2
keep polarisation constant but vary spatial position to collect a second signal
s2 from a displaced point, separated from the first by a spatial baseline B, as
R2
shown in Figure 5.1. Note that there are inherently two ways to achieve this.
R1
The first—‘single-pass interferometry’—occurs when we use two radars at the
H
same instant in time to collect signals s1 and s2 . Alternatively, it is often more
convenient (and more cost effective) to use a single radar system to collect a P z
signal at position 1 and then move the radar to position 2 at some time later
to collect s2 . In the airborne/space radar context this is called ‘repeat-pass h0
y
interferometry’, and the time difference between the data collections is called
the temporal baseline TB (in seconds), to be contrasted with the spatial baseline Fig. 5.1 Baseline geometry of radar interfer-
B (with units of metres). Note that we have another degree of freedom for ‘single ometry
pass’ sensors in choosing either to transmit and receive from both points (the
dual transmitter mode described in Figure 5.1 and often termed ‘ping-pong’), or
just to transmit from 1 and receive on both 1 and 2 (called the single transmitter
or standard mode). In this case we can avoid the expense and complexity of a
second transmitter. However, in this case the factor of 4π in equation (5.3) is
halved to 2π , since the phase difference now only depends on R1 − R2 rather
than 2R1 − 2R2 . In what follows we consider the dual transmitter ping-pong
mode to illustrate the main form of governing equations.
In either case, when we form the complex product s1 s2∗ for interferometry we
cancel the scattering phase terms (assuming they are the same; we shall see later
how good this approximation is in practice) and keep the geometrical phase. In
effect we now obtain a signal phase which depends only on the difference in
range between the two positions, R = R1 − R2 , as shown in equation (5.3):
.
s1 = a1 e−i( λ 2R1 +φS1 )
2π
4π
φ = arg{s1 s2∗ } = − R + 2πN N = 0, ±1, ±2, . . . (5.4)
λ
Note that this interferometric phase is 2π ambiguous; that is, for every shift of
half a wavelength λ/2 in range difference R the phase difference repeats itself.
We therefore face problems if we wish to invert this equation and determine R
from a measurement of phase, as there are multiple solutions. This is called the
‘phase unwrapping problem’ (Bamler, 1998), and its solution requires the use
of additional information, such as use of phase from nearby scattering points
to generate spatial phase gradients and force continuity, or the use of multiple
baselines (data from three or more spatial positions).
In what follows we consider three different types of radar interferometry
used in remote sensing: across-track, differential, and along-track modes.
210 Introduction to radar interferometry
z z
R R
ysin
zcos
y y Fig. 5.2 Approximate ray-path geometry for
P phase estimation
θ1 +θ2 θ2 −θ1 4π θ
s1 s2∗ = Aei y(sin θ2 −sin θ1 ) ) sin( ) y cos(θ )
4π 4π
λ = Aei λ
2y cos( 2 2 ≈ Aei λ
(5.10b)
θ +θ θ −θ
i 4πλθ
s1 s2∗ = Ae −i 4π
λ
z(cos θ2 −cos θ1 )
= Ae i 4π
λ
2z sin( 1 2 2 ) sin( 2 2 1 )
≈ Ae z sin(θ )
(5.10c)
In case the angular difference to the point θ = (θ2 − θ1 ) is small (that is,
B R), we can further simplify this expression, as shown in equation (5.10b).
Similarly, for points P shifted in the z direction (the right-hand side of Figure
5.2) we have an extra phase term, −z cos θ, which leads to a corresponding
interferometric phase, as shown in equation (5.10c). Combining these two, we
obtain the following expression for the total interferometric phase as a function
of y and z around the point P.
4π θ
φif (y, z) = (y cos θ + z sin θ) (5.11)
λ
Currently we see that the phase varies not only with height of the point P (the
coordinate z) but also for shifts in surface position (y). This somewhat disturbs
our desire to have a clean phase-to-height conversion. We can remove this
problem and obtain our desired result by shifting the frequency of the signal
collected at position 2 before forming the interferogram, using a process called
range spectral filtering, as follows.
212 Introduction to radar interferometry
θ
if f = f
tan θ
2θ 4π θ
⇒ β = β =
tan θ λ tan θ
⇒ φ = φif + β (z cos θ − y sin θ)
4π θ 4π θ (5.13)
⇒φ= (y cos θ + z sin θ) + (z cos θ − y sin θ)
λ λ tan θ
4π θ ∂φ
⇒φ= z= z = βz z
λ sin θ ∂z
4π θ
⇒ βz =
λ sin θ
z This transformation can be given a simple geometrical interpretation if we
s1
employ a two-dimensional wavenumber representation of interferometry as
shown in Figure 5.3. Here we show the complex signal from the first sensor
position as a vector in the β y ,β z plane with magnitude equal to 2π/λ and angle
s2 θ to the point P. The second position (before frequency shift) then yields a
vector with the same length but rotated about the origin by θ, as shown. If we
y consider the y components only, then in order to give s2 the same projection on
cos the y axis as s1 requires a radial shift in wavenumber by β (which is negative
if θ is positive) before forming the interferogram.
Fig. 5.3 Wavenumber space geometrical
Furthermore, this idea can be generalized to the bistatic surface scattering
interpretation of radar interferometry
case, using a concept known as the memory line, first described in Le (1998).
This can be motivated by considering the geometrical representation of a general
bistatic scattering geometry in wavenumber space, as shown in Figure 5.4. Here
we see that a bistatic angle ψ leads to a shift of the point representation in wave-
space in both angle and radial coordinates. To see this, just consider the general
z
r
r1 r2
Real space Wave-space
y
c c
Fig. 5.4 Geometry of generalised bistatic r = 2b cos f=u+
radar in real and wavenumber domains 2 2
5.1 Radar interferometry 213
c
phase term as shown in Figure 5.5. Now consider forming an interferogram 2|b|cos
2
ib.r i|b|(r1 + r2)
between signals for two different bistatic geometries with corresponding wave- e =e ⇒ |b|
2
c
space coordinates r1 ,φ 1 and r2 ,φ 2 . In order that the phase difference between |b|
signals has no y-dependence and depends only on height, then equal projections
onto the y axis are required, which in turn requires that equation (5.14) holds: Fig. 5.5 Bistatic triangle construction for
radar interferometry
ψ ψ
|β| cos cos θ + =C (5.14)
2 2
Consequently, we can now change any of the three parameters, θ, ψ and β,
B
but as long as the product shown in equation (5.14) remains constant then the
interferometric phase will be a function only of height. Equation (5.14) is the
equation of a line—the so-called memory line of the interferometer—and all
systems lying along this line satisfy our desired phase dependency. We see that
for the special case of backscatter (ψ = 0) this constraint reduces to βcosθ being
constant, which is the origin of the frequency shift introduced in equation (5.13).
B
Returning now to the case of backscatter, we can obtain a useful approxima- u =
R
tion for θ as found in equation (5.13) in the special case R B (as occurs
for spaceborne radar for example). In Figure 5.6 we show how the angular
baseline width θ can be approximated from the geometry as θ ≈ B⊥ /R.
Note that this expression does not include the absolute baseline B, but only its
component perpendicular to the range line. This leads us to consider the various
spatial components of the baseline, as follows. Fig. 5.6 Baseline geometry triangle
V
S1
B
BH = Bcosdb
H B BV = Bsindb
S2
B = Bsinf = Bsin ( p – u + db) = Bcos(u – db)
2
p
B = Bcosf = Bcos ( – u + db) = Bsin(u – db)
u 2
π λ sin θ λ R0 sin θ
hπ = = ≈ (5.16)
βz 4θ 4 B⊥
Note that since Ro B, this height is much greater than a wavelength. We
shall see later how in differential interferometry we can obtain much higher
sensitivities when the π -height itself reduces to λ/4. Before leaving this topic
we highlight three important issues.
1. The frequency shift used in equation (5.12) is actually a function of the
local angle of incidence at the point P. Hence, in the presence of topographic
variations it is affected by range slopes. For example, if the terrain is sloped
towards the radar then the effective angle of incidence required to determine the
frequency shift is reduced. In the extreme case, when the surface slope equals
the radar angle of incidence we ultimately determine effective normal incidence
onto the surface and lose sensitivity completely. In this case we obtain so-called
‘blind’ angles for the interferometer. In general, therefore, in the presence of
range slope η we must modify the expression for spectral shift, as shown in
equation (5.17):
cB⊥
f = − (5.17)
R0 λ tan (θ − η)
As an example we consider evaluation of the required spectral shift for typi-
cal values of a space radar (ERS-1) operating at 23◦ of incidence at C band
(λ = 5.66cm) and transmitting a signal with bandwidth 15.5 MHz with a slant
range of 840 km. Figure 5.8 shows the required frequency shift as a function
of surface slope. We note the following important features:
(a) As slope increases from zero, the required frequency shift eventually
exceeds the bandwidth of the pulse (f /W = 1). This is the start of
a band of ‘blind angles’, and the width of this band is defined from
this point until the slope increases to the extent that the frequency shift
returns to the pulse bandwidth. This high angle range, however, is
plagued by layover distortion in radar imaging, whereby points at the
top of a slope are at closer range to the radar than those at the foot of
the slope.
(b) In the negative range slope direction (away from the radar) the required
shift slowly decreases to zero for grazing incidence on the surface.
Thereafter the slope disappears into shadow, and there is no longer a
measurable radar backscatter return.
Correct application of the frequency shift concept therefore requires detailed
knowledge of the terrain slope. This can be obtained, for example, if a reference
digital elevation model (DEM) is available, from which the range slopes can
be calculated by spatial differentiation. In this way an adaptive variation of
spectral shift can be performed. In the absence of such topographic information,
5.1 Radar interferometry 215
15
Lay-over
10
Frequency Shift (MHz)
Shadow
5
0
Blind angles
–5
–10
–15
–20
–80 –60 –40 –20 0 20 40 60 80
Fig. 5.8 Example of frequency shift versus
Surface slope (degrees) surface slope
W-f
dashed = spectrum of s1
solid = shifted spectrum of s2
This is called ‘flat earth’ removal for interferometric processing. Note that it
is not a replacement for spectral shift processing, and its sole purpose is to
remove the high-frequency modulation of the underlying phase to facilitate
further processing and interpretation. Note that for airborne radar or other large
swath geometries, when the angle of incidence varies across the range swath,
the spatial frequency of the flat-earth signal is no longer constant but increases
with decreasing angle of incidence, leading to a ‘chirp’ or variable frequency
reference. In this case the same procedure can be used for flat-earth removal,
but fs must be evaluated more carefully by accurate evaluation of the θ term.
We now briefly consider two important variations on radar interferometer
design. The first—called ‘differential interferometry’—is designed to maximize
sensitivity to changes in repeat-pass sensors; while the second—along-track
interferometry (ATI)—is designed to sense the velocity rather than position
vector of scatterers.
5.1 Radar interferometry 217
s1s2 d
d
R2
H
R1
then remove its contribution from the phase signal, as shown in the lower
portion of equation (5.24). One potential problem with this approach is that
the reference DEM employed will have some errors, and these can propagate
into the displacement estimate. For example, if the DEM has an elevation error
z then this translates into an equivalent phase error of φ z which in turn is
interpreted as a ground motion dz , as shown in equation (5.25):
4π B⊥ λ B⊥
φz = z ⇒ dz = φz = z (5.25)
λR0 sin θ 4π R0 sin θ
This can be expressed more conveniently in terms of the critical baseline for
a sensor with bandwidth W operating at centre frequency f0 , as shown in
equation (5.26):
B⊥ B⊥ W z
dz = z = (5.26)
R sin θ Bcrit f0 cos θ
Clearly the error can be minimized by reducing the baseline, and in the limit
of zero spatial baseline the topographic error is removed. The second approach
(when a reference DEM is not available) is to employ three (or more) passes to
eliminate the topographic phase dependence. If n ≥ 3 passes are available then
we can define a set of interferometric parameters, as shown in equation (5.27):
We can then decompose the set of phases φ n−m into a part proportional to the
spatial baselines (topography) and a part proportional to the temporal baselines
(displacement at constant velocity). To illustrate this we consider the simplest
case of three passes, yielding complex signals s1 , s2 and s3 . The first signal,
s1 , is then used as a ‘master’ to generate interferograms with both s2 and s3 .
The first baseline B12 we wish to be dominated by topographic effects, and
so should be as large as possible, with a small temporal baseline so as to
minimize displacement effects. The second baseline, B13 , however, should be
dominated by temporal effects, and so should combine a small spatial with
long temporal baselines. With this combination the displacement phase can be
directly estimated as shown in equation (5.28). (Note that if the baseline ratio is
too large (>4) then phase unwrapping may be required of the φ 12 interferometric
phase before scaling.)
B13
φdisplacement = φ13 − φ12 (5.28)
B12
The second problem faced in differential interferometry is the effect of wave
propagation between the sensor and the surface. The propagation of microwaves
through the atmosphere causes a phase shift due to variations in refractive index.
For repeat-pass sensors the changing atmosphere causes a change in this phase
and hence a phase error in the interferogram. Such propagation effects can be
large, for example, for a spaceborne radar at C-band, as atmospheric delays
can cause an error of half a fringe (π radians), or in extreme cases up to three
5.2 Sources of interferometric decorrelation 219
fringes. For low-frequency radars, phase shifts due to propagation through the
ionosphere can cause similar problems (Freeman, 2004). In general, therefore,
the phase of an interferogram can be written in component form, as shown in
equation (5.29):
on), depending on the configuration used. However, so far we have ignored the
influence of noise and its impact on phase estimation. In polarimetry we saw
that noise arises from depolarisation and is manifest as an increase in scattering
entropy. In this section we first consider a formalism to include noise effects
in radar interferometry, and then consider a set of various potential sources of
noise (Zebker, 1992). Some of these are system related (signal-to-noise ratio,
for example), but others are related to wave scattering effects (volume and
baseline decorrelation in particular) and hence can be considered analogous
to wave depolarisation effects in polarimetry. By considering such coherent
scattering in detail, we will see how we can then turn the noise problem around
and use the interferometric coherence as a new radar observable to help estimate
surface and volume scattering parameters. This will then lead us to consider, in
the next chapter, combinations of polarimetry with interferometry.
We start with a general expression for interferometric phase φ as the sum
of ‘signal terms’ φ if and a noise term φn , (equation (5.31)) characterized by
its statistical moments. Generally the noise term will have zero mean, but is
characterized by non-zero standard deviation σ φ .
φ = φif + φn (5.31)
To see the impact of such stochastic fluctuations in the phase on surface param-
eters, consider the important special case of surface height estimation using
across-track radar interferometry. As shown in equation (5.32), the phase vari-
ance σ φ leads to a scaled height variance, derived from the relation of phase
to changes in slant range σ R and then using equation (5.15) to relate range to
height via the normal baseline.
λ λ R0 sin θ R0 sin θ λ
R= φ ⇒ σR = σφ ⇒ σh ≈ σR ≈ σφ (5.32)
4π 4π B⊥ B⊥ 4π
This relation can be used to estimate errors in surface height based on system
parameters (baseline geometry and angle of incidence) and the phase variance.
To proceed, we need to further investigate the different ways in which phase
noise can be generated in radar interferometry. To do this we first relate noise
variance to an underlying coherence.
We start by employing a coherency matrix formulation of radar interferom-
etry, as shown in equation (5.33). Here again we can define a useful secondary
parameter: the interferometric coherence, with a magnitude between 0 (pure
noise) and 1 (pure signal).
+ 2 , + ∗ , + ∗,
|s1 | s1 s2 s1 s2
[T2 ] = + , + 2 , ⇒ γ̃ = + ,+ , (5.33)
∗ |s2 |
s2 s1 |s1 |2 |s2 |2
From Appendix 3 it then follows that the phase variance can be related to the
coherence by the following Cramer–Rao bounds (Seymour, 1994):
"
1 − |γ |2 1 − |γ |2
σφ ≤ σ|γ | ≤ √ (5.34)
2L |γ |2 2L
5.2 Sources of interferometric decorrelation 221
s 1 = a + n1
(5.35)
s2 = a + n2
Now we must invoke some assumptions about the statistical distribution of the
noise terms. The most common assumption, based on the central limit theorem,
is that the noise terms are complex Gaussian random variables and hence have
uniform phase distributions and are uncorrelated both with the signal (a) and
with each other. Under this assumption the coherence can be evaluated as shown
in equation (5.36):
|a|2 SNR
γsnr = = (5.36)
|a| + |n|
2 2 1 + SNR
Here we see a simple relation between the signal-to-noise ratio (SNR) and
coherence. As the SNR tends to infinity (zero noise) then the coherence tends
to unity, while if the SNR tends to zero then the coherence also tends to zero.
Figure 5.12 shows how the coherence is related to SNR (expressed in dB).
Noise decorrelation
1
0.9
0.8
0.7
0.6
Coherence
0.5
0.4
0.3
0.2
0.1
0
–30 –20 –10 0 10 20 30
Fig. 5.12 Relationship between signal-to-
SNR (dB) noise ratio and coherence
222 Introduction to radar interferometry
γ = = (5.38)
|a|2 + |m|2 SCR + 1
where SCR is now the signal-to-clutter ratio. The most important application
of this model is to temporal decorrelation in repeat-pass interferometry. In this
case the ‘clutter’ noise is caused by motion of scatterers (such as wind-driven
vegetation) between passes. If the rms motion along the line of sight is δ rms ,
then the phase variance and hence coherence in equation (5.38) can be simply
related to this shift, as shown in equation (5.39):
16π 2
4π SCR + e− λ2 δrms
σφ = δrms ⇒ γ = (5.39)
λ SCR + 1
Importantly, we see that this coherence depends on the ratio of rms motion to
wavelength, and hence for a given shift the effect on coherence is worse for
higher frequencies. This drives us to consider lower frequencies to minimize
the effects of temporal decorrelation in repeat-pass interferometry (Hagberg,
1995; Askne, 1997, 2003, 2007).
Note that in the general case we can have a combination of these statistical
effects, such as temporal decorrelation γ t in combination with noise decorre-
lation γ snr . In this case the coherence is formed from products of triple sums,
as shown in equation (5.40). The most important consequence of this is that
coherence always decomposes in a multiplicative series of component terms
(to be contrasted with polarimetric decomposition which led to expansion as a
sum of component terms).
|a|2 + |m|2 e−σφ
2
s1 = a + meiφ1 + n1
⇒γ = 2
s2 = a + meiφ2 + n2 |a| + |m|2 + |n|2
|a|2 + |m|2 e−σφ
2
|a|2 + |m|2
= . = γt γsnr
|a|2 + |m|2 |a|2 + |m|2 + |n|2
(5.40)
Typically in radar applications the two signals s1 and s2 are formed from co-
registered synthetic aperture radar (SAR) images (often collected at different
times of a repeat orbit), and in practice it is impossible to exactly match the
two radar signals (see Chapter 9). There will always be some small residual
fractional offset in range and azimuth pixel size δ rg and δ az between the two
images (Krieger, 2005). This causes a coherence loss component given by
equation (5.41):
Current processing accuracies are limited to about 1/10 of a pixel in both range
and azimuth, and so we see that this error is independent of baseline and has
a value around 0.97. This error becomes particularly significant in the case of
small spatial baselines, where it can become the dominant source of decorre-
lation. To combine this error source with the other two we simply extend the
decomposition of equation (5.40), as shown in equation (5.42):
zo +hv z = z − z0 hv
s1 s2∗ = f (z)eiβz z dz −−−−−−−−−→ eiβz zo f (z )eiβz z dz (5.45)
zo 0
- hv
f (z )eiβz z dz
γ̃ = e iβz zo 0
- hv = eiβz zo |γ̃ | ei arg(γ̃ ) (5.46)
0 f (z )dz
Note that this is a complex coherence; that is, it has phase as well as magnitude,
and part of the phase arises from the integral of the structure function f (z)
shown in the numerator. This real non-negative function allows for arbitrary
profile of scattering between the bottom and top of the layer (Cloude, 2006b).
This relation shows that there is a direct relationship between the observed
coherence and vertical structure properties of the scattering layer. For example,
the height of the layer is found in the limits of the integral, the phase of the
surface, while not equal to the phase of the coherence, is contained therein, and
finally the structure function f (z) influences the coherence in both amplitude
and phase.
Special cases of the structure function are often used in practice: for example,
constant scattering amplitude or an exponential to more accurately model wave
extinction effects in the layer (Treuhaft, 1996). Here we first develop a general
theory of volume decorrelation based on arbitrary structure functions, and then
specialize our discussion to these important special cases. The approach we use
is to expand the bounded function f (z) in a Fourier–Legendre series, as follows.
We first normalize the range of the integral in the numerator by a further change
of variable, as shown in equation (5.47):
hv zL = 2z
hv − 1
1
f (z )e iβz z
dz −−−−−−− −−−→ f (zL ) eiβz zL dzL (5.47)
0 −1
where the first few Legendre polynomials of interest to us are given explicitly
as shown in equation (5.49). Figure 5.14 shows plots of these functions for
hv = 10 m. The first represents a simple uniform distribution, while the sec-
ond includes linear variations, then quadratic and so on, with the higher-order
226 Introduction to radar interferometry
Legendre polynomials
10
P0
9 P1
P2
8 P3
P4
7 P5
P6
6
Height
5
0
0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2
Fig. 5.14 The Legendre polynomials from
zeroth to sixth order Relative density
P0 (z) = 1
P1 (z) = z
1 2
P2 (z) = 3z − 1
2
1 3
P3 (z) = 5z − 3z
2 (5.49)
1 4
P4 (z) = 35z − 30z 2 + 3
8
1 5
P5 (z) = 63z − 70z 3 + 15z
8
1
P6 (z) = 231z 6 − 315z 4 + 105z 2 − 5
16
The numerator and denominator of the general expression for coherence can
now be written as shown in equation (5.50):
hv hv i βz hv 1 βz hv
f (z )eiβz z dz = e 2 (1 + f (zL ))ei 2
zL
dzL
0 2 −1
(5.50)
hv hv 1
f (z)dz = (1 + f (zL ))dzL
0 2 −1
5.2 Sources of interferometric decorrelation 227
-1 i βz2hv zL
iβz z0 i βz2hv −1 (1 + f (zL ))e dzL
γ̃ = e e -1
−1 (1 + f (zL ))dzL
-1 5 (5.51)
−1 1 + an Pn (z L ) eiβv zL dzL
n
= eiβz z0 eiβv
-1 5
−1 1 + an Pn (zL ) dzL
n
By expanding the series and collecting terms, this equation can be rewritten in
simplified form, as shown in equation (5.52):
-1 -1 -1
(1 + a0 ) −1eiβv zL dzL + a1 −1 P1 (zL )eiβv zL dzL + a2 −1 P2 (zL )eiβv zL dzL + · · ·
γ̃ = e iβz z0 iβv
e -1 -1 -1
(1 + a0 ) −1 dzL + a1 −1 P1 (zL )dzL + a2 −1 P2 (zL )dzL + · · ·
(1 + a0 )f0 + a1 f1 + a2 f2 + ..an fn
= eiβz z0 eiβv
(1 + a0 )
ai
= eiβz z0 eiβv (f0 + a10 f1 + a20 f2 + · · · ) ai0 = (5.52)
1 + a0
n βz eβz nz n−1 n(n − 1)z n−2 (−1)n n!
z e dz = z −
n
+ · · · (5.53)
β β β2 βn
As an example, equation (5.54) shows detailed calculation of the first two terms
in the series. The first, corresponding to the zeroth-order Legendre polynomial,
just yields a SINC function, while the first order linear polynomial gives a
slightly more complicated function.
1
1 1 eiβv z 1 sin βv
f0 = eiβv z dz = = eiβv z − e−iβv z =
2 −1 iβv −1 i2βv βv
1 1
1 eiβv z 1
f1 = zeiβv z dz = z−
2 −1 iβv iβv −1 (5.54)
1 1
= eiβv z + e−iβv z − eiβv z − e−iβv z
iβv (iβv )2
sin βv cos βv
=i −
βv2 βv
228 Introduction to radar interferometry
For reference we give the explicit form of all these functions up to sixth order
in equation (5.55).
sin βv
fo =
βv
sin βv cos βv
f1 = i −
βv2 βv
3 cos βv 6 − 3βv2 1
f2 = − + sin βv
βv2 2βv3 2βv
30 − 5βv2 3 30 − 15βv2 3
f3 = i + cos βv − + 2 sin βv
2vβv3 2βv 2βv4 2βv
35(βv − 6)
2 15
f4 = 4
− 2 cos βv
2βv 2βv
35(βv4 − 12βv2 + 24) 30(2 − βv2 ) 3
+ + + sin βv
8βv5 8βv3 8βv
−2βv4 + 210βv2 − 1890 30βv4 − 840βv2 + 1890
f5 = i cos βv + sin βv
βv5 βv6
42βv4 − 2520βv2 + 20790)
f6 = cos βv
βv6
6
2βv − 420βv4 + 9450βv2 − 20790)
+ sin βv (5.55)
βv7
1. The even index functions are real while the odd are purely imaginary.
We note also that the unknown coefficients an are all real.
2. The functions vary only with the single parameter β v , which itself
is defined from the product of two parameters: height hv , and the
interferometric wavenumber β z .
Graphs of these functions are shown in Figure 5.15. We see that the first
is a ‘SINC’ relation between coherence and increasing height–baseline prod-
uct. This is the expected functional relationship for scattering by a uniform
layer. However, we see that as the height–baseline product increases so the
other functions become more important. We can conclude, therefore, that the
interferometric coherence is sensitive to changes in the structure function f (z).
There are two special cases of structure function of particular importance due
to their widespread use in the literature. We now turn to consider these in more
detail.
0.4
0.2
–0.2
–0.4
0 0.5 1 1.5 2 2.5 3
Fig. 5.15 Coherence basis functions for
kz*h/2 Legendre expansion
There are two important features of this model. Firstly it shows that volume
scattering provides a phase offset, given in this case by half the volume height.
Hence in the presence of volume scattering the interferometric phase no longer
represents the true surface position but is offset by a bias. For vegetated terrain
this is called vegetation bias, and provides an error source in the use of radar
interferometry for true surface topography mapping. We see that the only way
to minimize this effect is to employ small baselines so that the product β z hv
remains small. However, this reduces the sensitivity of the interferometer and
is difficult to sustain over forested terrain, where hv can reach up to 50 m or
more. Note that we cannot simply use the phase of the interferogram to estimate
volume height hv , since the total phase involves addition of an unknown phase
shift due to the lower bound of the volume (z0 ). Only if we can provide an
estimate of this lower bound can we then use the phase to estimate height. We
shall see in Chapter 8 how to provide such an estimate.
The second key feature of the SINC model is that the coherence amplitude
falls with increasing height and hence the phase variance increases with hv . Note
that there is no effect of the lower bounding surface on coherence amplitude
(assuming range spectral filtering has been employed), and in principle we can
therefore use an estimate of measured coherence amplitude to estimate height
(for a known baseline). In particular, for short baselines we can expand the SINC
function in a series and obtain a useful direct height estimate from coherence,
as shown in equation (5.57):
"
sin x x2 24(1 − |γ̃ |)
x1⇒ ≈1− ⇒ hv ≈ (5.57)
x 6 βz2
230 Introduction to radar interferometry
z = z0 + hv
o
2sez
f (z) = e coso
z = z0
Fig. 5.16 Exponential structure function
5.2 Sources of interferometric decorrelation 231
equation (5.59):
- hv 2σe z
This example illustrates the important new idea that the coherence in general
depends not only on the volume depth hv but also on the shape of the vertical
structure function. The exponential model essentially allows a one-parameter
model for variation of structure (via σ e ). High extinction implies an effective
scattering layer at the top of the volume, such as a high-elevated forest canopy,
for example. Figure 5.17 shows an example of how the coherence varies for an
exponential profile with varying σ e and depth hv . We have selected a baseline
corresponding to β z = 0.1567 (which corresponds to a zero of the SINC model
at 40 m), and considered a 45-degree angle of incidence. Note that for zero
extinction we again obtain, as a special case, the SINC model.
However, as extinction increases so the coherence increases for a given
height. This arises physically as the effective scattering volume is being
squeezed into a smaller and smaller region close to the top of the volume
as extinction is increased. This can be confirmed by plotting the phase of the
coherence. We first define the fractional phase centre height Pc from the interfer-
ometric phase φ as Pc = hvφβz . Figure 5.18 shows how Pc varies with extinction
0.9
0.8
0.7
0.6
Coherence
0 dB/m
0.5
0.125 dB/m
0.4 0.125 dB/m
0.2
0.1
0
0 5 10 15 20 25 30 35 40
Fig. 5.17 Volume decorrelation versus
height (m) height for various extinctions
232 Introduction to radar interferometry
0.9
0.8
0.7
0.6
Coherence
0.5
0.4
0 dB/m
0.3 0.125 dB/m
0.2 0.125 dB/m
0.75 dB/m
0.1
0
0 5 10 15 20 25 30 35 40
Fig. 5.18 Phase centre height versus height
for various extinctions Height (m)
90 1
120 60
0.8
0.6
150 30
0.4
0.2
180 0 dB/m
0
0.75 dB/m
0.125 dB/m
210 330
and height. Note that for the special case of zero extinction we obtain a phase
centre halfway up the layer as expected in the SINC model. However, as the
extinction increases we see that the phase centre moves towards the top of the
layer, approaching Pc = 1 in the limit of infinite extinction.
We have seen from Figures 5.17 and 5.18 that the coherence amplitude and
phase variations are linked. Indeed, it is instructive to visualize both at the same
time by employing the coherence diagram representation (see Appendix 3).
Figure 5.19 shows how the complex coherence varies inside the unit circle
in the complex coherence plane for three extinction values and layer depths
varying from 0 to 40 m (using the same parameters used in Figures 5.17 and
5.18). Here we see that the SINC model spirals quickly to the origin—to zero
5.2 Sources of interferometric decorrelation 233
Combining these two ideas leads us to the following general procedure for gen-
erating a vector interferogram. We first project the complex scattering vectors
k 1 and k 2 , measured at ends 1 and 2 of the baseline, onto the conjugate of
6.1 Vector formulation of radar interferometry 235
transpose, called [2 ], where the subscript 2 now refers to the number of spatial
positions used.
= >
k 1 ∗T T11 12
[2 ] = . k1 k ∗T =
k2 2 ∗T
12 T22
w∗T
1 12 w 2
⇒ γ̃ (w1 , w2 ) =
w∗T
1 T11 w 1 . w∗T
2 T22 w 2
(6.7)
From here we can then start by considering single baseline polarimetric inter-
ferometry (SBPI) [2 ], which involves measurements at only two separated
spatial/temporal positions. Here the T ii matrices can still have polarimetric
dimension N = 1, 2, 3 or 4, but the matrix 12 contains additional information
about the variation of interferometric coherence. This procedure can then be
easily extended to multiple baselines (and frequencies), as shown in equation
(6.8) (Ferro-Famil, 2001, 2008).
Considering the special but important case of radar backscatter N = 3, there
are several special cases of unitary change of base U 3 to be distinguished.
Equation (6.9) is expressed in the linear Pauli basis (see equation
√ (6.3)). If we
wish to convert to the standard linear lexicographic base of HH, 2HV and VV,
6.1 Vector formulation of radar interferometry 237
We can then focus attention on a general change of wave base states: movement
of a reference point P over the surface of the Poincaré sphere. If the spherical
triangle coordinates of P are α w and δ w (see Figure 1.12), then the general
unitary matrix for use in equation (6.9) takes the form shown in equation (6.11):
cos αw
P= ⇒ [U3 ] = [U3L ][ULP3 ]
sin αw eiδw
√
cos2 αw − 2 cos αw sin αw e−iδw sin2 αw e−i2δw
√ √
⇒ [U3L ] = 2 cos αw sin αw eiδw cos2 αw − sin2 αw − 2 cos αw sin αw e−iδw
√
sin αw e w
2 i2δ 2 cos αw sin αw e w
iδ cos αw
2
(6.11)
For example, if we want to convert from the linear H,V √ basis to left and right
circular L,R so as to obtain the matrix in the basis LL, 2LR,RR, then we would
set α w = π /4, δ w = π /2 in equation (6.11) and obtain the composite change of
basis matrix shown in equation (6.12). This can then be used in equation (6.9)
to express all matrices in the circular basis.
√
0 1 1 √1 −1
2i √ 1 1 √0 √0 1 i
1 1
U3circ = 2i √0 2i √ 0 0 2 = √ 2i 0 0
2 2 2
−1 2i 1 1 −1 0 0 −1 i
(6.12)
Note, however, that equation (6.11) does not represent the most general unitary
transformation. It has only two free parameters, while the general 3 × 3 unitary
matrix has nine (see Appendix 2). These extra degrees of freedom are gener-
ated by combining triplets of orthogonal scattering mechanisms, as used in the
eigenvector decomposition of [T ], for example (see Section 4.22). Starting with
an arbitrary mechanism, we have five degrees of freedom; the second then must
be orthogonal to the first, and so has 5 − 2 = 3 parameters. The third must then
be orthogonal to the first two, and so has 5 − 4 = 1 parameter, producing nine
in total. If these are then combined into a special unitary matrix (det(U 3 ) = 1)
this reduces to eight parameters. The set of general unitary transformations is
then governed by the eight Gell–Mann matrices, as shown in equation (6.13)
(see Appendix 2) (Cloude, 1995b; Ferro-Famil, 2000).
det(U3 )=1
[U3 ] = w1 w2 w3 −→ [U3 ] = exp(iφn.G) (6.13)
The key conclusion is that arbitrary unitary matrices in the change of base
formulation—equation (6.9)—can be given a clear physical interpretation in
terms of triplets of polarimetric scattering mechanisms. The change of wave
polarisation base then forms only a subset of these matrices through equation
238 Polarimetric interferometry
1 ! ∗T
L
[Z] = uj uj (6.14)
L
j=1
0.9
0.8
0.7
0.6
Coherence
0.5
0.4
0.3
0.2
0.1
0
5 10 15 20 25 30 35 40 45 50
Fig. 6.2 Estimation of coherence triplet of
Number of looks (L) equation (6.15) versus number of looks L
0.9
0.8
Scattering entropy
0.7
0.6
0.5
0.4
0 10 20 30 40 50
Fig. 6.3 Estimated scattering entropy versus
number of looks for dipole cloud Number of looks (L)
w∗T
1 12 w 2
γ̃ w1 , w2 = → max |γ̃ | (6.16)
w1 w2
w∗T ∗T
1 T11 w 1 .w 2 T22 w 2
This represents the coherence obtained when polarisation w1 is used at the first
and w2 at the second end of the baseline. In general, therefore, it combines both
interferometric and polarimetric contributions to coherence. Our objective is
to find the extreme values of the magnitude of this function. There is a slight
complication in that equation (6.16) is a complex function, and so we must
6.2 Coherence optimization 241
decide whether to maximize the absolute value, the phase, or real and imaginary
parts. These various choices open up different forms of optimization as we now
consider.
This yields a set of coupled equations for the unknown vectors w1 and w2 and
the Lagrange multipliers λ1 and λ2 . There are now two important options for
solution of these equations. In the most general case we allow w1 and w2 to
be different and so allow full polarisation diversity. In this case we can find a
solution to the coupled equations as a pair of eigenvalue problems, as shown
in equation (6.18):
−1 2
w1 =w2
T22 ∗T −1 ∗
12 T11 12 w 2 = λ1 λ2 w 2 K1 w1 = νw1 = γopt w1
−→ λ1 = λ2 = γ̃opt ⇒ 2
−1 −1 ∗T
T11 12 T22 12 w1 = λ1 λ∗2 w1 K2 w2 = νw2 = γopt w2
(6.18)
This shows that the optimum scattering mechanisms can be obtained from
eigenvalue equations involving composite products of the elements of [y ].
Furthermore, the two Lagrange multipliers are complex but equal. To show this
we left multiply the top derivative equation in (6.17) by w∗T
1 and the lower by
w∗T
2 , and use the normalization condition on the Hermitian forms on the right-
hand side of L in equation (6.17) to show that λ1 =λ2 . Note that for backscatter
problems there are then three optimum values corresponding to the square mod-
uli of the three eigenvalues of K. Hence the 3 × 3 matrices K 1 and K 2 have the
same non-negative real eigenvalue spectra in the range 0 ≤ υ3 ≤ υ2 ≤ υ1 ≤ 1,
but different and non-orthogonal eigenvectors, as neither K 1 nor K 2 are gen-
erally Hermitian or unitary matrices. The maximum coherence is given by the
square root of the largest eigenvalue ν 1 with corresponding scattering mecha-
nisms given by the eigenvectors. Note that the optimum complex coherence can
be found by first calculating the eigenvectors w1 and w2 from equation (6.18),
phase normalizing using equation (6.4), and then using them directly in equa-
tion (6.16). Alternatively we may calculate the optimum directly by solving
the following generalized eigenvalue problem, obtained by a straightforward
242 Polarimetric interferometry
rewriting of the derivative equations in (6.17) and using the fact that λ1 = λ2 ,
as shown in equation (6.19):
0 12 w1 T11 0 w1 ∗T
= λ ⇒ γ̃opt = λmax e−i(arg(w1 w2 ))
∗T
12 0 w 2 0 T 22 w 2
(6.19)
where λmax is the eigenvalue with maximum modulus. The advantage of this
formulation is that it scales naturally to the multi-baseline case, as first shown
in Neumann (2008). When M-tracks are available we must use the generalized
coherency matrix [M ] shown in equation (6.8). In this case the unconstrained
optimization problem can be formulated by generalising the Lagrangian to a
sum of numerators with constrained denominators, leading to the generalization
of equation (6.19), as shown in equation (6.20):
!
M !
M M
!
L= w∗T
i ij w j + λ w∗T
i Tii w i − 1
i=1 j=i+1 i=1
0 12 . . . 1M w1 T11 0 ... 0 w1
∗T 0 . . . 2M w2 0 T22 ... 0 w2
12
⇒ . .. .. .. .. = λ .. .. .. .. ..
.. . . . . . . . . .
∗T
1M ∗T
2M . . . 0 w M 0 0 . . . TMM wM
(6.20)
Note that here λmax now corresponds to the weighted sum of optimized
coherence moduli—a type of average across all the baselines.
max|λ(φ)| w∗T
opt [H ]w opt
−→ wopt ⇒ γopt = (6.22)
w∗T
opt [T ]w opt
We see again an eigenvalue equation, but this time based on averages of the
sub-matrices. However, one drawback of this approach is, as shown on the
right-hand side of equation (6.22), that it maximizes only the real part of the
eigenvalue and so represents a phase-sensitive optimization. Hence this finds
only a local maximum, and to find the true global optima we need to introduce
a free phase parameter exp(iφ). By then repeating the optimization in equation
(6.22) for different values of φ we can then obtain the global maxima. The
general procedure for constrained optimization is then summarized in the lower
portion of equation (6.22).
The optimization process is then formally equivalent to a mathematical prop-
erty called the numerical radius of an N × N complex matrix [A] (Murnaghan,
1932; Li, 1994; He, 1997; Mengi, 2005). This itself is defined from the field of
values F(A), as defined in equation (6.23):
2 ? ? 3
F(A) = w∗T [A]w, w ∈ C N , ?w? = 1 (6.23)
The numerical radius is then the radius of the smallest circle that contains the
field of values, as defined in equation (6.24):
In our context we can formally relate the numerical radius to coherence opti-
mization by first generating the constrained form of the Lagrangian function
LC , as shown in equation (6.25):
LC = w∗T 12 w + λ w∗T T w − 1 (6.25)
This is almost in the form required, and needs only a pre-whitening transfor-
mation to remove the polarisation dependence of the constraint equation (the
factor T). Using the square root of T as a basis transformation, we then obtain
the following modified form:
T −1 w ⇒ Lc = w∗T −2
12 T − 2 wn + λ w∗T
1 1
wn = n T n wn − 1 (6.26)
There are many theorems and algorithms in the mathematics literature dealing
with the concept of numerical radius (Murnaghan, 1932; Li, 1994; He, 1997;
Mengi, 2005). Unfortunately there are no general analytical solutions avail-
able, but various numerical iterative algorithms have been proposed—one of
which involves exactly the phase transformation and repeated eigensolution
approach represented in equation (6.22). Furthermore, it leads to a third impor-
tant approach to optimization, based not on coherence amplitude but on phase
difference or coherence separation, as we now consider.
90 1
120 60
0.8
0.6
150 30
0.4
φ
0.2
180 0
Coherence region
210 330
240 300
Fig. 6.4 Definition of the coherence region
of a polarimetric interferometric matrix 2 270
in equation (6.22). For each value of φ this eigenvalue equation yields the
extreme values (through the maximum and minimum eigenvalues) of the real
part of the coherence. For each of these eigenvalues there corresponds an eigen-
vector, which can be used to estimate a corresponding complex coherence, as
shown in equation (6.30). These two coherences then define two points on
the boundary of the coherence region, as shown schematically in Figure 6.4.
Here we show an example elliptical coherence region (see equation (6.31)), and
show, for a specified value of φ, how we obtain two samples of the boundary.
By varying φ in the range 0 ≤ φ ≤ π we can then reconstruct the whole bound-
ary. This gives us a systematic way to visualize the boundary for an arbitrary
coherency matrix 2 .
[H ] = 12 12 eiφ + ∗T 12 e−iφ
[T ]−1 [H ]w = λw
[T ] = 12 (T11 + T22 )
w∗T
max 12 w max (6.30)
γmax (φ) = w∗T T w
λmax ,wmax
−→ max max
⇒
λmin ,wmin
w∗T w
−→
γmin (φ) =
min 12 min
∗T
wmin T wmin
As a more specific example, consider again the 2 matrix shown in equa-
tion (6.15). This has a region dominated by three points: the three diagonal
complex values of 12 . Figure 6.5 shows the corresponding triangular region
for this matrix defined by the three diagonal values as vertices. As we vary
the polarisation over all possible mechanisms the interferometric coherences
will always be contained within this triangle. The boundaries of the region
therefore define the various optimum values. In this case the constrained and
unconstrained maximum amplitudes are equal to the white vertex. The separa-
tion optimization yields the white and black vertices, with a maximum phase
difference of π /2.
6.2 Coherence optimization 247
90 1
120 60
0.8
0.6
150 30
0.4
0.2
180 0
210 330
240 300
Fig. 6.5 Example coherence region for
270 example matrix in equation (6.15)
This matrix can, by Schur’s theorem (see Appendix 1), always be written in
terms of an upper diagonal form and a unitary matrix [U ], as shown on the
right-hand side of equation (6.31). Here λ1 and λ2 are the eigenvalues of A.
It then follows that the field of values of A is an ellipse with two foci given
by λ1 and λ2 and minor axis length |δ|. The corresponding major axis length
is given by |λ1 − λ2 |2 + |δ|2 . For the special case that δ = 0 (in which case
the matrix A is termed ‘normal’—it can be diagonalized by a unitary trans-
formation) we obtain a linear field of values, varying along a line stretching
between the two eigenvalues. We shall see in Chapter 7 that such a limiting
case plays an important role in the description of mixed surface and volume
scattering.
248 Polarimetric interferometry
(6.32)
In SVD we allow different vectors at either end of the baseline and obtain
the singular values s1 and so on, which we have shown in equation (6.20)
correspond directly to optimum coherences. In the Schur approach, however,
we constrain the decomposition to a single unitary matrix (corresponding to
equal vectors at either end of the baseline), which leads to an upper triangular
3 × 3 matrix, also shown in equation (6.34). However, following the Schur
6.2 Coherence optimization 249
By setting the pair x,y equal to 1,2, 1,3, and 2,3 in equation (6.32), we then pro-
vide a link between the general Schur decomposition and projection approach.
Note that for each pairing we can directly calculate the shape of the coher-
ence region analytically, since it is always elliptical, with two foci λx and λy ,
2 2
minor axis length δxy and major axis length λx − λy + δxy . As a sim-
ple example, consider again the matrix shown in equation (6.15), with a region
illustrated in Figure 6.5. In this case the matrix has the following simple
form:
0.9 0 0
π
= 0 0.6ei 4 0 (6.36)
π
0 0 0.4ei 2
Here the three subspace regions reduce to line segments joining the pairs of
eigenvalues, as shown in Figure 6.5. We now return to the issue of numerical
bias in the context of these new coherence optimization techniques.
1! 1! 1!
L L L
T̂11 = k 1i .k ∗T
1i , T̂22 = k 2i .k ∗T
2i ,
ˆ 12 =
k 1i .k ∗T
2i (6.37)
L L L
i=1 i=1 i=1
250 Polarimetric interferometry
The question now is, what is the influence of the number of samples (‘looks’ in
radar imaging terms) L on the coherence optimization algorithms? As L → ∞
we should obtain the true matrices (since in this limit the matrices converge to
their correct values), but for small L we can expect overestimation of coher-
ence and hence distortion of the coherence region. To analytically study the
effects of L on optimum coherence estimation is a difficult task (see Touzi,
1999; Lopez-Martinez, 2005), and here we therefore employ some illustrative
numerical simulations based on use of Monte Carlo simulations (see Appendix
3 for details) to illustrate the nature of the problems involved, and to form
some general conclusions about bias effects in optimization methods. We again
make use of the random volume scattering example shown in equation (6.15),
and this time use the simulated data to estimate the matrices using equation
(6.37) before applying the various constrained and unconstrained optimization
algorithms.
Figure 6.6 shows the results of applying the unconstrained optimization algo-
rithm to the estimated matrices (again each point formed from an average of
256 coherence estimates). We note that the bias issues are more severe than for
the standard coherence estimation (shown dashed, for reference, in Figure 6.6).
This reflects the underlying higher dimensionality of the general unconstrained
optimization process. Direct coherence estimation implies that we have a priori
knowledge of the w vectors—in this case just the Pauli scattering vectors—
and so can project before we undertake the coherence estimation to obtain the
improved convergence shown in the dashed lines in Figure 6.6. However, for
the unconstrained optimization process we not only have to estimate coherence
values but also do not know the projection vectors themselves. These too must
be estimated from the data. Hence we need to estimate a larger number of param-
eters from the data itself. This increased dimensionality requires an increased
number of looks for convergence. This provides a qualitative explanation of
the increased bias seen in Figure 6.6.
0.9
0.8
Coherence
0.7
0.6
0.5
0.4
90 1
120 60
0.8
0.6
150 30
0.4
0.2
180 0
210 330
We note also that the triplet of optimum states (the three eigenvalues of
[K]) have different bias issues. The first and second optima are overestimated
for small L, but the smallest optimum value is actually underestimated. This
correlated bias behaviour between eigenvalues means that the apparent dynamic
range of coherence variability with polarisation is overestimated for small L.
We see that it takes in excess of L = 50 looks for the estimate bias to settle
down, but note again slow convergence even beyond this point.
This overestimation of dynamic range is also apparent in the estimation of
the coherence region. If we employ coherency matrix estimates for L = 6 looks
and then L = 50 looks we obtain typical region estimates as shown in Figure 6.7.
Here, in grey we show the estimated boundary region for L = 6, and note its
overestimation compared to the true region (shown as the black triangle). For L
= 50 we see a much better estimate (shown in black), more accurately reflecting
the bounds of the true coherence region, and hence prociding a better estimate
of the true dynamic range.
We now turn to consider how the physical structure of surface and volume
scattering controls the size and shape of the coherence region.
The coherence of
7 surface and volume
scattering
In this chapter we investigate in more detail the shape and structure of the
limiting form of the coherence region for vector surface and volume scatter-
ing problems. In Section 6.2.3.1 the coherence region is defined as the region
in the complex plane bounding the variation of interferometric coherence with
polarisation. Here we extend this idea to define a related concept: the coherence
loci, defined as the curves traced out by variation of interferometric coherence
with physical parameters of a scattering model (Papathanassiou, 2001; Cloude,
2003). Our objective is to relate the coherence loci as a limiting form of the
coherence region (as the number of looks tends to infinity) in order to estab-
lish strategies in Chapter 8 for using polarimetric interferometry for physical
parameter estimation.
We begin by looking at simple models of surface and volume-only scattering.
We then consider extension of these ideas to multilayer media which, impor-
tantly, will allow us to consider combinations of surface and volume effects.
We then look in detail at two important models widely used in the literature for
interpreting coherence diagram: the random-volume-over-ground, or RVOG
(Treuhaft, 1996, 2000a; Papathanassiou, 2001; Cloude, 2003), which is closely
related to an interferometric version of the water cloud model IWCM (Askne,
1997, 2003, 2007) (see Section 3.5.1); and the oriented-volume-over-ground,
or OVOG (Treuhaft, 1999; Cloude, 2000a). Both these models are character-
ized by having a small number of independent physical parameters, often fewer
than observables in the scattered field, so enabling consideration of methods
for estimation of these parameters from data (see Chapter 8).
As we shall see, both of these models (RVOG and OVOG) make assumptions
about the vertical variation of scattering in the layered media (through the struc-
ture function), which naturally leads us to consider a more general approach
termed coherence tomography (Cloude, 2006b, 2007a) that permits arbitrary
structure function and allows an efficient parameterization of the dependence
of coherence of changes in structure.
In general terms we note that the coherence loci must somehow be related to
variation of the vertical structure function f (z) with polarisation. For example,
if the scatterers in a scene do not change relative amplitude as w changes,
then the structure function, whatever shape it has, will be constant, and the
coherence will be constant with polarisation, so yielding a point coherence
loci. This point can then be stretched to a radial line by adding polarisation-
dependent temporal, or SNR, decorrelation, but the underlying physics will
be determined by a point in the complex diagram. We now investigate this
relationship between coherence loci shape and structure function variations in
more detail, for surface and volume scattering scenarios.
7.1 Coherence loci for surface scattering 253
The first issue we face is in defining surface scattering in the context of inter-
ferometry. By definition, surface scattering occurs at a discontinuity between
two media, and hence a good model for its vertical structure function would f (z) = (z – zo)
be a Dirac delta function located at the interface between the media, as shown
schematically in Figure 7.1, where the surface is clearly located at position
z = zo . However, we have seen that in microwave remote sensing of natural z
surfaces there is always some penetration of the wave into the lower medium,
Fig. 7.1 Idealized vertical structure function
depending on the effective dielectric constant (see Section 3.1.1.1), and hence for surface scattering
it is of interest to consider the circumstances under which this delta function
assumption is supported. In the context of interferometry, what is important
is not so much the absolute penetration depth but its value scaled to βz —the
vertical sensitivity of the interferometer. Hence by combining the definition of
penetration depth δp (equation (3.12)) with the baseline dependence of verti-
cal wavenumber βz , we can obtain a relationship between effective complex
material constant εr = ε − iε and baseline geometry, as shown in equation
(7.1). Here we set a threshold of 0.1 radians for the product, as this represents
a typical interferometric phase shift due to wave penetration of only 5◦ and a
maximum volume decorrelation of only 0.998. These are within the bounds
of estimation error for typical interferometer geometries, and thus represent a
somewhat arbitrary but realistic threshold for the delta function assumption to
apply.
β z δp
< 0.1
2
√ √ √
4π θ λ ε 2θ ε 2B⊥ ε
⇒ . = ≈ < 0.1
λ sin θ 2π ε sin θ ε Ro sin θ ε
√
ε Ro sin θ
⇒ < (7.1)
ε 20B⊥
If this inequality is satisfied (and we further assume that range spectral filtering
has been employed to remove any baseline decorrelation (see Section 5.1.1.1)),
the interferometric coherence can then be estimated as shown in equation (7.2),
where we have also, for the moment, ignored any temporal or SNR decorrelation
terms.
- hv
δ (z) eiβz z dz
γ̂ = e iβz zo o
- hv = eiβz zo = eiφo (7.2)
o δ (z) dz
Here we see that [K] is just the identity matrix, indicating an interferometric
coherence of unity for all polarisation states. This is just a consequence of
the fact that although the absolute level of the delta function in the structure
function f (z) varies with changes in polarisation, its position remains fixed at
the surface boundary z = zo , and hence the coherence remains the same for all
polarisations. In this case the corresponding coherence region shrinks to a point
on the circumference of the unit circle (the angular position of which depends
on βz zo ).
In practice this result must be extended to include variations in SNR with
polarisation. The backscatter power from smooth surfaces can be very low,
especially in the crosspolarised channel, and this variation will be apparent as
a polarisation-dependent coherence, as shown in equation (7.4):
w∗T [T11 ] w 1
s/n = ⇒ γsnr (w) = (7.4)
n 1 + w∗T [Tn 11 ]w
90 1 90 1
120 60 120 60
0.8 0.8
0.6 0.6
150 30 150 30
0.4 0.4
0.2 0.2
180 0 180 0
similarly to the polarisation with minimum SNR. The whole line position (and
scaled line length) is dictated by the temporal decorrelation γt . At extremes,
when γt = 1, the line maximum can approach the unit circle at the point φo as
shown. When γt = 0, the whole line reduces to a point at the origin.
Z We note that [K] is again a multiple of the identity matrix, indicating that
Top of layer zo + hv
the coherence does not change with polarisation and, as we found for surface
scattering, the coherence loci reduces to a point in the coherence diagram.
However, unlike surface scattering, the point does not lie on the unit circle.
Instead it lies within the circle at a point determined by the complex volume
f (z) = fv(z) decorrelation caused by the structure function f (z) (see Section 5.2.4). The
integral factors I1 and I2 in equation (7.8) can be expressed in terms of the
Surface position zo Legendre expansion of the structure function, as shown in equation (7.9). The
key consequence of the random symmetry assumption is that the Legendre
coefficients are independent of polarisation, and so the structure function, which
Fig. 7.3 Schematic representation of an can be arbitrary, as shown in Figure 7.3, must remain invariant to changes in
arbitrary vertical structure function polarisation.
hv
I2 = eiβz zo f (z )e iβz z
dz
0 I2
⇒ γ̃ =
I1
hv
I1 = f (z )dz
0
βz hv (1 + a0 )f0 + a1 f1 + a2 f2 + · · · + an fn
= eiβz z0 ei 2 (7.9)
(1 + a0 )
The effect of signal-to-noise ratio will be similar to that found for surfaces; that
is, to provide a polarisation-sensitive radial shift of the coherence towards the
origin. However, since volume scattering is generally more depolarising than
surface scattering, the variation of scattered power with polarisation will be
less, and hence SNR effects less important, than they are for surfaces. On the
other hand, temporal decorrelation can be much more important for volume
scattering, especially in vegetation applications, due to its susceptibility to
wind-driven motion on short time-scales. To further complicate issues, the
effects of temporal changes may not be uniform across the structure function.
For example, wind-blown motion may affect the top of the vegetation layer
more than the lower regions. To accommodate this we can modify the coherence
integrals to include in the numerator (I2 ) a new temporal structure function g(z),
as shown in equation (7.10):
hv
I2 = eiβz zo g(z )f (z )e iβz z
dz
0 I2
⇒ γ̃ = (7.10)
hv
I1
I1 = f (z )dz
0
The function g(z) will vary between 0 and 1, being zero in regions of maximum
change and 1 for zero change. In terms of the Legendre expansion, g(z) will of
course have its own expansion coefficients, which in general will be different
from those of f (z), and hence the effect of temporal decorrelation must formally
be evaluated as a product of Legendre series in the numerator I2 . In the simplest
case (and the one most often used in the literature) we can assume that g(z) =
γt —a constant function with height, in which case the overall coherence can
7.2 Coherence loci for random volume scattering 257
90 1 90 1
120 60 120 60
0.8 0.8
0.6 0.6
150 30 150
0.4 0.4 30
f = f0 + fb
0.2 0.2
180 0 180 0
the addition of the unknown phase of the surface φ(zo ). This acts essentially as
a new physical parameter (the location of the bottom of the layer) that must also
be estimated from the data. Hence we now have three unknowns and only two
observations. Nonetheless, this concept of reducing the number of parameters
required to describe the structure function so as to better match the number of
observations is an important one. We shall see in the case of layered media
that it leads us to a convenient solution for estimation of physical parameters
from data.
I2
γ̃ (w) = γSNR w γt
I1
- hv
e− cos θo
2σe hv 2σe z
e cos θo eiβz z dz
= γSNR w γt 0
- hv
e− cos θo
2σe hv 2σe z
0 e cos θo dz
2σe eiφ(zo ) hv 2σe z
= γSNR w γt / θ
eiβz z e cos θo dz
cos θo (e e v
2σ h cos o − 1)
0
2σe
p=
cos θ
p ep1 hv − 1
= γSNR w γt where 1 p = p + iβz
p1 ephv − 1
βz = 4π θ ≈ 4π Bn
λ sin θ λH tan θ
= γSNR w γt γ̃v (7.12)
This is a function of only one physical parameter: the mean extinction. The
coherence loci for this model has a simple form, forming a semicircle in the
coherence plane, starting on the unit circle at the top phase reference for infi-
nite extinction (set to zero phase for convenience in Figure 7.5), and moving
7.2 Coherence loci for random volume scattering 259
0.6
150 30
0.4
0.2
180 0
210 330
towards the origin (zero coherence) for zero extinction, as shown by the line in
Figure 7.5.
This represents one of the simplest possible models for volume decorrela-
tion, and when combined with a measured coherence (shown as the point in
Figure 7.5), provides a means for estimation of the mean extinction in the vol-
ume from the coherence magnitude of a single polarisation interferogram. (We
saw in the water cloud model (WMC), in Section 3.5.1, that this extinction is
often directly related to the water content mv , and hence we can often use this
extinction as a proxy for water content.) However, the assumption of an infi-
nite depth restricts this approach to applications where layer thickness greatly
exceeds wave penetration depth. Important examples are thick land-ice (Dall,
2003; Sharma, 2007) and high-frequency penetration of vegetation and snow;
but in general terms the assumptions of this model are not robust, and layer
thicknesses are often small compared to penetration, so that scattering from the
underlying bounding surface cannot be ignored. Treatment of these scenarios
will require multilayer scattering models to be developed in the next section, but
first we consider an important variation on the exponential structure function
assumption: the case of oriented volume scattering.
We can then also relate the [K] optimization matrices from equation (6.14) in
the two representations, as shown in equation (7.17):
−1 −1 ∗T
KT = T11 12 T11 12
−1 −1 −1 −1 −1 −1
⇒ KC = (ULP3 T11 ULP3 )(ULP3 12 ULP3 )(ULP3 T11 ULP3 )(ULP3 ∗T
12 ULP3 )
−1 −1
⇒ KC = ULP3 KT ULP3 ↔ KT = ULP3 KC ULP3 (7.17)
We make one further assumption: that the eigenpolarisations x and y are ortho-
gonal linear states; but we do allow for a mismatch in the angle between these
states and the radar coordinates by an angle ψ. We can now obtain an expression
for the coherency matrix [T11 ] = [T22 ] and interferometry matrix [12 ] for an
oriented volume extending from z = z0 to z = z0 + hv as vector volume
integrals shown in equations (7.18) and (7.19) (see equation (4.69):
8 .
hv (σx +σy )z
iφ(zo ) ∗
[12 ] = e R(2ψ) e iβz z
e cos θo P (τ ) TP(τ )dz R(−2ψ) (7.18)
0
8 .
hv (σx +σy )z
∗
[T11 ] = R (2ψ) e cos θo P (τ ) TP(τ )dz R (−2ψ) (7.19)
0
7.2 Coherence loci for random volume scattering 261
where for clarity we have dropped the brackets around matrices and define the
following terms:
1 0 0
R(ψ) = 0 cos ψ sin ψ (7.20)
0 − sin ψ cos ψ
cosh τ sinh τ 0 t11 t12 0
P(τ )TP τ ∗ = sinh τ cosh τ 0 t12 ∗ t22 0
0 0 1 0 0 t33
∗ ∗
cosh τ sinh τ 0
× sinh τ ∗ cosh τ ∗ 0 (7.21)
0 0 1
z
τ = νz = κy − κx − iβo nx − ny (7.22)
cos θo
where κx,y are the amplitude extinction coefficients of the volume for x and y
polarisations. Note that if we cannot align the radar coordinates with the vol-
ume then the matrix term R(2ψ), which multiplies the whole matrix integral
expression inside the brackets, causes a coherent mixing of terms that is dif-
ficult to interpret. We will show that the polarimetric optimizer automatically
aligns the radar to the oriented volume. This result follows from knowledge of
the explicit form of the matrix [K], which for this problem enables direct calcu-
lation of its eigenvalues and eigenvectors, and hence optimization parameters
in closed form.
h h ∗
h
I1 = e2σx z dz I2 = e2i(βy −βx )z dz I3 = e2(σx +σy )z dz
0 0 0
h h h ∗
I4 = e2σy z dz I5 = eiβz z e2σx z dz I6 = eiβz z e2i(βy −βx )z dz
0 0 0
h h h
−2i(βy∗ −βx )z
I7 = eiβz z e(σx +σy )z dz I8 = eiβz z e dz I9 = eiβz z e2σy z dz
0 0 0
(7.25)
Hence the first part of the optimization matrix [KC ] has the following form,
which is diagonal if I4 I6 − I2 I9 = I8 I1 − I2∗ I5 = 0.
c33 I4 0 −c13 I2
eiφ(z0 ) c11 I5 0 c13 I6
−1 f
C11 12 = 0 0 0 c22 I7 0
f c22 I3
∗ I
c13 0 c33 I9
∗ I∗
−c13 0 c11 I1 8
2
(7.26)
From equation (7.25) we can easily show that both equations are satisfied for
arbitrary medium parameters, as we have the following relationships:
h ∗
I 4 I6 = e2σy z eiβz z e2i(βy −βx )z dz = I2 I9
0
(7.27)
h
2σx z iβz −2i(βy∗ −βx )z
I8 I1 = e e e dz = I2∗ I5
0
∗ I I )
(c11 c33 I4 I5 − c13 c13 2 8 I4 I5 2σx eiφ(zo ) h 2σx z
γ̃1 = ∗ = = f (σx ) = eiβz z e cos θo dz
(c11 c33 − c13 c13 )I1 I4 I1 I4 cos θo (e2σx hv / cos θo − 1) 0
1 0 0
eigenvectors
KC ←→ w1 = 0 w2 = 0 w3 = 1 (7.29)
0 1 0
1 0 1
eigenvectors 1 1
R(ψ)KT R(−ψ) ←→ w1 = √ − cos 2ψ w2 = sin 2ψ w3 = √ − cos 2ψ
2 2
sin 2ψ cos 2ψ − sin 2ψ
(7.30)
We see that the eigenvectors of [K] are orthogonal for oriented volume scat-
tering, and also contain information about the orientation of the medium’s
eigenpolarisations, while the eigenvalues are related to the coherences for the
corresponding eigenwave extinctions. As expected from physical arguments, g~yy > g~xy > g~xx –> Ο
the highest (lowest) coherence is obtained for the polarisation with the high-
est (lowest) extinction. The higher the extinction, the less penetration into the
volume and hence the lower the effective volume decorrelation. This connec-
tion between the structure function and optimum coherences is summarized in Fig. 7.6 Summary of physical interpretation
Figure 7.6. Importantly, this provides our first example of an extended, non- of optimum coherence triplet for oriented
trivial coherence loci. Figure 7.7 shows how the loci, defined by the three volume scattering
optimum coherence points, can be constructed. Note the following important
features:
1) The three optima are rank ordered in coherence amplitude (radius inside
the unit circle), with the highest coherence associated with the highest
extinction propagation channel.
2) The three optima are also ranked in phase. If we take the phase of the
bottom of the layer (z = zo ) as reference (shown as the black point
90 1 90 1
120 60 120 60
0.8 0.8
0.6 0.6
150 30 150
0.4 0.4 30
0.2 0.2
180 0 180 0
90 1
120 60
0.8
0.6
150 30
0.4
0.2
180 0
210 330
on the unit circle in Figure 7.7) then the lowest coherence amplitude is
always closest in phase to this point, followed by the crosspolarised chan-
nel, and then the highest coherence point always has the highest phase
shift.
3) The loci extension is caused by physical changes in volume decorrelation
with polarisation, and so far we have ignored effects due to temporal and
SNR decorrelation. These can be included in the analysis by allowing
radial shifts, so extending the loci towards the origin, forming a new loci
bounded by the dotted lines on the right side of Figure 7.7.
As an important special case we can consider the coherence region and loci for a
semi-infinite oriented volume. Figure 7.8 shows the semicircular loci developed
for the infinite random volume (equation (7.14)), and superimposed we show the
three optima for the oriented volume. The coherence corresponding in Figure
7.6 to the polarisation with maximum extinction has the highest coherence and
position closest to the top of the layer (which now corresponds to zero phase
in this diagram). The minimum extinction lies furthest in phase, while the
crosspolarised channel has a coherence intermediate in phase and amplitude
between the maximum and minimum extinctions.
From a parameter estimation perspective the finite slab oriented volume (OV)
model is interesting. We see that we have six observables (the phase and ampli-
tude of coherence in three optimum polarisations), and yet we have only four
unknowns (the layer depth, two values of extinction and phase of the bottom
of the layer). This is a good starting point for developing robust algorithms for
estimating these parameters from experimental data (see Chapter 8). However,
there is one major problem to be addressed, in that we have so far ignored
the presence of a ‘hard’ boundary behind the layer. In radar applications this
is often a soil or rock surface beneath vegetation or snow/ice layer, and as we
shall see, this considerably distorts the coherence loci shape. To consider such
7.3 The coherence loci for a two-layer scattering model 265
issues we need to extend our approach to consider the coherence region for
two-layer scattering problems.
symmetric backscatter matrix for reciprocal media (see Section 3.4.3). Impor-
tantly, even though these are second order, because the two can be specular
(forward scattering with angle of incidence equal to angle of reflection), such
second-order scattering contributions can be as large as, or larger than, the
direct backscatter mechanisms themselves. Hence we cannot ignore them for
a full development of the coherence loci for this two-layer problem. As we are
considering coherent scattering, we must also concern ourselves with the phase
of such second-order mechanisms.
P
Figure 7.11 shows a schematic representation of typical second-order scatter-
ing from a volume scattering element P at height hp . We are concerned with the
S
P phase of the second-order signal from P (not the direct return) compared to R.
This will depend on the range difference relative to a direct surface return com-
hp Q R
ing from the surface at point R as shown. Point Q, being the specular reflection
Q R point on the surface, defines a triangle PQR as shown. It follows from the geom-
etry of this triangle, shown enlarged on the right-hand side, that the distance
Fig. 7.11 Ray geometry explanation of phase QS + SP = 2SR. Combining this with recognition that PQ represents a wave
centre location for dihedral scattering front of the incident plane wave, which by definition is an equi-range contour,
leads us to conclude that the range difference between P and R is zero for all
heights hp . This important result implies that in backscatter geometry, the phase
difference between the second-order and direct surface scattering components
is always zero.
There is, however, one further complication to be considered. If we now
extend our analysis to consider across-track radar interferometry, then the
second-order scattering behaves very differently for single and dual transmitter
modes. In dual transmitter mode (which includes repeat-pass interferometry as
a special case) we transmit and receive separately from each end of the base-
line B. From each position the second-order scattering effects (being exactly
in backscatter geometry) behave as shown in Figure 7.11, and consequently
the phase difference across the baseline will be exp(iβz zo )—the same as for
the direct surface scattering component from point R. Therefore, in dual trans-
mitter mode the second-order scattering effects behave like an effective and
additional surface component, with a phase corresponding to the underlying
surface position and a coherence of unity, indicating zero volume decorrela-
tion, even though the distributed volume is involved in the scattering mechanism
(see equation (7.31)).
Note that this all follows from the special geometry of the triangle PQR in
Figure 7.11. In terms of modifications of the structure function, we note that
the second-order effects add an additional delta function contribution at z = zo .
Note also that from a polarimetric point of view, the second-order components
have a polarisation signature (scattering α > π/4) very different from direct
surface scatter (α < π/4). Hence, for dual transmitter modes we conclude
that second-order surface–volume interactions and direct surface scattering are
separable in polarimetry but not in interferometry.
Now consider the case of single transmit/dual receive interferometry. By
definition this is a configuration that involves a small but non-zero bistatic
scattering angle δθ, as shown schematically in Figure 7.12. For the end of
the baseline operating with both transmit and receive modes, the second-order
7.3 The coherence loci for a two-layer scattering model 267
P P
hp hp
Q R Q R
scattering will again have an exact backscatter geometry, and the effective
phase centre lies on the surface at R (shown by the dash line in Figure 7.12).
However, for the end of the baseline with a receive-only mode we obtain a
height-dependent phase shift due to the small bistatic geometry. As the volume
scattering elements are distributed over a range of heights from 0 to hv , there
will consequently be a volume decorrelation effect in this mode. To analyse
further, consider separately the geometry of the surface–volume and volume–
surface contributions shown in Figure 7.12. On the left we show the ray path
for the surface–volume term. The scattering into a small bistatic angle δθ gives
rise to a height-dependent phase given by equation (7.32). (Note the factor of
2π in place of 4π , because we are considering the single transmitter case.)
2π sin θ δθ 2π sin θ δθ
φSV (hp ) = hp = sin2 θβz hp φVS (hp ) = − hp
λ λ
= − sin2 θβz hp (7.32)
uniform profile ranging from –hv to hv from which we can calculate the decor-
relation caused by second-order scattering interactions for single transmitter
configurations as a SINC function, as shown in equation (7.33):
- hv i sin2 θβ z -0
dz + −hv ei sin θβz z dz
2
iφ(zo ) 0
e z
γ̃sv1TX =e -h
2 0 v dz
- hv 2
θβz z
iφ(zo ) −hv
ei sin dz
=e
2hv
sin(sin2 θβz hv )
= eiφ(zo ) (7.33)
sin2 θβz hv
Note that the mean phase centre still lies on the surface at z = zo (as it does
for the dual transmitter case). However, we now have a radial shift towards the
origin of the coherence diagram, with the amplitude of the coherence decreasing
with increasing depth of layer 1.
We have seen in the above that second-order interactions cause some com-
plexity in the analysis of coherent scattering from two-layer media. This is
further compounded when we realize that in theory there is an infinite cas-
cade of such higher-order surface–volume interactions to be considered. For
example, on the far right of Figure 7.10 is a typical third-order interaction of
surface–volume–surface scattering. Fortunately, for random media, such higher
order interactions are usually very small compared to first and second order.
Physically we can see that this arises because the backscatter level of such high
order interactions is determined by a cascaded product of small quantities. The
surface–volume–surface interaction, for example, involves the product of two
surface reflections, which will be small, all multiplied by the backscatter rather
than specular forward scatter from particles in the volume, which will generally
be smaller. Add to this the increased effective propagation distance inside the
lossy material of layer 1, and we can see, at least qualitatively, how interactions
higher than second order can often in practice be ignored (although there are
some notable exceptions such as scattering from complex man-made structures
such as bridges and buildings). Nonetheless, for most remote sensing applica-
tions this result will allow us to justify the use of simpler second-order models
in deriving the coherence loci for such problems.
First, however, for completeness we consider in detail the coherence prop-
erties of third-order surface–volume–surface contributions. For simplicity we
P
consider only the dual transmitter configuration. Figure 7.14 shows a schematic
of the geometry concerned. Again we are interested, for interferometry, in how
hp the effective range difference to a point P across the baseline varies as the height
u
R
hp is changed. In this case we see that the third-order effect has the same range
Q
variation as scattering from a virtual point Pm located a distance hp beneath the
surface, as shown. Therefore, even in the dual transmitter case we now obtain
–hp a height-dependent phase that will cause (complex) volume decorrelation and
a loss of coherence amplitude combined with a (negative) phase bias.
Pm We can calculate the level of this coherence by realizing that it has a cor-
responding structure function f3 (z) that is extended below the surface into the
Fig. 7.14 Ray diagram to locate phase centre range 0 < z < −hv . From this structure function we can then calculate the
for third-order scattering contribution corresponding complex coherence from an integral of the general form shown
7.3 The coherence loci for a two-layer scattering model 269
z z z
z = z 0 + hv z = z0 + hv z = z0 + hv
z = z0 – hv z = z0 – hv z = z0 – hv
z z
z = z0 – hv z = z0 – hv
2TX
fsvs (z) fsv1TX (z)
f(z) f(z)
in equation (7.34). Note that the superscripted N represents the number of trans-
mitters (1 or 2), and we have extended the range of the coherence integral from
−hv to hv to accommodate calculations for the virtual scattering points.
- hv NTX
iφ(zo ) −h
f (z )eiβz z dz
γ̃iNTX = e - v i (7.34)
hv NTX
−hv fi (z )dz
Figure 7.15 summarizes the form of the structure functions for various com-
ponents of the two-layer scattering problem. In the first diagram we show the
direct volume term, fv2TX , which has some arbitrary shape, bounded by the
surface and the top height of the layer. Next we show the corresponding direct
surface return, fs2TX , which is a simple delta function located on the surface. For
the two-transmitter scenario this delta function also matches the second-order
surface–volume scattering contribution, fsv2TX . However, the single-transmitter
case has a uniform structure function extending across the full range, as shown
in the lower diagram fsv1TX . Finally we show the structure function for third-
2TX . This has a totally negative extent, but again can lead
order scattering fsvs
to volume decorrelation with an arbitrary structure function bounded by the
surface and top layer.
Of more general interest is the way in which these components combine to
provide the overall coherence variation for a two-layer problem. To see this we
need to incorporate all the mechanisms into a single generalized polarimetric
interferometric formulation. The starting point is to define the coherent scat-
tering vector k as the sum of contributions at ends 1 and 2 of the baseline, as
shown in equation (7.35). In these expressions, [P] is the vector propagation
270 The coherence of surface and volume scattering
We can then combine these vectors with averaging (which removes all
cross-products between mechanisms—an expression of independent scatter-
ing mechanisms) to express the generalized polarimetry and interferometry as
the sum of component matrices, as shown in equation (7.35). From these matri-
ces we can then determine a general expression for the observed coherence as
a function of polarisation, as shown in equation (7.36):
w∗T 12 w w∗T (v + Ps s Ps∗T + Psv sv Psv
∗T + · · · )w
γ̃ w = ∗T = ∗T
w T11 w w (Tv + Ps Ts Ps∗T + Psv Tsv Psv
∗T + · · · )w
m0v (w)γ̃v (w) + ps (w)m0s (w)γ̃s (w) + psv (w)m0sv (w)γ̃sv (w) + · · ·
=
m0v (w) + ps (w)m0s (w) + psv (w)m0sv (w) + · · ·
(7.36)
Here we have rewritten each term as a product of three components: its nor-
malized radar cross section mo , its coherence contribution γ̃ , and a propagation
factor p that attenuates each contribution according to the propagation paths
involved. Note that we can write this expression as the product of total radar
backscatter times total observed coherence, as shown in equation (7.37):
γ̃ (w)m0 (w) = m0v (w)γ̃v (w) + ps (w)m0s (w)γ̃s (w)
+ psv (w)m0sv (w)γ̃sv (w) + · · · (7.37)
This gives us a procedure for deriving the coherence loci for two-layer prob-
lems, by which we first need to calculate the three elements for each mechanism,
and then combine them as shown in equation (7.36). As mentioned earlier,
for lossy layers this series converges quickly, and we need not consider the
complexity of scattering higher than second order. To illustrate this we now
develop three particular forms of this model that are widely used in the lit-
erature: a coherent two-layer version of the water cloud model (IWCM),
the closely related random-volume-over-ground or RVOG model, and the
oriented-volume-over-ground or OVOG model.
Significantly, only the parameter µ changes with polarisation. This is the ratio
of effective (sum of direct and second-order) surface-to-volume scattering. We
can develop an explicit form for µ as shown in equation (7.40):
This can be further simplified by realizing that the propagation factors for direct
surface and second-order interactions are the same and given by equation (7.41):
cos θ
mov w = 1 − e−2σe hv sec θ mv w (7.42)
2σe
From this we see that the cross-section can vary in the range mHH +VV ≤
mv (w) ≤ smHH +VV . For a dipole cloud, for example, we note there is only
a 3-dB variation of RCS with polarisation. The factor µ, however, can have a
much wider dynamic range than this, as shown in equation (7.44):
Here we see that the numerator depends directly on the variation of ‘effective’
surface scattering with polarisation. We can assume that this has reflec-
tion symmetry and a corresponding variation with polarisation, as shown in
7.4 Important special cases: RVOG, IWCM and OVOG 273
equation (7.45):
Here the subscripted es denotes the effective surface components. The polarisa-
tions that maximize and minimize the µ ratio will be of interest in establishing
the coherence loci for RVOG. To find these we need to solve the following eigen-
value equation arising from optimization of the µ ratio, as shown in equation
(7.46) (see Section 4.2.2.2):
w∗T [Tes ] w
max ⇒ [Tv ]−1 [Tes ] wopt = λwopt (7.46)
w∗T [Tv ] w
90 1
120 60
0.8
g~vo 0.6
150 30
0.4
0.2
180 0if (z )
e o
210 330
240 300
Fig. 7.17 Coherence loci for RVOG model 270
7.4 Important special cases: RVOG, IWCM and OVOG 275
0.9 f = 0º
0.8
0.7 f = 90º
0.6
Coherence
0.5
0.4
0.3
0.2
0.1 f = 180º
0
–30 –20 –10 0 10 20 30
Fig. 7.18 Variation of coherence magnitude
Mu (dB) in RVOG model for various phase angles
We see that for small volume phase shifts φ the minimum is given by small
µ < −30 dB, but as the phase increases so that Re(γ̃vo ) > |γ̃vo |2 or φ >
cos−1 (|γ̃vo |), then the minimum coherence occurs for higher values of µ. Figure
7.19 shows how the µ for minimum coherence changes for the example shown
in Figure 7.18. Note that for phase angles up to cos−1 (0.8) = 37◦ the minimum
276 The coherence of surface and volume scattering
–5
–10
Mu (dB)
–15
–20
–25
–30
0 20 40 60 80 100 120 140 160 180
Fig. 7.19 µ required for minimum coherence
in RVOG model versus phase Phase angle (degrees)
coherence is given by the volume only (µ = 0) point. However, for phase shifts
above this the minimum occurs for a mixture of surface and volume scattering.
For high phase shifts the minimum occurs for an almost equal mixture of surface
and volume scattering.
However, we have seen that the RVOG assumes an exponential structure
function in the volume, and hence the coherence amplitude and phase are not
independent quantities. In fact the phase centre for the volume-only component
of RVOG must lie between halfway and the top of layer 1 (see Figure 5.18).
In other words, for RVOG it follows that the phase of the interferogram for the
volume-only channel φ ≥ βz2hv . Coupled to this is the realization that for RVOG
the coherence amplitude can never be less than the zero extinction limit; that
is, |γ̃vo | ≥ sinc( βz2hv ).
This then begs the question as to whether the minimum coherence for the
RVOG model can ever be given by µ = 0. For this to be possible the following
inequality must apply:
βz hv βz hv βz hv
φmin < cos−1 (|γ̃min |) ⇒ < cos−1 sinc 0≤ ≤π
2 2 2
(7.50)
However, this inequality is never satisfied for the RVOG model, and hence we
conclude that for RVOG the minimum coherence is never given by the µ = 0
volume scattering channel. There is always some mixture of surface and volume
scattering that combines to produce a minimum. This rather surprising result
follows from the assumed form of the structure function (an exponential). This
exposes a weakness of the RVOG model: that its assumption of a uniform layer
with a simple extinction profile ignores any variations due to vertical structure
in volume scattering. We now turn to consider, in more detail, how structure
variations are dealt with in RVOG.
7.4 Important special cases: RVOG, IWCM and OVOG 277
(1 + a0 )f0 + a1 f1 + a2 f2 + · · · + an fn
= eiβz z0 eikv
(1 + a0 )
For this more general SVOG model the relationship between coherence
amplitude and phase is not as restrictive as it is for RVOG, and so the min-
imum coherence can indeed be the volume-only coherence, under the general
condition that the phase φ < cos−1 (|γ̃vo |).
The classical RVOG model can be made to accommodate changes in struc-
ture by varying the extinction over a sufficiently wide range. However, this
gives rise to a structural ambiguity in RVOG, in that we can fit RVOG to
non-RVOG situations simply by adjusting the model parameters. A simple
example of this ambiguity is shown in Figure 7.21. Here we show a simple
case of vertical structure: a layering of the volume, with scattering from a thin
top layer, and surface reflection from a position separated from this volume
by a gap.
The coherence model for this three-layer problem can be written as shown
in equation (7.52). This still has a linear coherence loci, but the coherence
amplitude of the volume-only channel will be small, while its phase will be
z = hv
Constant
number hv, 1
density
z=0
1 >> 2
z = hv
hv, 2
Vertical
structure z = hc
large. It is possible to fit RVOG to this structure (as shown in the upper portion
of Figure 7.21). However, to explain the combination of high coherence and
large phase offset we need to employ an effective extinction for the medium
that is much larger than the actual value.
µ(w)
γ̃ (w) = e iβz zo
e iβz hc
γ̃vo + 1 − eiβz hc γ̃vo
1 + µ(w)
- hv 2σe hv
hc e cos θ eiβz z dz
γ̃vo = - hv 2σe hv
(7.52)
hc e cos θ dz
There are two forms of this model that deserve special attention. In the first
we consider its form for short spatial/long temporal baselines and high radar
frequency, where volume decorrelation is very small and temporal effects for
both surface and volume dominate. We shall call this the high-frequency or
HF-IWCM. In the second form—more closely linked to RVOG—we consider
longer spatial/shorter temporal baselines and low radar frequency when tem-
poral effects are mixed with significant levels of volume decorrelation. This we
call the low-frequency or LF-IWCM.
In the first case the βz value is small and the mean wave extinction is
high (because of the high-frequency approximation), and the volume decor-
relation can therefore be approximated by a unitary phase shift, as shown in
7.4 Important special cases: RVOG, IWCM and OVOG 279
equation (7.54):
2σe
p1 =
p1 (ep2 hv − 1) cos θo
γ̃vo =
p2 (ep1 hv − 1)
2σe
p2 = + iβz
cos θo
lim σe →∞ p1 iβz hv 1
−→ γ̃vo ≈ e = eiβz hv (7.54)
p2 1 + i βz 2σ
cos θo
e
θo
−i βz 2σ
cos θo iβz hv − cos
≈e e e iβz hv
=e 2σe
γ̃ (w)
mv (w) cos θ cos θo
γ̃tv 2σe 1 − e−2σe hv sec θ eiβz (hv − 2σe ) + e−2σe hv sec θ γ̃es m0es (w)
=
mv (w) cos θ
2σe 1 − e−2σe hv sec θ + e−2σe hv sec θ m0es (w)
cos θo
γ̃tv eiβz (hv − 2σe ) + γ̃es µ(w)
= (7.56)
1 + µ(w)
Again we see that the parameter µ—the surface-to-volume scattering ratio—is
important in determining the coherence loci for this model. The loci in this
case is a triangle, with two vertices on the unit circle and one at the origin
(γtv = γtes = 0), as shown in Figure 7.23. An important simplified case of
HF-IWCM arises in the limit γtv = 0 and γes = 1. This occurs in vegeta-
tion problems, for example, when wind-driven temporal change destroys the
coherence completely from layer 1, while the underlying surface scattering
contributions show no change. In this case we move along the radial line OP in
Figure 7.23, for which the coherence loci depends entirely on µ, as shown in
equation (7.57):
µ(w)
γ̃ (w) = eiφ(zo ) (7.57)
1 + µ(w)
Note that this has same form of coherence variation with polarisation as SNR
decorrelation, as demonstrated in Figure 7.24 (compare this with Figure 5.12).
In the second form of this model—the LF-IWCM—we consider the limit
of larger spatial/lower temporal baselines and low radar frequency combined
with volume-only temporal decorrelation γtv due, for example, to short-term
wind-blown effects in vegetation cover. This model is closely related to RVOG,
since the volume coherence can no longer be approximated by a phase shift.
280 The coherence of surface and volume scattering
90 1
120 60
0.8
0.6
150 Q 30 if (z ) ib (1– cosuo )
0.4 e o e z 2se
0.2
180 0 0
P eif (zo)
210 330
240 300
Fig. 7.23 Coherence loci for the IWCM
model 270
0.9
0.8
0.7
0.6
Coherence
0.5
0.4
0.3
0.2
0.1
0
Fig. 7.24 Coherence variation with surface- –30 –20 –10 0 10 20 30
to-volume ratio (µ) for the IWCM model Mu (dB)
Here we see that the coherence loci remains a non-radial line segment, but that
the fixed point representing the volume is shifted towards the origin by the scale
7.4 Important special cases: RVOG, IWCM and OVOG 281
90 1
120 60
0.8
g~v
0.6
150 30
0.4
gtg^v 0.2
if (zo)
180 0 e
210 330
240 300
Fig. 7.25 Coherence loci for the RVOG
270 model with temporal decorrelation
factor γtv . This amounts to a rotation and stretch of the coherence line about
the surface topography point, as shown in Figure 7.25 (Papathanassiou, 2003).
In conclusion, we have shown that the coherence loci for a two–layer random
volume over ground scattering problem is a line segment in the complex plane.
This line is radial when temporal effects dominate, and shifts to a non-radial
phase variant line as volume decorrelation becomes more important. In both
cases we note two important features. The first is the importance of the ratio µ,
being the ratio of effective surface-to-volume scattering. The second key point
is that the line passes through the unit circle at the surface phase point. We shall
see later that this provides us with a method for correcting for vegetation bias
in radar interferometry by line fitting and by finding this intersection.
Here both volume and surface scattering have polarimetric coherency matrices
with reflection rather than azimuthal symmetry, and the surface-to-volume scat-
tering ratios include the effects of propagation distortion. The dynamic range
of µ can be developed using a modification of the procedure used in equation
(7.46), as shown in equation (7.60):
90 1
120 60
0.8
c = f ( x, y, hv)
0.6
150 30
0.4
0.2
180 0
210 330
240 300
Fig. 7.27 Coherence loci for the OVOG
model 270
7.4 Important special cases: RVOG, IWCM and OVOG 283
is constrained to move up and down its own straight line inside the unit cir-
cle, depending on the µ ratio. The three lines for the eigenstates define the
boundary of this region, coming to a focus at the ground topography point
φ, and having a spread ψ, as shown in Figure 7.27. Importantly, this spread
depends on the differential extinction in the volume layer and not on the µ val-
ues or surface topography. In the special case that ψ = 0 we again obtain the
random-volume-over-ground RVOG model. We also note that the OVOG model
requires that the crosspolarised XY coherence line lies between the XX and
YY lines.
The coherence loci for this problem can be obtained as an extension of the OV
region, as shown in Figure 7.29. Here again we see a region formed by three
bounding lines for the eigenpolarisations emanating from the surface point, with
variations along each line given by the fraction of surface-to-volume scattering,
F(w). Again the RVUG or random volume version of this model is obtained
as a limiting case when the extinctions become equal and we obtain the single
line coherence region, as shown on the right-hand side of Figure 7.29.
Coherence loci for infinite volume Coherence loci for infinite volume
90 1 90 1
120 60 120 60
0.8 0.8
0.6 0.6
150 30 150
0.4 0.4 30
0.2 0.2
180 0 180 0
240 300 240 300 Fig. 7.29 Coherence loci for the OVUG
270 270 model (left) and OVOG model (right)
Parameter estimation
8 using polarimetric
interferometry
from the top layer. Finally, we can use multiple polarisations and least squares
correction to phase bias. We now consider each of these in turn.
In the simplest case the phase of the surface component can be estimated
directly from the coherence, as shown in equation (8.1):
&
'
R0 sinθ λ ' 1 − γw 2
σh ≈ σφ σφ ≤ ( 2
s
(8.2)
B⊥ 4π
2L γws
where the Cramer–Rao bound (minimum value) of the phase variance σφ for
a given number of looks L is given on the right-hand side of equation (8.2)
(Seymour, 1994). This, of course, resorts to conventional interferometry in the
limiting case of bare surfaces (hv = 0), but in other cases is made complicated
by the phenomenon of phase bias.
We have seen that the volume coherence for layer 1 is complex and hence
contributes a phase offset from the surface itself, and so equation (8.1) will
generally overestimate the surface position for RVOG and OVOG, and under-
estimate it for OVUG. Hence it is clear that polarisation ws should be chosen
so as to minimize this bias and optimize the accuracy of the estimate. A second
objective must be to choose ws to also maximize the SNR, so as to minimize the
decorrelation due to noise and hence optimize the precision. In RVOG, OVOG
and OVUG these requirements amount to maximization of µ, the surface-to-
volume ratio. The best polarisation to use would therefore be that given by
equation (7.47) or (7.60). However, as we have no a priori knowledge of the
separate volume and surface component coherency matrices, we cannot make
direct use of this equation. Instead we must employ an indirect solution, as we
now investigate.
The problem is that there is no single polarisation that always maximizes µ.
For a bare surface (hv = 0) at low frequencies when the Bragg surface scatter-
ing model is valid, a good choice is VV, as HH has less scattered power and
thus lower SNR, and HV is zero. A better choice still is HH+VV, as the zero
polarimetric phase difference leads to an even better SNR than VV. Clearly, the
optimum would weight the Bragg scattering matrix elements to maximize the
SNR. For higher frequencies and rougher surfaces the depolarisation increases,
the difference between HH and VV becomes less, and polarisation plays less of
a role in bare surface parameter estimation. In general, therefore, an unbiased
coherence optimizer would provide a suitably adaptive solution. For bare sur-
faces the constrained optimizer of equation (6.22) would provide a good choice
(as shown again in equation (8.3)). Note, however, that in order to implement
such an optimizer we require access to full scattering matrix data, so as to be
286 Parameter estimation using polarimetric interferometry
ˆ 12 .
able to estimate the component matrices T̂11 , T̂22 ,
[ ˆ H] = 1 ˆ 12 eiφ +
ˆ ∗T e−iφ
−1 ˆ
[T̂ ] [H ]w = λ (φ) w
2 12
[T̂ ] = 1 T̂11 + T̂22
2
(8.3)
max|λ(φ)| w∗T ˆ
opt [H ]w opt
−→ wopt ⇒ γopt =
w∗T
opt [T̂ ]w opt
The above analysis breaks down, however, in case layer 1 has non-negligible
thickness. We can set a suitable threshold on the product of wavenumber and
layer thickness to estimate this breakpoint, such as βz hv < 0.1 (see equation
7.1), for the surface approximation to hold. If the product exceeds this threshold
then we require a different strategy to optimize estimation of surface topography
as follows.
When the layer thickness can no longer be ignored, we face complica-
tions arising not only from phase bias and increased volume decorrelation,
but a change in scattering mechanism. This arises especially when the dihedral
second-order scattering is dominant. In this case HH is often preferred to VV
(the opposite of the bare surface case), as it has a higher specular reflection
coefficient at the surface, and VV will in this case have a lower µ. By the
same reasoning, the Pauli choice HH–VV is sometimes selected in preference
to HH+VV, as this has an even higher µ than HH, due to the 180-degree polari-
metric phase shift that occurs in the case of a dominant second-order scattering
scenario.
A second problem also arises in the case that layer 1 is a random volume,
when it acts to depolarise the scattered wave with a high entropy. This implies
that the volume scattering ‘contaminates’ every polarisation vector w, and con-
sequently that it is impossible to find a candidate ws which contains surface-only
scattering. This means that the phase bias due to volume scattering in layer 1
will be present across the whole of polarisation space. The best we can do is
again try to select the ws with maximum µ. However, there is no longer a guar-
antee that the coherence amplitude optimizer of equation (8.3) will correspond
to the maximum µ (see the discussion in Section 7.4.2). While the optimizer
will still guarantee the highest coherence and hence the highest precision, it no
longer guarantees the highest accuracy, because of the presence of phase bias.
To proceed further we need to consider methods for the removal of this bias.
circle point from a line fit as follows. The idea is to use γ̃1 as a fixed point on
the line and relate γ̃2 by a scale factor F2 along the line towards the unit circle.
In this way the two coherences can be related as shown in equation (8.4). We
see that the desired ground topography point is embedded in these equations,
and we can solve for it directly.
γ̃1 γ̃2 − γ̃1 (1 − F2 )
⇒ eiφo = 0 ≤ F2 ≤ 1 (8.4)
γ̃2 = γ̃1 + F2 (eiφ − γ̃1 ) F2
Here we see that the phase term is obtained as a weighted average of the two
complex coherences. If F2 tends to unity then it represents the desired surface
point, and γ̃2 is taken as the solution. However, in general there will be phase
bias present in both channels (because of depolarisation in layer 1), and hence
this mixture formula is required to compensate for this bias. There remains a
problem in that to solve for the surface topography we first require an estimate
of the factor F2 . This can be obtained directly from the estimated coherences
γ̃ˆ1 and γ̃ˆ2 by forming the product eiφ e−iφ = 1, using equation (8.4) to obtain a
quadratic. Taking the root that makes F positive, we obtain the solution shown
in equation 8.5:
φ̂ = arg(γ̃ˆ2 − γ̃ˆ1 (1 − F2 )) 0 ≤ F2 ≤ 1
√
−B − B2 − 4AC (8.5)
AF2 + BF2 + C = 0 ⇒ F2 =
2
2A
A = |γ̃1 |2 − 1 B = 2Re((γ̃2 − γ̃1 ).γ̃1∗ ) C = |γ̃2 − γ̃1 |2
Note that unlike the simple phase algorithm (equation (8.1)), this requires
coherence estimates in both amplitude and phase, and hence is susceptible
to non-compensated errors in coherence such as those due to SNR or tempo-
ral effects. Also, of course, the estimates of coherence themselves have some
variance due to their stochastic nature and the finite number of looks L used
in the estimator (the coherence region). For stability of the phase estimate we
therefore need to ensure that F2 is as large as possible. (If F2 tends to zero so
that the two points are close together, then large errors can result.) Some care
is therefore required in the selection of w1 and w2 . There are three strategies
used in making an appropriate selection:
In physics-based selection we use our understanding of surface and volume
scattering to select the two channels. For example, the low-frequency Bragg
surface model predicts zero or (for higher-order forms of the model) very low
levels of crosspolarisation HV from a flat surface. On the other hand, volume
scattering from a cloud of anisotropic particles can yield high levels of HV (for
a dipole cloud only 3 dB below the maximum RCS). Hence HV is often chosen
as a candidate for the w1 channel. The w2 channel can likewise be selected on
the assumption that specular second-order scattering is dominant, and so HH
or HH–VV are good choices as they are likely to satisfy the requirement that
µ2 > µ1 . In summary, the direct physics-based approach produces allocations
of the two channels such as those shown by the two examples in equation (8.6):
γ̃1 = γ̃HV γ̃2 = γ̃HH −VV
or (8.6)
γ̃1 = γ̃HV γ̃2 = γ̃HH
288 Parameter estimation using polarimetric interferometry
Note that the lower option is particularly well suited to dual polarised active
systems that can transmit only linear H polarisation but receive H and V
components (see Chapter 9).
If more specific information is available about the scattering problem to
hand, based, for example, on direct EM scattering model simulations, then these
assignments can of course be modified as appropriate. Although such selections
may match very well a specific application or dataset, they are generally not
sufficiently robust for widespread application. For this reason we turn to a
second approach, based instead on phase optimization, that adapts itself to
variations in the data.
For each phase angle φ we find the distance between the maximum and min-
imum eigenvalues. By finding the maximum of this distance as a function of
φ we then automatically align the solution with the axis of the linear coher-
ence region and use these as an estimate of the coherence loci bounds. We then
find the polarisation scattering mechanisms wa and wb from the corresponding
eigenvectors. From these we can derive the two coherence γ̃max and γ̃min for
use in the topographic phase estimation algorithm of equation (8.5).
However, we face a potential problem with all these line-fitting ideas to ensure
that we always choose the correct rank ordering of the two coherences, remem-
bering that we must ensure that µ2 > µ1 to find the correct topography point.
In fact this is a general problem with all phase bias removal algorithms based on
the assumption of a linear coherence loci. By definition, a line intersects the unit
circle at two points (see Figure 8.2), one of which is the true topographic phase,
while the other represents a false solution obtained for the line fit technique by
exchanging the rank order of coherences. There are several ways to resolve this
8.1 Surface topography estimation 289
90 1
120 60
eifmax 0.8
0.6
150 30
g^ max 0.4
0.2
g–min
eifmin
180 0
210 330
240 300
270 Fig. 8.2 Unit circle surface phase ambiguity
T ∗T
ws = cos α sin αeiϕ 0 µ >µ γ̃1 = γ̃min if w∗T
max w v < w min w v
−→
s v
T γ̃ = γ̃ ∗T ∗T
wv = 0 0 1 1 max if wmax wv > wmin wv
(8.8)
Similarly, the orthogonal state wv should match the volume scattering (and have
a lower µ). In this way we can develop an algorithm for assigning the optimum
phase states in the correct rank order for the line fit algorithm, as shown on the
right-hand side of equation (8.8).
In the second interferometric approach we decide on rank ordering by
employing the phase difference between the calculated unit circle intersection
point and assumed low µ coherence. Knowing that layer 1 is above (or below)
the surface allows us to calculate these phase differences using the same clock-
wise (or anticlockwise) rotation around the coherence diagram. If we repeat
this for both rank permutations, then one of the phase shifts will be much larger
than the other and can be rejected (especially if it is known that the layer depth
is less than the π height of the interferometer). If we define φmax as the unit
circle phase estimate obtained when we propose γ̃max as the high µ channel
estimate, and likewise φmin when we propose γ̃max , then we can decide on the
290 Parameter estimation using polarimetric interferometry
So, for example, in Figure 8.2, if we assume that layer 1 is above the surface for
anticlockwise phase rotation, then max = 270◦ and min = 90◦ , and therefore
according to equation (8.9) we would select γ̃1 = γ̃max on the assumption that
the layer thickness should be less than the π height of the interferometer.
y y
Ri = y cos θ = =√ (8.10)
1 + tan2 θ 1 + M2
~
y=Im(g)
~ ~
Im(g) = MRe(g) + C
y = Mx+C
Ri
y
~
x = Re(g)
If we then make the simplest assumption that the unknown fluctuation errors
are the same in x and y (the modifications, if they are not, are straightfor-
ward, but complicate the notation), the function to be minimized now has the
following form:
! 1 !
R2i = (yi − C − Mxi )2 (8.11)
1 + M2
i i
This differs from the conventional LS approach only in the pre-multiplier, which
is itself a function of M . We can then find the stationary points of this function
by differentiation, to yield the following:
∂R 1 !
0= = −2(yi − C − Mxi )
∂C 1+M 2
i
∂R 1 !
0= = −2(yi − C − Mxi )xi (8.12)
∂M 1 + M2
i
2M !
− (yi − C − Mxi )2
(1 + M )
2 2
i
From the first term we obtain a direct solution for the estimate of C as
follows:
) *
1 ! !
Ĉ = yi − M̂ xi = ȳ − M̂ x̄ (8.13)
N
i i
Then, by substituting the first equation in the second and collecting terms we
obtain an estimate of M as the root of a quadratic, as shown in equation (8.14):
−c1 ± c12 − 4c2 c0
c2 M̂ 2 + c1 M̂ + c0 = 0 ⇒ M̂ = (8.14)
2c2
where the three coefficients are defined in terms of the data points as
follows:
!
c0 = − (xi − x̄)(yi − ȳ)
i
!@ A
c1 = (xi − x̄)2 − (yi − ȳ)2 (8.15)
i
!
c2 = (xi − x̄)(yi − ȳ)
i
This then provides us with a method for fitting a line to an arbitrary number
of polarisation channels. We can then use the estimates of M and C to find the
two unit circle intersection points as shown in Figure 8.4. These two points can
292 Parameter estimation using polarimetric interferometry
90 1
120 60
0.8
0.6
150 30
0.4
0.2
180 0
210 330
240 300
Fig. 8.4 Example of total least squares line
fit to complex coherence data 270
Clearly, this TLS approach will suffer from the same rank-ordering dichotomy
encountered in Section 8.1.2. Indeed, in the TLS case this problem is arguably
more serious, as there is no easy way to employ the physical selection process
described in equation (8.8). As there are multiple polarisation channels, and
not just two being used, it is difficult to decide which are surface- or volume-
dominated. For this reason the TLS approach is often combined with the phase
approach of equation (8.9) to decide which phase point to use as the topography
estimate.
Finally, we note that some care is required in the choice of N polarisations to
ensure that there is sufficient diversity of µ along the line segment in the com-
plex plane. For this reason, typical selections involve the three Pauli channels
(HH+VV, HH–VV and HV), augmented by the linear channels HH and VV,
as well as the three optimum states from either constrained or unconstrained
algorithms to produce a sample set in the region N = 8–12 for estimation.
azimuthal symmetry. While this may be a valid assumption for many applica-
tions, it is violated for an important class of problems when layer 1 displays
reflection scattering symmetry and behaves as an oriented volume. In this case
we have seen that the coherence loci is formed by a fan of three lines, emanating
from the unit circle topography point (see Figure 7.27). In theory, therefore, we
cannot strictly apply the above algorithms to the OVOG model. However, there
are two important classes of OVOG applications that deserve special mention
(Ballester-Berman, 2005, 2007; Lopez-Sanchez, 2007).
The first involves applications with only weak differential extinction, in
which case the fan angle ψ (see Figure 7.27) is small and the OVOG region
approaches a straight line, or at the other extreme, high differential extinction
combined with small minimum extinction. This latter scenario is very important
in that it can lead to a wide dynamic range in µ, with the low extinction channel
dominated by surface scattering and the high extinction by volume scattering.
In this case the line fit approach of TLS or phase optimization still provides a
good, if approximate, solution to surface topography estimation, as shown in
Figure 8.5. However, in all OVOG cases it must be realized that there is an
additional source of error due to the separation of volume terms in the complex
plane, and this can lead to large errors if extra care is not taken to make full
use of the µ spectrum for the problem. In this case, therefore, it is good to
use either the TLS with a wide diversity of polarisations, or the coherence
separation optimization technique to ensure that the maximum and minimum
µ values are being fully exploited.
Another class of OV problems of interest are for layers of effective infinite
depth, when we can assume that µ = 0 in all polarisation channels. In this
limit we essentially obtain the oriented volume or OV problem as a limiting
case of OVOG. We have seen that the coherence loci for this problem is a
semicircle (Figure 7.8), and we can devise a simple algorithm for top surface
phase estimation based on a circle fit to the data, as follows. The first point to
note is that the circle must intersect the unit circle and the origin, and hence has
90 1
120 60
0.8
0.6
150 30
0.4
0.2
Large m
180 0
210 330
240 300
Fig. 8.5 Line fit for topography estimation
270 under the OVOG model
294 Parameter estimation using polarimetric interferometry
0.6
150 30
0.4
c = po + iqo
q 0.2
⇒ f = tan−1 ( o )
po
180 c 0
210 330
240 300
Fig. 8.6 Topography estimation for the OV
model 270
a fixed radius of 0.25. The second point is that the topographic phase is simply
related to the coordinates of the centre C of the circle as shown in Figure 8.6. By
combining these two observations we can set a linear least squares formulation
for the two unknown coordinates of C as shown in equation (8.17):
(p − po )2 + (q − qo )2 = r 2
r 2 = po2 + qo2 = 0.25
2 2
pxx qxx pxx + qxx
p
⇒ 2 pxy qxy . o = pxy
2 + q2
xy (8.17)
qo 2 + q2
pyy qyy pyy yy
⇒ [A]x = b
⇒ x̂ = ([A]T [A])−1 [A]T b
Here we make use of the real and imaginary parts of the coherence in the three
eigenpolarisation combinations XX, XY and YY to fit the best-constrained
circle to the data. This then allows us to estimate the top surface position,
even though we are assuming there is no scattering from this interface and only
volume scattering is occurring. This is useful when either the top surface is very
smooth or there is a small dielectric contrast between free space and layer 1.
Note that this technique does not work if µ > 0 in any channel, as this has
the effect of pulling the coherences off the circle and towards the topography
point. In this case we must resort to the approximation used in Figure 8.5.
We have seen that there are several possibilities for using a priori assumptions
about the coherence loci for the two-layer problem, to devise algorithms for
estimation of the surface topography and hence to effect phase bias removal. We
now turn to consider a similar approach to the estimation of a second important
physical parameter: the height of the top layer.
8.2 Estimation of height hv 295
arg(γ̃wV e−iφ̂ )
ĥv = (8.19)
βz
Note that φ̂ can be estimated either from the data itself or from some external
source such as a reference digital surface model (DSM). We can further improve
this algorithm by matching it to the optimization process used in equation (8.7).
In particular we can make use of the optimum coherence furthest in phase from
the surface topography point as the ideal wv channel. This then acts to maximize
the height of the phase centre of wv in layer 1. However, there still remains
the problem of compensating the volume scattering channel for variations in
structure function f (z). For example, if the structure function is uniform then
the phase optimum will still only reach halfway up layer 1, and so the phase
estimate of equation (8.19) will be only one half the true layer depth.
To try to resolve this, we note that γ̃wv is complex and so has two degrees
of freedom, of amplitude as well as phase. However, we have so far made use
296 Parameter estimation using polarimetric interferometry
only of the phase information. The idea is therefore to try to use the coherence
amplitude of γ̃wv to help compensate for variations in the structure function to
obtain a better estimate of hv . We shall see that there are various ways of doing
this, but a good starting point is to use a Fourier–Legendre expansion of the
structure function f (z) in terms of an infinite series with coefficients aio , which
are then related to the coherence as shown in equation (8.20) (see Section 5.2.4
for a derivation of the functions fi ).
ai βz hv
γ̃ = eiφ eikv (f0 + a10 f1 + a20 f2 + ...) ai0 = , kv = (8.20)
1 + a0 2
This is a more balanced set suitable for inversion; that is, for estimation of the
four unknown parameters φ0 , kv , a10 (w1 ) and a10 (w2 ) from two observations
of complex coherence. The following strategy then follows immediately from
equation (8.22). First we estimate the surface phase term (noting that f1 is
imaginary), not by line fitting as in equation (8.5), but by differencing the
complex coherences as shown in equation (8.23):
We can then calculate kv from the real part of the phase-shifted coherence, as
shown in equation (8.24):
sin kv
Re(γ̃ (w1 )e−iφ ) = Re(γ̃ (w2 )e−iφ ) = 0 ≤ kv ≤ π (8.24)
kv
Note that to invert this relation to estimate kv we can use the following con-
venient invertible approximation for the SINC function, valid over the range
0 to π:
sin(x) (π − x) 1.25
y= ≈ sin 0 ≤ x ≤ π, 0 ≤ y ≤ 1
x 2
⇒ x ≈ π − 2 sin−1 (y0.8 ) (8.25)
⇒ k̂v ≈ π − 2 sin−1 (Re(γ̃ (w2 )e−iφ )0.8 )
Note that this approach, when combined with equation (8.23), allows us to
calculate the surface phase φ0 without the need for a separate straight-line
coherence region assumption. We can then calculate the structure parameter for
arbitrary polarization w from the imaginary part of the phase-shifted coherence,
as shown in equation (8.26):
Im(γ̃ w e−iφ ) Im(γ̃ e−i(kv +φ0 ) )kv2
â10 (w) = = (8.26)
|f1 | sin kv − kv cos kv
Finally, we can reconstruct the vertical profile by knowing the interferometric
wavenumber βz to calculate the height from kv and using the Legendre coef-
ficient to reconstruct the profile with unit integral over this height range, as
shown in equation (8.27):
2k̂v 1 2â10
ĥv = ⇒ fˆL1 (z) = (1 − â10 ) + z 0 ≤ z ≤ ĥv (8.27)
βz ĥv ĥv
Figure 8.7 shows a schematic summary of the types of structure function we can
construct from this simple first-order truncation. We note that the maximum and
minimum of this first-order structure function are given simply in terms of the
parameter a10 , as shown in Figure 8.7. Note that for a10 > 1 this has a negative
minimum on the surface. This may seem to violate the important physical
requirement that f (z) be non-negative (since physically it represents scattered
power as a function of depth). However, such a restriction is not necessary when
^ max
fL1
z = hv
^ ^ ^ min
fL1(0) = (1– a10) = fL1
^ ^ ^ ^
fL1(hv) = (1+ a10) = fL1max
a10 > 1 a10 < 0
a10 = 0
a10 < 1
Normalized height
0.6
0.5
0.4
0.3
0.2
0.1
0
Fig. 8.8 Examples of bipolar Legendre struc- –0.4 –0.2 0 0.2 0.4 0.6 0.8 1 1.2 1.4
ture function estimates of a non-negative step
structure function Relative scattering density
For positive a10 , zc lies between the surface and half the height, while for
negative it lies between the top and the half height.
Before proceeding further it is important to consider the range of the three
unknown parameters φ0 , kv and a10 with a view to investigating the uniqueness
of this model inversion. The phase φ0 is defined in the range 0 to 2π, with
ambiguities arising for surface variations in excess of this. These correspond
to the classical phase unwrapping problem in radar interferometry. However,
here we are more concerned with phase shifts relative to φ0 , and can therefore
ignore the phase unwrapping problem—at least initially. If we consider scenar-
ios where βz is always positive (by appropriate selection of master and slave
8.2 Estimation of height hv 299
tracks for the baseline generation) and hv is also positive, then a good working
range for kv is 0 ≤ kv ≤ π. Although, mathematically, kv can go to infinity, in
practice we would wish to restrict it to avoid phase ambiguities in the layer; that
is, we design the interferometer to ensure that the depth of the layer is always
less than the 2π ambiguity height of the interferometer. This then restricts kv
to the range specified.
The Legendre coefficient a10 , on the other hand, can be bipolar—negative
or positive—but must be constrained so that the magnitude of the right-hand
side in equation (8.21) is always less than or equal to 1 (to match the limits of
coherence on the left-hand side). This requires the following inequality to hold:
" &
'
1 − f02 ' kv2 − sin2 kv
|γ̃ | ≤ 1 ⇒ |a10 | ≤ =( (8.29)
|f1 |2
2
sin kv
kv2 − 2 sin kvkcos
v
kv
+ cos2 kv
minimum
3 maximum
2
Maximum bounds on a 10
–1
–2
–3
90
1
120 60
0.8
0.6
150 30
0.4
0.2
180 0
210 330
positive phase shifts. In any case there are two important observations from
this result:
1. The first-order model does not cover the whole coherence diagram, and
severely limits the possible set of valid coherences. In fact, for a fixed
kv the valid coherences are constrained to lie along a single straight line
going through the appropriate SINC point.
2. It follows from this model that the coherence variation with polarisation
should also lie along a line in the complex plane. However, this line
does not intersect the unit circle at the topographic phase point, and
hence is not the same line as used in other coherence models, such as
the two-layer random-volume-over-ground or RVOG model.
We also note from Figure 8.10 that there is also some ambiguity for phase shifts
less than π/2, where we see the intersection of different lines. However, these
ambiguities can be explained physically as the equivalence of a thick layer with
all scattering from the surface region having the same phase centre as a thin layer
with all scattering coming from the top. In order to enable a unique inversion
we can restrict the model to positive values of a10 ; that is, to scenarios where
the scattering increases rather than decreases with height into layer 1. These
positive loci are shown in Figure 8.11, where we also superimpose the linear
coherence loci of RVOG and its distortion for temporal effects on volume-
only scattering. Clearly, from this the first-order Legendre series cannot be
used to fully represent coherences with large µ values in the RVOG model.
Nor can it even be used for all temporal decorrelations in the volume-only
scattering case.
From these observations we conclude that the conditions are very restrictive
for the inversion scheme of equations (8.23)–(8.27) to apply. Consequently,
we can consider the first-order Legendre model to be inappropriate for general
physical applications. At the very least, for fixed kv , we require better coverage
8.2 Estimation of height hv 301
90 1
eif2 120 60
0.8
0.6
150 30
0.4
0.2
180 0
eif
210 330
240 300
Fig. 8.11 Overlay of RVOG model on posi-
270 tive first-order Legendre model
of the complex plane so that we can model a wider range of physical scenarios
(including the RVOG model). To do this we must extend the model to second
order, as we now consider. However, we shall see that in doing so we can still
make use of some of the inversion ideas from this first-order truncation.
Here R2 is again the truncation error. This error is of the order of the absolute
value of the next term in the coherence series normalized by |f0 |; in this case,
R2 ≈ |f3 |/|f0 |. This is generally small. From Figure 5.15 we see that for 0 ≤
kv ≤ π this is much smaller than the first-order truncation error R1 . Hence
the second-order truncation offers a much more accurate model; but it has an
increased number of parameters, and so we now turn to consider its suitability
for inversion.
We see that we now have two polarisation dependent coefficients, a10 and
a20 , which together with kv and φ0 constitute a set of four unknowns to be
determined from the two observations (amplitude and phase of coherence).
Here, polarisation diversity does not seem to help us, as each additional w adds
two observations but also adds two new unknowns. Instead we need to develop
a different strategy to invert equation (8.30). First, however, we investigate the
expanded coverage in the complex plane of this second-order model.
302 Parameter estimation using polarimetric interferometry
In this form we see that a10 again generates a line in the complex plane passing
through a fixed-point z with direction d , but that now the fixed point is itself
determined by the second-order coefficient a20 . Since the function f2 is always
negative (see Figure 5.15), it follows from equation (8.31) that for negative a20
the fixed point moves radially outwards towards the unit circle, and for positive
a20 it moves inwards towards the origin of the coherence diagram. Thus, for
a fixed kv the result is a family of straight lines, all with the same slope and
defined by a set of fixed points generated by a radial line passing through the
SINC point, as shown in Figure 8.12. Here we show in thick black line the
spiralling SINC locus for φ0 = 0. For each point on this curve there is now
a family of lines generated for positive and negative a10 (shown as solid and
dashed lines respectively) and moving up and down the radial line through the
SINC point according to positive or negative a20 . As we move along the SINC
locus this pattern of lines is rotated and shifted accordingly. The bounds of a10
for fixed a20 can then be derived by setting the coherence to unity, as shown in
a10 > 0
90 1
120 60 a10 < 0
0.8
a20 < 0
0.6
150 30
0.4
180 0
210 330
240 300
Fig. 8.12 Coherence loci for the second-
order Legendre approximation 270
8.2 Estimation of height hv 303
equation (8.32):
1 − f0 (1 + f0 )
(f0 + a20 f2 )2 = 1 ⇒ ≤ a20 ≤ − (8.33)
f2 f2
We note that for a fixed kv value these bounds now lead to full coverage of
the unit circle. Hence, if we know the kv value then we can use the position of
any sample coherence to estimate the two parameters a10 and a20 , as shown in
equation (8.34):
Im(γ̃ w e−iφ ) Re(γ̃ w e−iφ ) − f0
a10 (w) = a20 (w) = (8.34)
|f1 | f2
Here we see, for example, if a10 = 0 then the minimum (maximum) of the
profile occurs at half the layer depth for positive (negative) a20 . In general,
if the ratio a10 /a20 is positive then the extreme point will occur in the lower
half of the layer, while for a negative ratio the extreme point will occur in
the upper half. In this way we can represent a much wider variety of structure
functions than is possible using the classical RVOG model (which assumes
that the maximum always occurs at the top of the volume). In particular we
note that since a20 can be positive or negative, we can now represent functions
with a maximum response below the top of the layer (negative a20 ) or with an
enhanced response from the surface position at z = 0 (a20 positive). The former
is useful for representing non-exponential volume scattering profiles, while the
latter can be used to represent changes in µ—the effective surface-to-volume
scattering ratio.
304 Parameter estimation using polarimetric interferometry
90 1
120 60
0.8
0.6
150 30
0.4
0.2
Fig. 8.13 The RVOG model (stars) superim-
posed on the second-order Legendre coordi-
nate system 180 0
3
Legendre coefficients
–1
–2
0.8
10
0.7
0.6
8
0.5
6 0.4
Height (m)
0.3
4 0.2
0.1
2 0
–0.1
Fig. 8.15 Variation of second-order Legen-
0
–30 –20 –10 0 10 20 30 dre structure function approximation of the
RVOG model with surface-to-volume scatter-
Mu (dB) ing ratio µ
in the positive direction. This reflects changes in the structure function itself.
Figure 8.15 shows an image of how the second-order approximation to the
RVOG structure function varies with µ. Each vertical profile extends over
10 m, and is normalized so that its integral is unity (as in equation (8.35)). On
the left we see the uniform volume scattering profile obtained when µ = 0. As
µ increases we see a shift in the structure function to more localized surface
scattering, as physically expected in the RVOG model. Again we note that as the
surface contribution increases, the second-order approximation is forced to go
negative at some points in the volume. This again reflects the approximate nature
of the truncation rather than any physical interpretation of negative scattering
amplitudes.
In conclusion, we have seen that a second-order Legendre expansion is the
lowest-order truncation capable of providing full unit circle coverage. We have
demonstrated its application to the widely used RVOG coherence model to
demonstrate its ability to reflect changes in the underlying structure function.
There remains, however, one issue with this model. It has too many unknowns
to be inverted and hence to be directly applied for height estimation. In the next
section we turn to consider methods for resolving this limitation.
the true height βz hv , and the second term will approach zero. Half the sum
then still produces the correct kv estimate. The idea is that equation (8.38) will
provide a reasonable estimate for arbitrary structure functions between these
two extremes. It requires estimates of only two parameters: the truesurface
topographic phase φ0 , and the volume-only complex coherence γ̃ wv .
We can extend this idea further and estimate an optimum value of weighting
factor η by using the second-order Legendre structure model for the volume
coherence channel, as shown in equation (8.39):
γ̃ (wv ) = ei(kv +φo ) (f0 + a10 wv f1 + a20 wv f2 ) (8.39)
Now, if we allow a10 and a20 their full range we can fit this coherence to
any kv value, and hence seem to undermine the approximation proposed in
equation (8.38). However, by making some reasonable physical assumptions
about volume scattering we can reduce the working range of a10 and a20 as
follows.
The basic idea is that in the selected special volume-only channel we assume
there is zero surface scattering component, and hence its structure function
should have a local minimum at the surface (at z = 0). This in turn requires
that in this polarization channel wv we restrict a10 ≥ 0 and a20 ≤ 0. When
combined with the limits derived in equations (8.32) and (8.33) we will see that
this constrains the kv values satisfying equation (8.39). Figure 8.16 illustrates
the results. Here we show along the abscissa a set of true kv values in steps of 0.1
over the range 0–2. The corresponding ordinate shows the spread of estimated
kv values obtained using equation (8.38) with η = 0.8. This value is selected to
fit the a20 = 0 variation, and the underestimates we see are then entirely due to
the presence of non-zero a20 structure. It also ensures that equation (8.38) will
always estimate the minimum height consistent with the data. While the trend
in Figure 8.16 is encouraging, some of the errors can apparently be quite large.
For example, if the estimate yields a value of 1 (in the ordinate of Figure 8.16)
1.8
1.6
1.4
Estimated kv value
1.2
0.8
0.6
0.4
0.2
0.2
0.1
0
–0.1
–0.2
–0.3
–0.4
–0.5
0
–1 2
1.5
–2 1
Fig. 8.17 Fractional error for full range of 0.5
–3 0
volume structure functions for kv = 2 a 20 a 10
then we see that the true value could actually lie anywhere between 1 and 1.6,
depending on the volume structure. However, this simple interpretation masks
an important issue. The underestimation errors for each abscissa point in Figure
8.16 increase in proportion to a20 , as shown for the kv = 2 case in Figure 8.17.
Here we plot the fractional error in kv estimate for all valid values of a10 and a20
(but note that this behaviour is typical of all other values, and so our conclusions
will apply equally for all values of kv ).
The key observation from Figure 8.17 is that the largest errors always occur
for large a20 values. This makes physical sense, as this occurs for a quadratic
profile, which is a second-order approximation to a Dirac delta function located
halfway up the volume. In this case we can achieve a combined high-coherence
and low-phase centre, in contradiction to the assumptions behind equation
(8.38). This is a problem faced by all height estimation techniques based
on interferometry. If the volume has a top height hv but the bulk of the
scattering comes from halfway up the volume, then interferometry ‘sees’ a
smaller effective height. In order to resolve such ambiguities we need to add
extra information beyond a single baseline single wavelength interferometer.
Other possibilities include adding more baselines or using a second frequency
(Treuhaft, 2000b, 2004; Reigber, 2000, 2001; Neumann, 2008).
If we accept that such errors can occur for a single baseline, but only for rather
extreme (and unlikely) cases of the structure function, then we can proceed to
employ equation (3.38) as a reasonable approximation, with around 10–15%
estimation errors for a wide range of structure function variations.
We now turn to consider depth retrieval under the more extreme assumption
of a fixed structure function: namely, the exponential assumed by the RVOG
class of models.
Shown in the lower portion of equation (8.40) is the simplified version of this
algorithm obtained in the case λ = 0; that is, when we can assume that the
observed coherence γ̃wv itself has µ = 0—that it contains volume-only scatter-
ing. We have also shifted the topographic phase estimate onto the observable
coherence. Again it is interesting to investigate the range of the two parameters
hv and σe with a view to determining coverage and uniqueness of the solution in
the complex plane. The range of hv is again capped by the requirement to avoid
phase ambiguities in the layer, so that 0 ≤ βz hv ≤ 2π . The extinction can also
have non-negative infinite extent; but in practice, as extinction becomes large
enough so the coherence amplitude becomes insensitive to changes in height,
and varies only the phase of the top of the layer. These trends are apparent in
the loci shown in Figure 8.18. Here we see the variation of coherence over
90 1
120 60
–if 0.8
g^ w e
v
0.6
150 30
0.4
0.2
180 0
210 330
240 300
Fig. 8.18 Overlay of SVOG linear region on
270 RVOG coherence loci
310 Parameter estimation using polarimetric interferometry
the full range of height for varying extinction. The inside spiral corresponds to
the reference zero extinction or SINC loci. We see that for zero extinction the
coherence falls to zero at the 2π height. However, as the extinction increases,
the curvature of the spiral reduces until for large extinction the loci is almost
circular, maintaining high coherence amplitude because of the small effective
volume contributing to decorrelation.
Although the loci are no longer straight lines we note that they provide a set of
non-intersecting curves, and so again if we overlay a sample volume coherence
(shown as the point in Figure 8.18) then we can find a unique solution to
equation (8.40) (for fixed λ), and thus secure an estimate of hv . Note again that
here the extinction parameter is acting as a structure compensation parameter,
allowing height estimation for a wide range of structure functions approximated
by exponentials with varying extinction rates. Note also that the coverage is
not complete. If we draw a line through our sample coherence (shown as the
point in Figure 8.18) we see that if µ is above a certain level the coherence can
still fall outside coverage of the simple extinction model. In this case we have
to employ the free parameter λ in equation (8.40) to move the coherence back
into the valid region, but generally we have no idea which value of λ to use and
hence are likely to make errors in the height retrieval. In this sense we see that
the SINC curve defines a boundary for coverage, and any volume scattering
candidate coherence must lie above the SINC curve to enable a clear solution
in RVOG inversion.
The above algorithm uses the RVOG model to match the observed coherence
in both amplitude and phase. However, to do this we require estimates of the
topographic phase φ0 . We saw in Section 8.1 how to devise algorithms based
on RVOG and OVOG to estimate this parameter. However, sometimes the
coherence can be so low that this phase estimate is too noisy to use. In this
case we would like to employ an algorithm that does not make use of phase
information and relies only on coherence amplitude. We can devise such a
model based on the RVOG approach, as long as we assume that the structure
function has a known form (known extinction, in this case). We then have
only one unknown (the depth of the layer) and one observable, and can solve a
minimization problem as shown in equation (8.41):
? ?
? p ep1 hv − 1 ? p = 2σe
? ?
min G = ? γ̃wv − ? where cos θ (8.41)
hv ? p1 ephv − 1 ?
p1 = p + iβz
In the RVOG case, a known structure function implies knowledge of the mean
extinction coefficient in the medium σ e , as shown in equation (8.41). This
can sometimes be estimated from physical models of the environment or from
measurements and inversions from previous datasets (see Section 3.5.3). Note,
however, that matching coherence amplitude calls for good calibration and
compensation for SNR and temporal effects.
In conclusion, we have seen three important algorithms for estimating layer
depth (or height) from single baseline polarimetric interferometric data. Equa-
tion (8.38) represents an approximate method that makes minimal assumptions
about the layer structure function. Equation (8.40) assumes an exponential
structure via the RVOG model, and consequent matching of the coherence
in both amplitude and phase provides an estimate of both height and mean
8.2 Estimation of height hv 311
z = zo + hv
Oriented volume
Fig. 8.19 Geometry of scattering from an
z = zo
oriented volume with finite depth
312 Parameter estimation using polarimetric interferometry
90 1
120 60
0.8
0.6
150 30
0.4
0.2
180 0
slope = cot(kv + )
210 330
equation (8.21).
3 ?
! ?
? iφ ikv sin kv j sin kv cos kv ?
min G = ? γ̃j − e e + ia − ? (8.42)
kv ,a1 ,a2 ,a3 ? kv kv2 kv ?
j=1
Note that here there are five unknowns and six observables. The parameter
bounds are the same as those derived earlier (0 ≤ kv ≤ π , 0 ≤ aj ≤ π , 0 ≤ φ <
2π), and the only difference now is that for each fixed pair of values kv , φ we
seek a triplet of Legendre coefficients aj that minimize the above function. The
global minimum of such searches will then give us estimates of the parameters
of the layer. There is, however, a simple geometrical interpretation of this
inversion, as we now consider.
The first point to make is that the model of equation (8.42) implies that the
triplet of coherences lie along a line in the complex plane (see Figure 8.20).
These lines are shown in Figure 8.11, and derive from the assumption of a trun-
cated Legendre series. Hence a starting point for the suitability of the inversion
of (8.42) is to test whether or not the components of the triplet are collinear. If
they are, then equation (8.42) will have a good match with the data, otherwise
the assumptions of the truncated Legendre series may be invalid. Note, inci-
dentally, that this line does not intersect the unit circle at the surface topography
point. This is in contrast to the RVOG line fit employed in Section 8.1 to find sur-
face topography from this intersection. It is therefore important in applications
to be able to differentiate between RVOG and finite OV before applying the
appropriate parameter estimation. This can be accomplished in several ways—
for example, by checking the orthogonality of the optimum coherence states:
for single layer OV they will be orthogonal, while for two-layer RVOG they
will not be orthogonal.
The second point to note is that the slope of the line joining the three coher-
ences is given simply in terms of the two fixed parameters kv and φ, as shown in
equation (8.43). Therefore, by fitting a line through the three volume coherences
8.2 Estimation of height hv 313
and measuring its slope, we can obtain directly an estimate of the parameter
kv + φ. We can then employ a single channel model fit—as in the random
volume case—to any one of the three coherences (for example, the maximum
coherence γ1 ) to obtain an estimate of kv from minimization of the function
shown in equation (8.43).
sin kv sin kv cos kv
γ̃ = eiφ eikv + iaj − = z + aj d
kv kv2 kv
ikv +φ sin kv cos kv
⇒d =e i −
kv2 kv
1
⇒m=− (8.43)
tan(kv + φ)
1
⇒ (kv + φ) = − tan−1
m
? ?
? iφ+kv sin kv 1 sin kv cos kv ?
?
⇒ min G = ?γ̃1 − e + ia − ?
kv ,a 1 kv 2
kv kv ?
This value of kz will then, by definition, satisfy the other two polarisation
channels, and when combined with the slope estimate will provide an estimate
for the surface topography φ.
Having determined a simple algorithm based on the Legendre coherence
expansion, we now consider a solution based on exponential structure functions.
z = zo + hv
Oriented volume
? ?
Govog = ?γ̃xx − eiφ (γ̃vo
xx xx ?
+ Fxx (1 − γ̃vo ))
? xy ?
+ ?γ̃xy − eiφ (γ̃vo + Fxy (1 − γ̃vo ))?
xy
? yy ?
+ ?γ̃yy − eiφ (γ̃vo + Fyy (1 − γ̃vo ))?
yy
(8.45)
This has six observables but seven unknowns, and hence is not well suited to
solution as it stands. To be able to make inversion tractable we need to make
some further assumptions about one or more of the parameters in the model.
The simplest of these is the high µ dispersion assumption—also used in Section
8.1.4 for topography estimation. In this case we assume that the extinction in
the x polarisation is so high as to reduce µ in this channel to zero, while the
low extinction channel y maintains a high surface-to-volume scattering ratio.
In this case we can simplify the cost function as shown in equation (8.46):
? ?
Govog = ?γ̃xx − eiφ γ̃vo
xx ?
? xy ?
+ ?γ̃xy − eiφ (γ̃vo + Fxy (1 − γ̃vo ))?
xy
? yy ?
+ ?γ̃yy − eiφ (γ̃vo + Fyy (1 − γ̃vo ))?
yy
(8.46)
This now has six unknowns and six observables. This can be further simplified
if we use the xx and yy channels to estimate surface topography using the line
fit technique, as described in Section 8.1.2. In this way we can reduce the
balance of equation (8.46) to five unknowns and six observables. Problems
arise with this approach if µ remains high in all channels. This can occur, for
example, in applications involving thin layers with low to moderate extinctions
at high angles of incidence. This causes several problems, as the coherences
then all migrate down their lines towards the surface topography point, and it
becomes more difficult to fit lines to the topography point or find the true volume
scattering channel for depth estimation. The only approach then is to constrain
the range of extinctions expected in the problem (by physical modelling or
external measurements), and then use these as known parameters in equation
8.3 Hidden surface/target imaging 315
(8.46) to leave five unknowns for the six observables. Of course, if additional
further information is available (for example, the phase φ of the surface from
external measurements) then this can be added to further reduce the parameter
imbalance. Whichever course is taken, the result is a set of inversions across
the parameter range of extinctions. These provide us with a mean solution and
error bars associated with the spread of solutions.
Note that this a just a special case of the general phase bias removal algorithm
of equation (8.5), the main difference being that now the reference point is
assumed to have µ = 0 and F is the desired parameter, rather than just being
an intermediate step towards phase estimation.
This approach works acceptably well for large µ, but for small µ the two
coherences are close in the complex plane, and so any small errors in coherence
estimation can lead to large errors in line fit. A more robust approach is therefore
to find the estimated surface phase φ using large µ separations or a least squares
line fit (as in equation (8.16)), and to then project all coherences onto the best-
fit line joining the volume coherence to the estimated unit circle point before
estimating F for a given polarisation w, as shown in equation (8.48):
γ̃p w − γ̃v
F(w) = (8.48)
eiφ̂ − γ̃v
Here the projected value of the coherence is given in terms of the line fit
parameters m and c, as shown in equation (8.49):
Re γ̃ (w) + m̂.Im γ̃ (w) − m̂.ĉ
xi =
1 + m̂2 γ̃p w = xi + iyi (8.49)
yi = m̂.xi + ĉ
This relation is derived by minimising the distance between the coherence γ̃ (w)
and the line with known slope and estimating y = m̂x + ĉ, posed as shown in
equation (8.50):
? 2 ?
? ?
min ?(Re(γ̃ ) − xi )2 + m̂xi + c − Im(γ̃ ) ? (8.50)
xi
F(w) g~p(w) This algorithm is summarized schematically in Figure 8.23, where we show the
end points of the linear coherence region for the SVOG model assumption. The
~ parameter F is then just the fractional distance along the line from the volume-
g w ~ eif
v g(w) only coherence point passing through the projected coherence towards the unit
circle. Clearly, if the coherence approaches the unit circle F = 1 we have 100%
Fig. 8.23 Projection of general coherence
onto the line model surface scattering. This fractional parameter can then be directly used with an
estimate of the total scattering cross-section σ to isolate the effective surface
component, as shown in equation (8.51):
σ w = σv w + σes w = (1 − F)σ w + Fσ w
⇒ σes w = Gs (w) = Fσ w (8.51)
However, this result masks a problem with this approach: the confusion
of direct and specular surface scattering in the polarimetric response, as
we now consider.
2) The surface component in polarimetric interferometry is ‘effective’ in
that it includes everything with a phase centre located on the surface.
As we have seen in Section 7.3, this includes not just the direct surface
return but also the specular second order scattering. Note that this acts
to enhance F(wHH−VV ) at the expense of F(wHH+VV ) because of the
π polarimetric phase change on specular reflection, and so will distort
the moisture and roughness estimates of equation (8.52). A method of
correcting for this is to estimate the full polarimetric coherency matrix
for the surface components and use an incoherent decomposition to sep-
arate the direct and dihedral components before applying appropriate
surface parameter techniques. One way to do this is to model the surface
components as a rank-3 reflection symmetric coherency matrix, which
can then be reconstructed from seven separate Gs estimates, as shown
in equation (8.53):
1
Gs (w1 ) (Gs (w4 ) − Gs (w5 ) − i(Gs (w6 ) − Gs (w7 ))) 0
1 2
[Ts ] = m
(Gs (w4 ) − Gs (w5 ) + i(Gs (w6 ) − Gs (w7 ))) Gs (w2 ) 0
2
0 0 Gs (w3 )
where
1 0 0 1 1
1 1
w1 = 0
w2 = 1
w2 = 0
w4 = √ 1
w5 = √ −1
(8.53)
2 2
0 0 1 0 0
1 1
1
1
w6 = √ i w7 = √ −i
2 2
0 0
–2
–4
Mu max (dB)
–6
–8
–10
–12
–5
–10
mu (dB)
–15
–20
–25
∂ γ̂ (1 − γ̃vo )
= = f (µ)g (γ̃vo ) (8.55)
∂µ (1 + µ)2
Maximising sensitivity would then seem to require that γvo = −1; that is, the
baseline is chosen so that the phase centre for volume scattering lies at the π
height of the interferometer. To be realized, however, this phase must also occur
with a coherence magnitude of unity. This is not a realistic scenario, requiring
as it does infinite extinction in RVOG or a very localized structure function in
the Legendre approximation. Adopting the latter, we can express the line length
more realistically as a function of three parameters, kv a20 and a10 , as shown in
equation (8.56):
|1 − γ̃vo | ≈ 1 − eikv (f0 + a10 f1 + a20 f2 ) (8.56)
Before considering this in more detail, we first include the change of minimum
coherence along the line. This is important, as it impacts on the number of sam-
ples required to estimate coherence and hence on the accuracy and resolution
of any estimation. The µ for minimum coherence was found in equation (7.49).
Inserting this into the line coherence model we obtain the following expression
for the minimum coherence:
(8.58)
We can now investigate the upper and lower bounds of this function as we
change structure for a given baseline/height product. Note that a10 and a20
have a limited range, as we are considering the volume-only component of the
structure function (and so only structure functions that increase with height).
8.3 Hidden surface/target imaging 321
1.8
maximum value
1.6 minimum value
Bounds on line length
1.4
1.2
0.8
0.6
0.4
0.2
maximum value
Bounds on minimum coherence
minimum value
0.8
0.6
0.4
0.2
This limits the Legendre spectrum, so that a10 is non-negative and a20 is non-
positive. In Figures 8.26 and 8.27 we show how the two components of this
function vary with kv . The line length in equation (8.26) starts at zero, and
then increases to a maximum of 2 before falling again for high kv . However, at
the same time we see the minimum coherence value start at 1 for low kv , and
decrease to zero when the line goes through the origin. Indeed, we see that the
minimum can be zero for all structure configurations beyond kv = 2.
Finally, in Figure 8.28 we show the corresponding variation of the product
of these two components, and see a clear optimum range 1 ≤ kv ≤ 1.5. This
range then represents the best compromise between line-length (sensitivity) and
322 Parameter estimation using polarimetric interferometry
0.8
maximum value
0.4 minimum value
0.2
Note that the coherence so obtained can be much higher than the polarimetric
only coherence, and so by adopting interferometry we obtain a better situ-
ation (effectively a lower entropy) than the polarimetric approach based on
equation (8.54).
In the RVOG and OVOG models, polarisations other than volume-only are
related by a scale factor: the surface-to-volume scattering ratio, µ. This parame-
ter can be estimated for an arbitrary polarisation using the techniques described
in equation (8.48). The corresponding vertical profile is then obtained as a
weighted sum of an exponential and delta function, as shown in equation (8.60):
2σ̂e
cos θe− cos θ z
fˆrvog (w, z) = 2σ̂e
+ µ̂ w δ (z) 0 ≤ z ≤ ĥv (8.60)
2σ̂e (e cos θ
ĥv
− 1)
In the Legendre approach, more interesting possibilities arise and lead to a gen-
eralization of the structure estimation problem, termed coherence tomography
(Cloude, 2006b), as we now consider.
This can then be written in matrix form in terms of the real and imaginary parts
of the phase normalized coherence, as shown in equation (8.62):
1 0 0 a00 1
0 −if1 0 . a01 = Im(γ̃k ) ⇒ [L]a = b (8.62)
0 0 f2 a02 Re(γ̃k ) − f0
The next important idea is that we can now invert this relationship to obtain
estimates of the polarisation-dependent Legendre parameters from coherence
and knowledge of the matrix [L], as shown in equation (8.63):
â = [L]−1 b̂ (8.63)
From the vector â we can then estimate the normalized vertical structure
function for a known layer depth hv , as shown in equation (8.64):
/
ˆ 1 2z 6z 2
0 ≤ z ≤ hv ⇒ fL2 (z) = 1 − â10 + â20 + (â10 − 3â20 ) + â20 2
hv hv hv
(8.64)
Note that when the quadratic term (a20 ) is zero this reverts to the linear approx-
imation fˆL1 (z), as developed in equation (8.27). Equations (8.62) and (8.63)
constitute a method for reconstructing the function f (z) from coherence, and
324 Parameter estimation using polarimetric interferometry
are therefore termed coherence tomography, or CT. We shall see that the matrix
formulation can be extended to arbitrary order of Legendre polynomial, and
hence to higher and higher resolution reconstructions by adding multiple base-
lines to the interferometer. However, there remains the important issue of how
to obtain estimates of the polarisation-independent terms kv and φ0 . We now
turn to consider this in more detail.
In equation (8.61) we separated coherence into polarisation-dependent and
independent components. The latter set comprises three parameters of interest:
the layer depth hv and interferometric wavenumber βz (which are then used
to calculate kv ), and the phase of the bottom of the layer, φ0 . The wavenum-
ber can be estimated from knowledge of the baseline geometry and operating
wavelength of the interferometer (see Section 5.1), but the other two parame-
ters require special attention. There are two principle ways of obtaining these
parameters.
⇒ â(w) = [L]−1 b̂
/
ˆ 1 2z 6z 2
⇒ fL2 (w, z) = 1 − â10 (+â20 (w) + (â10 (w) − 3â20 (w)) + â20 (w) 2
hv hv hv
(8.65)
This approach makes no assumptions about the shape of the coherence region,
and so can be used to investigate the most general profiles. We will consider
an example of such an approach based on laboratory anechoic chamber mea-
surements of maize plants in Chapter 9. Often, however, especially in remote
8.4 Structure estimation: extinction and Legendre parameters 325
90 1
120 60
0.8
0.6
150 30
0.4
a20 a20
a10 a10
a wide range of kv values. Here we show the set of coordinate lines for two kv
values, with the two origins (a10 = a20 = 0) shown as black points, and see
that the sample coherence point, although it has different a10 a20 values, can be
made to fit either. Hence a single baseline cannot be used to estimate the two
Legendre structure parameters uniquely. We can, however, estimate a family
of solutions using single baseline data. As we move around the SINC spiral in
Figure 8.29 we obtain a set of solution pairs a10 , a20 for the given coherence
(in grey). We can represent this family geometrically as a set of solution points
in the a10 , a20 plane, generated as kv varies from 0 to π . These generate a curve
in the plane, as shown schematically in Figure 8.30. We can then use this idea
to propose a method for estimating the true kv value by combining data from a
second baseline.
If we now consider that coherence data is available for a second additional
baseline, related to the first by a baseline ratio Br , then we can generate a second
set of a10 , a20 points by simultaneously solving the matrix equation for kv for
the first baseline and Br kv for the second. The correct kv value then occurs
when these two curves intersect, as shown schematically in Figure 8.30. Here
we show two possible scenarios. On the left, the solid curve is the curve of
solutions obtained for the first baseline, and the dashed curve is the solutions
for the second. This is a well-conditioned case, when the intersection point
occurs for nearly orthogonal curves. Such a scenario will be robust to errors in
the two coherence estimates. On the right of the figure is shown the opposite
case of a poorly conditioned solution where the intersection point occurs for
nearly parallel curves. For example, in the limiting case, if we take the two
baselines equal (Br = 1) then the loci will exactly overlap and a solution is
not possible. When nearly overlapping, any small perturbation of the curves
will lead to a large change in the solution. This ill-conditioning can undermine
the uniqueness of a solution using dual baselines by making the algorithm so
sensitive to noise that it cannot be used in practical applications (Hopcraft,
1992). The level of such ill-conditioning will be a function of the baseline
ratio Br . We shall examine the conditioning of coherence tomography and its
dependence on Br in more detail in Section 8.4.3. First, however, we consider
a formalization of this approach and how to generalize it for unknown surface
topography φ0 .
We can formally write the solution to the dual baseline kv estimate as mini-
mization of the coherence error as defined in equation (8.66), where subscripts
8.4 Structure estimation: extinction and Legendre parameters 327
γ̃1 = eiφ1 eikv1 (f0 (kv1 ) + a10 f1 (kv1 ) + a20 f2 (kv1 )) ⇒ γ̃k1 = γ̃1 e−iφ1 e−ikv1
−1
â00 1 0 0 1
â01 = 0 −if1 (kv1 ) 0 . Im(γ̃k1 )
â20 0 0 f2 (kv1 ) Re(γ̃k1 ) − f0 (kv1 )
iBr φ1 ikv2
⇒ γ̃2 = e
est
e f0 (kv2 ) + â10 f1 (kv2 ) + â20 f2 (kv2 ) where kv2 = Br kv1
? ?
⇒ Coherence error = ?γ̃2 − γ̃2est ? (8.66)
The procedure (for φ1 = 0) is then to vary kv1 from 0 to π , calculating for each
value the Legendre spectrum a10 , a20 . We then use these values to estimate the
second baseline coherence and select the triplet kv1 , a10 , a20 that minimizes the
difference between this complex estimate and the true second baseline coher-
ence γ̃2 , as shown in equation (8.66). When topographic phase is also unknown,
the only difference we face is to search for the intersection of a10 , a20 loci in
a two-dimensional space of φ0 and kv1 rather than just kv1 . The estimate of
the second baseline coherence is then phase shifted by a scaled topography, as
shown in equation (8.66). (Note that we are also assuming that there are no
residual phase errors between baselines, as can occur, for example, in repeat
pass sensors. If not true, then we must add an extra unknown phase parameter
to equation (8.66).)
As a typical example of dual baseline performance, consider the choice (φ1 =
φ2 = 0) Br = 0.5 and kv = π/2, with a profile defined by a10 = 0.5 and a20 =
1. These lead to very distinct coherences of 0.87 for the smaller baseline and 0.54
for the larger. These two points can then be used to estimate a pair of solution
curves for a10 , a20 , as shown in Figure 8.31. Here we see a scenario that seems
poorly conditioned, despite the fact that the coherences from the two baselines
are very different, with the intersection point of the two curves occurring for
2.5
1.5
a 20
0.5
–0.5
–5
–10
–20
–25
–30
–35
300 –5
Topography phase (degrees)
250
–10
200
–15
150
–20
100
–25
50
number of samples to minimize residual estimation noise. Such noise could lead
to large errors in the estimate of kv , and we need to balance the amplification
of error caused by the ill-conditioning against the number of samples required
in order to correctly assess these errors. This example nicely illustrates how
uniqueness is not the only criterion required for an assessment of algorithm
performance, and that the level of numerical stability or ill-conditioning must
also be quantified. We will consider such an analysis in Section 8.4.3.
Equation (8.66) represents one method of estimating kv , but it requires data
for two baselines. But this is often not available, and so it is of interest to consider
alternative single baseline strategies for estimation of kv that still enable use of
the second-order Legendre approach to CT.
1@ 0.8 A
kv = arg(γ̃wv e−iφ0 ) + 0.8(π − 2 sin−1 (γ̃wv ) (8.67)
2
The estimated φ0 , kv values then establish a unique coordinate system for the
whole coherence diagram, so that the structure function for all other polari-
sations can be reconstructed up to second order using the matrix inversion of
coherence tomography.
This approximation works with only a single baseline, but requires the com-
bination of at least two interferograms formed for different polarisations: one
volume and the other surface dominated. It therefore requires some polarisation
diversity in measurements, and hence we term it polarisation coherence tomog-
raphy, or PCT, to distinguish it from standard CT. The basic steps involved in
PCT are summarized in Figure 8.34. Here we show how, starting from a pair
of polarisations, ws and wv , we can use the SVOG assumption to fit a line
330 Parameter estimation using polarimetric interferometry
^ ^
– –
kv ≈ 12 arg(gwv e–if) + 0.8(p – 2sin–1(gwv0.8)
2kv
hv = 0 < z < hv
kz
1 2z 6z2
Fig. 8.34 Summary of algorithm for single- fL2 (w, z) = {1 – a10 (+a10 (w) + (a10 (w) – 3a20 (w)) +a20 (w) 2 }
hv hv hv
baseline polarization coherence tomography
(PCT)
and find the topographic phase. (The two-state line fit is shown in Figure 8.34,
although the total least squares or other approach could equally be applied if
more polarisation channels were to be made available.) We must then select
one of the polarisations as a volume-dominated channel to obtain an estimate
of kv . We then use the kv and φ0 estimates to scale the coherence for arbitrary
polarisation w to generate a second order estimate of the structure function, as
shown. This can then be repeated for arbitrary polarisations as desired.
The procedure has a simple geometric interpretation as shown in Figure 8.35.
The process begins by identifying the volume coherence (grey point) and cor-
recting for surface topography (white point) using the SVOG linear coherence
assumption (black line). From the grey volume coherence we then estimate kv ,
which locates the origin for the a10 , a20 coordinate system (the black point in
Figure 8.35). This origin then defines a grid of a10 a20 values that can be used
to estimate the structure function for arbitrary location in the complex unit cir-
cle. However, under the SVOG assumption, coherence varies with polarisation
8.4 Structure estimation: extinction and Legendre parameters 331
90 1
120 60
0.8
0.6
150 30
0.4
180 0
210 330
along the black line and so we can access only a limited portion of the a10 /a20
space. Note, for example, that while a10 can be positive or negative, the line
permits only positive a20 values in the reconstruction. This is the price to pay
for using such a bootstrap approach.
We now turn to consider more quantitatively the ill-conditioning of the matrix
inversion embedded in both CT and PCT algorithms.
where the functions f0 , f1 and f2 are given in Figure 8.34. Some care must
be taken with this inversion, as it can lead to an amplification of any errors
in b so that the resulting a may represent a very poor estimate of structure.
An alternative way of thinking about this is to estimate the level of noise we
can tolerate in the coherence vector b so as to keep the fractional error in
a below a prescribed value. This will then allow us to estimate the number of
coherence samples required for good estimation. In this section we develop such
an algorithm for analysing the stability of matrix inversions like equation (8.68).
Key to quantifying this amplification process is the condition number (CN)
of the matrix [L] (see Appendix 1). The larger CN, the larger any amplification
of errors in b. As [L] is diagonal in equation (8.68), we can obtain an explicit
expression for the condition number of the matrix [F] as a ratio of the functions
332 Parameter estimation using polarimetric interferometry
104
103
CN
102
101
1 kv2
CN = − =− (8.69)
f2 3 cos kv − (3 − kv2 ) sin kv kv
Figure 8.36 plots this function versus normalized wavenumber kv . Note that for
small baseline/height products the inversion is very poorly conditioned (a large
CN). For baseline/height products around unity, the condition number is around
10–20. Since [F] is diagonal we can also identify the worst-case scenario, when
the system becomes most sensitive to errors in b. From equation (8.68) this
arises for perturbations of a true solution of the form
1 1
b = 0 ⇒ b + δb = 0 (8.70)
0 δ
0.9 L = 25
L = 50
0.8 L = 100
Maximum fractional error
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 0.5 1 1.5 2 2.5 3 Fig. 8.37 Maximum bound on fractional
error in Legendre estimates versus kv for
kv different number of looks
where L is the number of looks. In this way we can estimate an upper bound on
the fractional error in the estimate of the Legendre coefficients as a function of
just two parameters, kv and the number of looks L, as shown in equation (8.72):
)? ?*
?δa? sin2 kv − kv2
max ? ? = CN .β = √ (8.72)
?a?
2L 3 cos kv − (3 − kv2 ) sinkvkv
Figure 8.37 shows how this bound varies as a function of kv for various number
of looks L. We should note that this represents a worst-case scenario, and
generally the errors will be better than this. This approach assumes that the
layer is almost a uniform volume scatterer (and so the volume coherence lies
along the bounding SINC curve in Figure 8.35). If this is not true, and the
volume channel has some other structure, then the errors will be less than this
bound.
This conditioning error is due to amplified noise in coherence estimation.
However, we have two other main sources of system noise to consider: the
effects of SNR, and temporal decorrelation in the interferometer. These can
now be incorporated into the CT and PCT formulations as follows.
90 1
120 60
0.8
0.6
150 30
0.4
0.2
180 0
210 330
sin kv
fo =
kv
sin kv cos kv
f1 = i −
kv2 kv
3 cos kv 6 − 3kv2 1
f2 = − + sin kv
kv2 2kv3 2kv
30 − 5kv2 3 30 − 15kv2 3
f3 = i + cos kv − + sin kv
2kv3 2kv 2kv4 2kv2
35(kv2 − 6) 15 35(kv4 − 12kv2 + 24) 30(2 − kv2 ) 3
f4 = − cos kv + + + sin kv
2kv4 2kv2 8kv5 8kv3 8kv
−2kv4 + 210kv2 − 1890 30kv4 − 840kv2 + 1890
f5 = i cos kv + sin kv
kv5 kv6
42kv4 − 2520kv2 + 20790) 2kv6 − 420kv4 + 9450kv2 − 20790)
f6 = cos kv + sin kv
kv6 kv7
(8.75)
We now see that there is a natural extension of this idea to multiple baselines,
adding two new structure parameters per baseline, so that in general N baselines
yields 2N + 1 terms of the Fourier–Legendre series. Returning to the N = 2
case, CT inversion then takes the form of a matrix equation based on the use of
equation (8.74) for two baselines ‘x’ and ‘y’, as shown in equation (8.76):
1 0 0 0 0 a00 1
0 −if x 0 −if x 0 a10 Im γ̃ x
1 3 k x
0 f4 . a20 = Re γ̃k − f0
−1
⇒ â = [L] b
0 f2x 0 x x
0 −if y
0 −if3 0 a30 Im γ̃k
y y
1
y y y y
0 0 f2 0 f4 a40 Re γ̃k − f0
(8.76)
336 Parameter estimation using polarimetric interferometry
Note that the real matrix [F] is now 5 × 5 (for N baselines it is (2N + 1) ×
(2N + 1)), and is no longer diagonal in structure. From the estimated vector of
Legendre coefficients we can determine the shape of the corresponding structure
function up to fourth order, as shown in equation (8.77):
Extending this to N = 3 (to three baselines) leads to the following model for
coherence, where again the new functions f5 and f6 are given in equation (8.75).
When applied across the three baselines ‘x’, ‘y’ and ‘z’, this leads to a cor-
responding 7 × 7 matrix inversion for coherence tomography, as shown in
equation (8.79). Note again that [F] is a real non-diagonal matrix with elements
a function of kv .
1 0 0 0 0 0 0 a00 1
0 −if x 0
−if3x −if5x a10 Im x γ̃k x
x
1 0 0
0 0 f2x 0 f4x 0 f6 a20 Re γ̃k − f0
x
0 −if y 0 −if3
y
0 −if5
y
0 a30 = Im γ̃ y (8.79)
1
y k
0 0 f2
y
0 f4
y
0 f6 a40 Re γ̃k − f0
x y
0 −if z 0 −if3z 0 −if5z 0 a50 Im γ̃kz
1
0 0 f2x 0 f4x 0 f6x a60 Re γ̃k − f0
z z
Figure 8.39 summarizes the differences between single, dual and triple base-
line reconstructions by plotting the polynomials employed in the corresponding
reconstructions. We can clearly see the improvement in resolution with increas-
ing baseline. However, while formulation of CT in this way is straightforward,
note from Figure 5.15 that the functions fi tend to zero with increasing order
and hence anticipate problems with the conditioning of the inversion. This will
provide a practical limit to the achievable resolution, as eventually we will
demand impossible limits on the control of error in coherence estimation for
the vector b. In order to assess this, we now turn to quantify the condition
number of multi-baseline CT.
In general we can analyse the conditioning of multi-baseline CT using a
singular value decomposition of the matrix [L]. This allows us to represent
the inversion in terms of a (2N + 1) × (2N + 1) diagonal matrix [], just as
we did for the single-baseline case in equation (8.68), but now with different
8.4 Structure estimation: extinction and Legendre parameters 337
Height
Height
0.5 0.5 0.5
0 0 0
0 0.5 1 0 0.5 1 0 0.5 1 Fig. 8.39 Legendre functions for single,
Relative density Relative density Relative density dual, and triple baseline inversions
orthogonal frames [U ] and [V ] for the vectors a and b. Formally we can write
the matrix [L] in the form shown in equation (8.81) (see Appendix 1):
where the (2N + 1) real parameters si are the singular values. The formal solu-
tion to the CT estimation problem can be written in terms of these matrix
components, as shown in equation (8.82):
â = [V ].[]−1 [U ]∗T b,
−1
s1 0 0 .. 0
0 s−1 0 .. 0
2 (8.82)
0 s3−1
[]−1 = 0 .. 0
..
: : : . :
−1
0 0 0 0 s2N +1
The condition number of the inversion is then defined as the ratio of maximum
to minimum singular values, as shown in equation (8.83):
s1
CN = (8.83)
s2N +1
2.5 40
35
2
30
Kvy
25
1.5
20
1 15
10
0.5
5
2.5 40
35
2
30
Kvy
25
1.5
20
1 15
10
0.5
5
One way to deal with this high condition number is to filter the [L] matrix,
which can be achieved by removing the smallest singular value to reconstruct
a profile with a matrix [f ], as shown in equation (8.84). Here we obtain
a matrix with a lower condition number, given by s1 /s2N . Note that in this
case we lose some resolution (given by one pair of singular vectors), but still
gain resolution over the reduced baseline case. Figure 8.41 shows the condition
number of this filtered matrix. Here we see an order of magnitude improvement
8.4 Structure estimation: extinction and Legendre parameters 339
in conditioning, with condition numbers around 100 for the useful part of the kv
range.
2
=
Transmitter Receiver =
p 2
z r=2 cos
r 2
FT
e 2 cos
2
A O x-axis A
Radar flight trajectory
z-axis
z = zo
∂ 2d ∂ 2d 1 ∂ 2d
+ = (9.1)
∂x2 ∂z 2 v2 ∂t 2
342 Applications of polarimetry and interferometry
If we now take the Fourier transform of this equation with respect to time and
to x, we obtain the following (ordinary) differential equation for the transform
quantity D(βx , ω, z):
d 2D ω2 d 2D ω2 β 2 v2
βx2 D − = 2D⇒ = − 2 (1 − x 2 )D (9.2)
dz 2 v dz 2 v ω
This ‘ODE’ can then be factored in terms of upward and downward (±z) prop-
agating waves. The latter we can then use to ‘migrate’ the field from the line
AA back to the source line at z = zo . This will render the wave field sensed
along AA as an ‘image’ of the apparent ‘sources’ along z = zo ; that is, it will
focus the radar image. We shall see that the larger the aperture AA , the higher
the resolution in this image. For the downward propagating waves we have the
following factorization:
"
dD ω βx v 2 ω
= f .D = i 1− D=i 1 − η2 D (9.3)
dz v ω v
This equation has a simple plane wave solution. Therefore, if we know D across
a surface AA then we can propagate or ‘migrate’ the data to any other z value,
as shown in equation (9.4):
ω
√
1−η2 z
D(βx , ω, z + z) = D(βx , ω, z)ei v (9.4)
Finally, we can obtain the image of the sources by an inverse Fourier transform
(FT) with respect to βx and a summation w.r.t. ω as follows:
!!
d (x, t = 0, z) = D(βx , ω, z)eiβx x (9.5)
kx ω
If the velocity is constant (v = c/2 to account for the two-way propagation) then
this summation can be performed very efficiently by using a two-dimensional
Fourier transform as follows. We first propagate the data from AA to z = zo in
one single step by phase rotation, as shown in equation (9.6):
ω
D(βx , ω, zo ) = D(βx , ω, z = 0)eiβz zo βz = 1 − η2 (9.6)
v
We then perform the summation and inverse transform to obtain the following
integral:
cβz
d (x, t = 0, zo ) = D(βx , ω, z = 0).ei(βz z+βx x) d βx d βz (9.8)
βx2 + βz2
9.1 Radar imaging 343
SAR processing
pixel
2-D IFT
z
bx Image
Space
x
2dFT
d(x,t) D(bx,v)
Stolt Interpolation c 2 2 o2
D(bx,v) U(bx,bz) bx , +( + – o
2 c
2dIFT Fig. 9.3 Schematic geometry and key steps
U(bx,bz) u(x,z) in SAR processing
344 Applications of polarimetry and interferometry
bx
x
bz
Image
2p
= bz
4pW
⇒ z =
c
Pixel z
z c 2W
2p 2v l
= bx ≈ bu = u ⇒ x =
x c 2u u
depends on the angular width of the sector. At first sight this may seem to
depend on several system variables, but there is a simple relationship between
the maximum angular width (and hence best resolution) and real antenna size,
as shown schematically in Figure 9.5. The beamwidth (in radians) of the real
antenna is approximately equal to the size of the aperture in wavelengths. When
dr we substitute this result into Figure 9.4 we obtain the well-known result, from
SAR theory, that the cross-range resolution is given by half the real antenna
= aperture size:
dr
λ dr
θ ≈ ⇒ x ≈ (9.9)
dr 2
This result has important implications for the exploitation of polarisation effects
in radar imaging, as we now show.
4v sin θ 4v θ 2v λ 2v
fd = ⇒ fd max ≈ · = · = (9.10)
λ λ 2 λ dr dr
The maximum Doppler shift therefore occurs at the outer edges of the real beam
(positive for approaching points, and negative for receding). By again using our
approximation for the beamwidth of the antenna in terms of the real aperture
size (Figure 9.5), we can obtain a simple expression for the maximum Doppler
9.2 Imaging interferometry: InSAR 345
shift as a function of the ratio of platform speed to real aperture size, as shown
on the right-hand side of equation (9.10). One important constraint required
for undistorted SAR imaging is therefore that the PRF be greater than or equal
to this maximum shift. For small antennas on fast-moving platforms this can
require a very high PRF. However, there are two consequences of operating
at high PRF. The first is the requirement for a transmitter with higher mean
power (given by the peak power times the duty cycle of the radar or τ *PRF,
where τ is the pulse width), which may be expensive or difficult to obtain at the
desired operating frequency. The second issue, however, is more important for
imaging, in that the PRF also impacts on the range extent of the image or range
swath in pulsed systems. The problem here arises from range ambiguities. If
the PRF is too high and the range variation across the image too large, then
there can be an ambiguity as to which transmitter pulse any particular received
pulse actually belongs. A quantitative analysis (Curlander, 1991) shows that the
PRF must be bounded by the following inequality in order to avoid such range
ambiguities:
c
PRF ≤ (9.11)
2Ws
where Ws is the width of the image in the range direction (the range swath size).
This tends to demand low PRF for wide image coverage of the system, which,
as we have seen, is in direct contrast with the requirements for high resolution
in the cross-range direction. The compromise between such conflicting PRF
requirements is one of the central engineering steps in imaging radar design.
Polarisation switching places further constraints on this relationship, as we
show in Section 9.3; but first we turn to consider radar interferometry.
2-D
IFT
wave-space P
P
T
IF
2-D
Image space
N
2-D
IFT
wave-space P
P P M
T
IF
2-D Image space
Fig. 9.7 Boxcar estimation of complex
coherence in radar imaging interferometry
x
HH
HV
VH
VV P
P
P
P
z
Image space
Fig. 9.8 Wave-space interpretation of POL-
SAR imaging
1 M N *T
T = k-ij k-ij
NM i =1 j =1
N
Image space
k
ij
P
M pixel
T̂ = T + f (k i k ∗T
i − T ) (9.13)
left and right circular (L and R). In order to measure the first column of [S]
we illuminate the scatterer with x polarisation and measure, simultaneously in
amplitude and phase, the scattered field components in the orthogonal x and
y channels. Simultaneous dual reception can be achieved using a two-channel
receiver preceded by an orthogonal polarisation splitter, although, as we shall
see, this complicates the calibration, as multiple channels need to be balanced
in both amplitude and phase (Freeman, 1992).
The second column of [S] can be similarly measured by illuminating the
scatterer with orthogonal y polarisation and again measuring coherently the
x and y components of the scattered radiation. In this way, all four complex
matrix elements are obtained. Ideally, as SAR involves a moving platform we
should transmit x and y polarisations simultaneously. This could in princi-
ple be achieved using suitable orthogonal coding. However, by far the most
common method is to employ a single carrier frequency and time multiplex
the two orthogonal states on a pulse-by-pulse basis, as shown schematically in
Figure 9.10. Here we first transmit a horizontal or H polarised pulse, and receive
in the co- and crosspolarised channels the first column of [S], as shown. The
next pulse is then transmitted with V polarisation, and we measure the second
column of [S]. In this switching scheme there is an inherent time delay of one
PRI (pulse repetition interval) between the first and second column, and so the
bandwidth of the transmitter switch needs to be much faster than any decorre-
lation time of the scattering process, so as to maintain coherence between the
columns of [S]. Bandwidths in the kHz region are typical for imaging radar
applications.
We see, however, that this interleaved switching arrangement also inter-
feres with the sampling requirements for SAR processing. There are two main
options. The first is to keep the same PRF (and hence mean power) as for a
single polarisation system. However, this means that the effective PRF for each
column measurement is halved, as H is transmitted only on every second pulse
in Figure 9.10. This in turn means that the azimuth resolution is also halved,
and a larger antenna (with twice the size in the azimuth direction) is required
to avoid Doppler aliasing. Both of these are unattractive options for imaging
radar systems.
In the second scenario we can instead double the PRF of the system so as
to maintain the column sampling at the same rate as before, and so maintain
azimuth resolution and keep the same antenna size. However, in this case the
mean power of the transmitter is doubled (unless the pulse length is also halved
to maintain the same duty cycle, which then has further implications for the
range resolution, which may have to be halved). This option also leads to a halv-
ing of the range swath and hence a smaller image, due to the possibility of range
ambiguities between columns of [S] (especially in the crosspolarised channels).
In equation (9.11), therefore, we need to use the full PRF in determining the
range swath, leading to reduction by a factor of 2. All these considerations are
worsened by the fact that systems always have finite isolation between orthog-
onal channels; that is, there will inevitably be some y component radiated when
350 Applications of polarimetry and interferometry
x is selected, and vice versa. Methods for dealing with such practical issues via
calibration are dealt with in the next section.
There are two strategies for dealing with these distortions. In design, every
effort can be made to reduce cross-talk by good antenna design and careful
system layout. As a second strategy we can employ the process of calibration
to estimate the distortion matrices [R] and [T ], and remove them by matrix
inversions, so that in the absence of noise, for example, we can obtain an
estimate of the true scattering matrix, as shown in equation (9.15):
There are various methods available for the estimation of the elements of [R] and
[T ], involving a combination of internal and external calibration techniques.
Internal methods involve monitoring of signals by test channels inside the radar
to estimate imbalances, while external methods (which have the advantage
9.3 Polarimetric synthetic aperture radar (POLSAR) 351
that they include the full system, including antenna and propagation effects)
involve measuring signals from external active and passive reflectors, which
send signals back through the radar system from an object with known polari-
metric behaviour. By arranging for a set of four such reflectors with orthogonal
scattering vectors (see Section 4.1.4), a set of sixteen equations in the sixteen
unknowns of [R] and [T ] can be obtained. In practice, simpler deployments
are favoured, often of only one or two types of reflector with additional con-
straints (such as symmetry assumptions in the scattering from random media;
see Section 2.4.2.1) used to solve the remaining calibration equations. To see
how these arise we now turn to a vector formulation of the system calibration
equations.
The two key features of this formulation are the presence of quadratic products
of the distortion matrices appearing in [Z], and the simple mathematical form
of the correction process. If we can estimate the elements of [R] and [T ], then
their distortions can be offset by a single matrix inversion as shown. There are
two important special forms of this calibration matrix to consider, both of which
stem from the important case of reciprocal backscatter when SHV = SVH . In
this case the matrix [Z] is no longer square and has dimension 4 × 3, as shown
in equation (9.17):
OHH r11 t11 r11 t21 + r12 t11 r12 t21
OHV r11 t12 SHH
r11 t22 + r12 t12 r12 t22
k obs =
OVH = [Z] k s = r21 t11 . SHV
r21 t21 + r22 t11 r22 t21
SVV
OVV r21 t12 r21 t22 + r22 t12 r22 t22
−1 ∗T
⇒ k s = Z ∗T Z Z k obs (9.17)
Note that the observed scattering vector k obs violates reciprocity, though this is
due entirely to the effect of system distortions. This observation can be used,
352 Applications of polarimetry and interferometry
for example, to test the quality of system calibration. If the Pauli channel image
OHV − OVH is formed, it should, for properly calibrated backscatter data,
behave like noise (have zero coherence, and so on). If it contains structure,
then the calibration is not perfect. Note that calibration of the data does not
now involve matrix inversion directly, but instead a pseudo-inverse based on
a least squares solution can be employed, as shown in equation (9.17). This
arises because we are using reciprocity to reduce the number of unknowns
below the number of observations, and hence have an overdetermined system of
equations.
In practice, a further simplification can often be made by assuming that the
cross-talk terms (the off-diagonal elements of [R] and [T ]) are small compared
to the copolar distortion terms. In this case we can set to zero elements of the
[Z] matrix that involve products of the small crosspolar terms. In this ‘small-
coupling’ assumption the [Z] matrix takes on the simplified form shown in
equation (9.18):
r11 t11 r11 t21 + r12 t11 0
r11 t12 r11 t22 r12 t22
[Z] =
r21 t11
(9.18)
r22 t11 r22 t21
0 r21 t22 + r22 t12 r22 t22
A special case of this matrix occurs in the limit of zero cross-talk. If the design
isolation of the system is very good (typically better than –30 dB), then we
can set all off-diagonal terms of [R] and [T ] to zero and establish a simplified
calibration matrix as shown in equation (9.19):
1 0 0
0 t22 /t11 0
[Z] = r11 t11
0
(9.19)
r22 /r11 0
0 0 r22 t22 /r11 t11
Note that this matrix still causes lack of reciprocity in the observed vector due
to differences in the receiver and transmitter copolar distortion channels.
To illustrate how this matrix vector formulation can be used to derive practical
calibration algorithms, we summarize here the main steps involved in a widely
used POLSAR calibration algorithm, first derived in Quegan (1994), and then
further modified by Papathanassiou (1998a) and Kimura (2004). It employs
the small cross-talk hypothesis of equation (9.18) to express the relationship
between calibrated and observed (uncalibrated) scattering four-element vectors
in the form shown in equation (9.20).
−1
hh k2 0 0 0 hh
hv 0 0
=Y k 0 [Z]−1 . hv (9.20)
vh 0 0 k 0 vh
vv calibrated 0 0 0 1 vv uncalibrated
Here two scalar factors Y = r22 t22 k = rr2211 have been factored from the matrix
inverse. These are found by imaging a single point target with a known scattering
9.3 Polarimetric synthetic aperture radar (POLSAR) 353
matrix. Often a trihedral corner reflector (Figure 1.21) is used for this purpose.
It has the identity matrix as a true scattering matrix, and hence by measuring
the ratio of apparent copolarised elements for the pixel we can establish a direct
estimate of ‘k’. The radiometric factor Y can then be determined from the
known radar cross-section (RCS) of the trihedral. The 4 × 4 matrix [Z]−1 can
be written in terms of a set of four cross-talk ratios u, v, w and z, and a factor
‘a’ as shown in equation (9.21):
1 −v −w vw
1 −az a azw −aw
[Z]−1 =
a −u uv 1 −v
azu −au −az a (9.21)
1. The first is reciprocity, so that for the true matrix HV = VH. Hence any
departure from this in the measured data is attributed to the effects of
the system distortions u, v, w, z, and a.
2. In addition, we also need to assume that the depolariser has reflection
symmetry, so that the true covariance matrix has zero elements whenever
co- and crosspolarised channels are multiplied (see Section 2.4.2).
The second is a more restrictive assumption, as it may not be true in the presence
of surface slopes or in heterogeneous regions, where discrete point scatterers
exist (in urban areas, for example). It also forces the use of a large number of
looks to reduce bias in the estimation of zero cross-products. For this reason,
calibration must be applied over very flat homogeneous regions containing
strong depolarising effects (volume scattering). Flat, forested areas (such as the
Amazon basin for spaceborne sensors) are typical of regions of choice. On the
other hand, urban and mountainous regions (even if vegetated) must be avoided
in the calibration process, as they are likely to violate reflection symmetry. In
practice these are masked out of the calibration by first estimating the co-/cross-
correlations and rejecting pixels where this is high (Papathanassiou, 1998a;
Kimura, 2004).
The detailed algorithm is shown in equation (9.22). Starting on the right
we form the Quadpol scattering vector for each pixel, and then average over
azimuth to obtain a 4 × 4 covariance matrix [C] (having also masked out pixels
with low SNR as well as those which violate reflection symmetry). These
elements can then be used to solve for all the unknowns:
hh c11 c12 c13 c14
hv + , c c22 c23 c24
k clutter = ∗T 21
vh ⇒ [C] = k clutter k clutter = c31
c32 c33 c34
vv c41 c42 c43 c44
354 Applications of polarimetry and interferometry
s1 S SXY px
= XX (9.25)
s2 SYX SYY py
+ ∗ , + ∗ ,
s1 = SXX px + SXY py |px |2 +|py | =1
2
s s +s1 s2∗ , = JXX JXY
−→ [J ] = + 1 1∗ ,
s2 = SYX px + SYY py s2 s1 s2 s2 JYX JYY
(9.26)
This matrix has only four parameters, while the full scattering coherency matrix
has up to sixteen. However, by assuming two symmetries in the scattering
process we can reduce this discrepancy. The first—reciprocity in backscatter
forces Sxy = Syx and the full coherency matrix—then has rank 3 with nine
parameters. The second—reflection symmetry with an axis aligned parallel
to x or y—forces cross-products involving mixed co- and crosspolar terms to
zero; that is, SXX SXY = SYY SYX = 0. This reduces the scattering coherency
matrix [T ] and covariance matrix [C] in the xy basis to the reduced 3 × 3 forms
shown in equation (9.27), both of which have only five unknowns.
t11 t12 0 c11 0 c13
[T ] = t12
∗ t22 0 ⇔ [C] = 0 c22 0 (9.27)
0 0 t33 ∗
c31 0 c33
Here we have four equations in five unknowns, and so cannot solve for all
five elements of [C], whatever the choice of px and py . There are then three
important special cases of compact polarimetry that arise in practice. They all
derive from the choice of xy = HV; that is, for linear horizontal and vertical on
receive. In the simplest case we can then choose px = 1, py = 0—fixed horizontal
transmit. In this case the [P] matrix takes the following form:
1 0 0 0 0
0 1 0 0 0
px = 1, py = 0 ⇒ [P] = 0 0 0 0 0
(9.29)
0 0 0 0 0
Here we see that [J ] then only contains information about the scattered power in
co- and cross-channels. (Remember that we are assuming reflection symmetry
and so the HH and HV channels are uncorrelated.) In order to access information
related to the other elements a different choice of px and py are required. In the
π/4
B√ compact mode, for example, the transmitter is set to 45◦ linear—px = py =
1 2—and the matrix [P] takes the form shown in equation (9.30):
1 1 0 0 0
1 1 1 0 1 1 0 0
px = √ , py = √ ⇒ [P] =
(9.30)
2 2 2 0 1 0 1 0
0 0 0 0 1
We note two important aspects of this mode. The first is that we now have access
to linear combinations of all the elements of [C], and hence some sensitivity
to all the elements of the covariance matrix. The second is the factor of 1/2 in
front of the matrix [P]. This implies a 3-dB loss of signal compared to a full [S]
matrix system. Such signal loss is an inevitable consequence of mismatching the
transmitter and receiver bases. Finally, another√ mode that has√ been proposed
is to transmit circular polarisation:px = 1/ 2, py = ±i/ 2. This case is
very similar to the π /4 mode, but with a [P] matrix of the form shown in
equation (9.31):
1 1 0 0 0
1 ±i 1 0 1 1 0 0
px = √ , py = √ ⇒ [P] = (9.31)
2 2
2 0 0 0 0 ±1
0 ±1 0 ∓1 0
Again we see that there is a 3-dB loss of signal due to antenna mismatch, but
again there is information from all components of [C] present in the mixture.
9.3 Polarimetric synthetic aperture radar (POLSAR) 357
[T ] = [Ts ] + [TV ]
t11 0 0 9 : 9 :
[TV ] = 0 t22 0 ⇒ 4 |SXY |2 = |SXX − SYY |2
0 0 t22
9 : 9 : 9 :
∗ ∗
= |SXX |2 + |SYY |2 − 2Re(SXX SYY ) = 2 |SXX |2 − 2Re(SXX SYY )
+ , ∗ )
4 |SXY |2 2Re(SXX SYY
⇒+ , + , = 1 − + , + ,
|SXX |2 + |SYY |2 |SXX |2 + |SYY |2
2Re(SXX SYY ∗ )
=1− + , = 1 − |γXXYY |
2 |SXX |2
+ ,
|SXY |2 1
⇒+ , + , = (1 − |γXXYY |) (9.32)
|SXX | + |SYY |
2 2 4
To check this we first ask what happens in the limit as the volume tends to
zero and we are left with bare surface scattering. In this case (according to the
RVOG model) the coherency matrix is rank-1 (or with very small secondary
eigenvalues), and is therefore represented by a symmetric scattering matrix,
which we also assume is diagonal in the XY basis (due to Bragg scattering
358 Applications of polarimetry and interferometry
from a flat surface, for example). Hence it has zero crosspolarisation combined
with a high polarimetric coherence equal to unity. We see that this combination
is still consistent with equation (9.32). For the general mixed case between
these two extremes we can adopt a simple two-component decomposition as
shown in equation (9.33):
cos2 α sin α cos αeiδ 0 0.5 0 0
[T ] = [Ts ] + [TV ] = ms sin α cos αe−iδ sin2 α 0 + mv 0 0.25 0
0 0 0 0 0 0.25
1 + sin 2α cos δ 0 cos 2α + i sin 2α sin δ 3 0 1
ms mv
⇒ [C] = 0 0 0 + 0 2 0
2 8
cos 2α − i sin 2α sin δ 0 1 − sin 2α cos δ 1 0 3
(9.33)
Here we combine two components: one a rank-1 surface mechanism with mag-
nitude ms , and the second a random dipole cloud with scattering cross section
mv . This is very similar to the Freeman decomposition (see Section 4.2.4) or the
RVOG model (see Section 7.4.1). We can now express the cross-to-copolarised
scattering ratio and HH/VV coherence as functions of the surface-to-volume
scattering ratio µ = ms /mv and scattering mechanisms α and δ, as shown in
equation (9.34):
+ ,
4 |SHV |2 1
+ , + ,=
|SHH | + |SVV |
2 2
2µ + 32
(9.34)
1
+ µ(cos 2α + i sin 2α sin δ)
1 − |γHHVV | = 1 − 4
9
µ2 (1 − sin2 2α cos2 δ) + 32 µ + 16
We can now use this to check the equality of equation (9.32) for arbitrary
mixtures of surface and volume scattering mechanisms. Figure 9.12 shows
some example calculations. Here we plot, along the x axis, the cross-to-copolar
ratio, and along the y axis one minus the coherence amplitude. For equation
(9.32) to be valid, therefore, we require the points to lie along a line at 45◦ .
We show how the two parameters vary for µ ranging from –30 dB to +30
dB; that is, from the limiting cases of zero surface to zero volume scattering.
We show the results for steps of 15◦ in alpha (always with δ = 0 to simplify
the situation), starting from zero. We note that for α = 0 the equality holds
for all mixtures, and that the two limiting points of zero surface (the origin)
and zero volume (when both approach 2/3) are also satisfied for all scattering
mechanisms, as expected. However, for alpha angles greater than 30◦ we note
significant departures from the model. In particular, for alpha = 45◦ we see that
we have a situation where the coherence can be zero even when there is low
crosspolarisation. This arises because one of HH or VV scattering coefficients
goes to zero for this mechanism.
More significantly we see that for all alpha values greater than 45◦ —
for dihedral scattering of all types—there is always a µ value that leads to
zero coherence, and consequently to large deviations from the simple linear
9.3 Polarimetric synthetic aperture radar (POLSAR) 359
0.9
0.8
0.7
1 - HHV Coherence
0.6
0.5 0 degrees
15 degrees
0.4 30 degrees
45 degrees
0.3 60 degrees
75 degrees
0.2 90 degrees
Compact Pol
0.1 Fig. 9.12 The variation of cross-to-
copolarised ratio versus 1-HHVV coherence
0 for varying mixture of surface and vol-
0 0.2 0.4 0.6 0.8 1
ume scattering and different scattering
Cross-to-Copolarised ratio mechanisms
relationship. We can see from equation (9.34) that this arises when cos2α is
negative, as it can then cancel the positive numerator contribution from the
volume scattering. As such this effect has its origin in the 180-degrees phase
shift caused by double reflection.
Despite these limitations, relations such as equation (9.32) are widely used
in the compact polarimetry community. The reason is that by combining this
result with [P] leads to a set of five (non-linear) equations in five unknowns
(the elements of [C]), as shown in equation (9.35), which can then be solved
by iteration for all the elements of [C] and therefore [T ] or the Mueller
matrix [M ].
c11
px px∗ py py∗ 0 0 0 JXX
c22
0 px px∗ ∗
py py 0 0
c33 = JYY
0 Re(px py∗ ) 0 Re(px py∗ ) −Im(px py∗ )
Re(c13 )
Re(JXY )
0 −Im(px py∗ ) 0 Im(px py∗ ) Re(px py∗ ) Im(JXY )
Im(c13 )
4c22 c13
+ √ −1=0
c11 + c33 c11 c33
(9.35)
In this way, compact polarimetry can be used to provide estimates for the
full coherency or covariance matrix elements without compromising the PRF
requirements of a single channel SAR system. Note, however, that this is
only true for the class of depolarisers satisfying the combined assumptions
of reciprocity, reflection symmetry and especially the equality between copolar
coherence and cross-to-copolarised power (which, as we have shown in Figure
(9.12), is the weakest assumption). In addition to the case shown above, the
compact assumptions also do not apply to general point scatterers (such as
rotated dihedrals or dipoles, for example) or to scattering from sloped terrain—
both of which introduce correlation between co- and cross-polarisations and
360 Applications of polarimetry and interferometry
x
kz 2-D IFT
2-D IFT P
P
P
1 P
2-D IFT z
kx W1
2-D IFT
k-space P
Image space P
x
kz 2-D IFT
2-D IFT P
P Interferometry
P W2
1 P
2-D IFT z Fig. 9.13 Schematic of steps involved
kx in polarimetric SAR Interferometry, or
2-D IFT
POLInSAR
Z1 T11 Z1∗T Z1 12 Z2∗T
[2 ] =
Z2 ∗T
12 Z1
∗T Z2 T22 Z2∗T
w∗T ∗T
1 Z1 12 Z2 w 2
⇒ γ̃ (w1 , w2 ) = (9.36)
w∗T ∗T ∗T ∗T
1 Z1 T11 Z1 w 1 · w 2 Z2 T22 Z2 w 2
Z2∗T −1 T22
w1 =w2
−1 ∗T −1
12 T11 12 Z2∗T w2 = λopt w2
−→
∗T −1 −1 −1 ∗T ∗T
Z1 T11 12 T22 12 Z1 w1 = λopt w1
−1 ∗T −1
T22 12 T11 12 w2 = λopt w2
⇒
T −1 T −1 ∗T w = λ w
11 12 22 12 1 opt 1
(9.37)
C
T11 C 12
= (9.38)
C ∗T T C
12 22
We can also derive the coherence region for compact polarimetry using con-
strained optimization, as we did in Section 6.2. The key difference here is
that we are now working in a two-dimensional rather than three-dimensional
complex space, and hence the coherence region will always be an ellipse (see
Section 6.2.4). It is interesting, from an applications point of view, to deter-
mine if we can use these ideas to approximate the true region shape by using
such a compact system. We return to this point in Section 9.5.3, when we
consider the coherence region for vegetation scattering obtained from chamber
measurements.
We now turn to consider some illustrative applications of polarimetry and
polarimetric interferometry.
– z +
< 0º > 0º
m
.56
inc
=9
sct
RO
2.0 m Y
0.3 m
Turntable
Surface backscatter
Wideband scattering matrix measurements were made in monostatic mode
(backscatter) over the frequency range 1–19 GHz and incidence angles θ of
10–50◦ (in 5 or 10-degree steps). For each angle of incidence the turntable is
rotated through 360◦ in 5-degrees steps. In this non-imaging case, averaging
9.5 Applications of polarimetry and interferometry 365
–5
Sattering coefficient (dB)
–10
–15
–20
HH
HV
–25 VH
VV
–30
–35
–40
2 4 6 8 10 12
Fig. 9.16 Rough surface backscatter in linear
Frequency (GHz) basis as a function of frequency
is made over 360◦ of azimuth coverage (72 samples) combined with some fre-
quency smoothing over a 160-MHz bandwidth (sixteen frequency steps). By
averaging over such combined azimuth/frequency variations we ensure that
the surface has a scattering coherency matrix of the reflection symmetric form
shown in equation (2.75). Starting with the rough surface, Figure 9.16 shows
the backscatter cross-section as a function of frequency for 30-degree angle of
incidence. Here we show the cross-section in the linear basis, HH, HV VH and
VV, being the diagonal elements of the covariance matrix [C]. We see that there
is little change with frequency over the band, and that the copolarised chan-
nels HH and VV are some 10 dB greater than the crosspolarisation channels.
As expected, we see that HV = VH (due to reciprocity and good polarimetric
calibration of the data). Although the copolarised channels are also equal in
amplitude (HH = VV), this is due to another reason: the particular scattering
366 Applications of polarimetry and interferometry
–5
–15
–20
lambda 1
lambda 2
–25 lambda 3
lambda 4
–30
–35
–40
2 4 6 8 10 12
Fig. 9.17 Coherency eigenvalue spectra for
rough surface scattering Frequency (GHz)
–5
–10
Scattering amplitude (dB)
–15
–20
–25
–30
–35
–40
–45
10 15 20 25 30 35 40 45 50
Fig. 9.18 Variation of coherency eigenvalue
Angle of incidence spectra with angle of incidence
10–1
10–2
be estimated from the data for the purposes, for example, of surface parameter
estimation.
Another way to represent this information is to use the entropy/alpha
approach. Figure 9.20 shows the results when applied to this rough surface
scattering data. In the upper graph we show the entropy, which reduces with
increasing angle of incidence. This is just another way of representing the angle
of incidence dependence of the depolarised eigenvalues in Figure 9.19. We see
that at small angles of incidence the entropy is over 0.6, while at larger angles
it falls to around 0.1. However, by comparing Figures 9.19 and 9.20 we realize
368 Applications of polarimetry and interferometry
Entropy
0.4
* *
0.2 *
*
0
10 15 20 25 30 35 40 45 50
Angle of incidence
5
*
* * *
Fig. 9.20 Entropy (upper) and alpha angle 0 *
(lower) values for rough surface scattering as 10 15 20 25 30 35 40 45 50
a function of angle of incidence Angle of incidence
that an entropy of 0.6 still represents a strongly polarised signal. In the lower
portion of Figure 9.20 we see the corresponding alpha angle. If the eigenvector
were truly of the form (1,0,0), this should be zero. We see that experimentally
it lies around 3◦ .
In conclusion, we have seen that rough surface backscattering represents
only a weak depolariser, with an isotropic noise-like depolarisation subspace
and a dominant eigenvector with scattering entropies below 0.6. However, the
polarised eigenvector itself seems rather trivial: just the coherent sum of the HH
and VV channels. The question is, do we ever obtain more interesting variation
of eigenvectors, allowing us to use variation of the polarised ratios for parameter
estimation? To answer this we turn attention to the smooth surface scattering
behaviour.
The smooth surface is characterized by a smaller rms roughness s (although
the same correlation length), and hence at a given frequency the product βs
will be smaller and the surface electrically smoother. With this in mind we
show, in Figure 9.21, the variation of linear basis backscatter as a function of
frequency (at a 30-degree angle of incidence). Here we see two features of
interest. The first is a lower backscatter level compared to the rough surface,
with the crosspolarised response now around 20 dB below the copolarised.
The second feature of interest is the separation of copolar coefficients at low
frequencies. Here we see that VV is a few dB above HH, in qualitative agreement
at least with the predictions of the small perturbation or Bragg scattering model.
Figure 9.22 shows the corresponding eigenvalue spectra of the coherency
matrix. Again we note the small fourth eigenvalue, as expected from reci-
procity, and also the presence again of a dominant eigenvalue, showing that
again the backscatter is strongly polarised. However, now we see that the depo-
larised subspace is anisotropic; that is, that the second and third eigenvalues
are not equal, demonstrating that the depolarisation caused by the surface is
9.5 Applications of polarimetry and interferometry 369
HH
0 HV
VH
VV
–10
Scattering coefficient (dB)
–20
–30
–40
–50
–60
2 3 4 5 6 7 8 9 10
Fig. 9.21 Smooth surface backscatter in lin-
Frequency (GHz) ear basis as a function of frequency
lambda 1
–10 lambda 2
lambda 3
lambda 4
Scatterinng coefficient (dB)
–20
–30
–40
–50
–60
2 3 4 5 6 7 8 9 10
Fig. 9.22 Coherency eigenvalue spectra for
Frequency (GHz) smooth surface scattering
now polarisation dependent. Again we can expose the polarised nature of the
scattering by normalizing the eigenvalue spectra to have unit sum. Figure 9.23
shows the results as a function of angle of incidence (for a fixed low frequency
of 2 GHz). Here we see a strongly polarised response, with the depolarised
power making up less than a few percent of the scattered signal. The aver-
age phase angle between HH and VV is small—typically a few degrees, as
shown in Figure 9.24—again agreeing qualitatively with predictions from the
Bragg surface scattering model. However, the main new feature of interest is
the change in the eigenvector parameters with incidence angle. Figure 9.25
shows how the amplitude of the Pauli components of the dominant eigenvector
370 Applications of polarimetry and interferometry
10–1
10–2
10–3
Fig. 9.23 Variation of normalized eigenvalue 10 15 20 25 30 35 40
spectra with angle of incidence for smooth
surface scattering Angle of incidence
HH/VV Phase
20
15
10
5
Phase (degrees)
-5
-10
-15
-20
Fig. 9.24 HH/VV phase for the smooth sur- 2 3 4 5 6 7 8 9 10
face at 30-degree angle of incidence Frequency [GHz]
vary across the spectrum. Here we see that while the first component is still
dominant, the second Pauli component is non-zero and increases with angle of
incidence. This is reflected in the corresponding entropy/alpha variation shown
in Figure 9.26. Here we see low scattering entropy at all angles, but combined
with an alpha parameter that steadily increases with angle of incidence. This
variation is directly related to the dielectric constant of the surface, as we now
demonstrate.
We begin by noting that the simple first-order Bragg surface scattering model
can be used to estimate an alpha parameter, which is a function of the ratio of sum
9.5 Applications of polarimetry and interferometry 371
0.8
0.6
0.4
0.2
0.2
0.1 * *
* * *
*
0*
10 15 20 25 30 35 40
Angle of incidence
15
*
10 *
*
5 *
* *
*
0 Fig. 9.26 Entropy (upper) and alpha angle
10 15 20 25 30 35 40
(lower) values at 2 GHz for smooth surface
Angle of incidence scattering as a function of angle of incidence
18
16
14
10
surface backscatter. We saw in Section 3.2 how the extended or X-Bragg model
provides a method for parameterising the effects of depolarisation on the Bragg
model.
According to this approach (see equation (3.40)), the alpha parameter must be
corrected for depolarisation by estimating it—not directly from one eigenvector,
but from a ratio of diagonal terms of the full coherency matrix, as shown in
equation (9.40):
t22 + t33
αb = tan−1 (9.40)
t11
Figure 9.27 shows how these three estimates compare for low-frequency (2-
GHz scattering). In the solid line we again show the eigenvector estimates from
the lower part of Figure 9.26. We show too the reference Bragg values obtained
from the dielectric constant. Finally, in dash we show the corrected estimated
alpha values using the X-Bragg model. We see that the correction is of the
order of 2◦ , and for angles of incidence greater than 20◦ the correction brings
the estimates into close agreement with the Bragg predictions. However, for
angles less than 20◦ , both methods seem to overestimate alpha. This can be
traced to the small separation of HH and VV scattering coefficients for angles
near normal incidence, which makes estimation of small alpha values from
experimental data more difficult.
The importance of this depolarisation correction can be made even more
apparent by considering a higher frequency. At 10 GHz, for example, the sur-
face roughness leads to βs = 0.84, which is well outside the usual bounds for
validity of the simple Bragg model. Figure 9.28 shows the various alpha esti-
mates for this high-frequency case. Here we see that the maximum eigenvector
dramatically underestimates the alpha parameter, especially at high incidence
angles, whereas when we add the correction for depolarisation (dashed line)
9.5 Applications of polarimetry and interferometry 373
18
16
14
Alpha angle (degrees)
12
10
we see much better agreement. This result can, for example, be used to esti-
mate surface dielectric constant by first using the X-Bragg model to estimate
alpha corrected for surface depolarisation. From alpha, and knowing the angle
of incidence, we can then obtain a direct estimate of the dielectric constant by
using the standard Bragg relations.
In this section we have seen that surface backscattering provides a polarised
return over a wide range of roughness scales and angles of incidence. We now
turn to consider the generalization of these ideas to bistatic surface scattering—
the situation in which transmitter and receiver are separated.
Metres
0 + + +
+ + +
–2
–4
–6
–8
10–1
Eigenvalue level
10–2
10–3
Fig. 9.30 Normalized eigenvalue spectra for –30 –20 –10 0 10 20 30 40 50
bistatic scattering from rough surface at
10 GHz Scattering angle
The Pauli components of the dominant eigenvector are shown in Figure 9.31.
For angles close to backscatter we again see a dominant first component, corre-
sponding to alpha around zero. However, as the scattering angle increases so the
second Pauli component becomes more important. We note that the third and
fourth remain equal and close to zero over the full range of angles. Figure 9.32
shows the corresponding entropy and alpha variations. Note that the entropy
is now defined in base 4, to account for the fourth eigenvalue. We see a dip in
9.5 Applications of polarimetry and interferometry 375
1
Eigen vector component level
0.8
0.6
0.4
0.2
Bistatic entropy
0.5
0.4
Entropy
0.3
0.2
0.1
0
–30 –20 –10 0 10 20 30 40 50
Scattering angle
Principal eigenvector of T
Alpha angle (degrees)
40
30
20
10
0
–30 –20 –10 0 10 20 30 40 50
Fig. 9.32 Bistatic entropy (upper) and alpha
Scattering angle angle (lower) for bistatic
the entropy for specular scattering. The alpha angle is taken from the dominant
eigenvector, and shows a steady increase with scattering angle, with a trend
similar to that found in Figure 9.28. However, in this case the physical origin
of this variation is quite different. Here we must use the bistatic geometry and
dielectric constant in the BRDF model of Section 3.2.2 (Priest, 2000). This
model can be used to predict the variation of alpha for the bistatic scattering
matrix. In Figure 9.33 we show the amplitude of the Pauli components of this
matrix as a function of scattering angle. Superimposed, we again show the Pauli
376 Applications of polarimetry and interferometry
0.6
0.4
0.2
components for the dominant eigenvector. We see that the simple BRDF model
is quite good at explaining the trend of the alpha variation.
In conclusion, we have seen from our analysis of rough and smooth surface
scattering data that bare surface scattering is characterized by a strong polarised
return (the entropy is seldom above 0.5), maintained over a wide range of
angles of incidence and frequencies. The corresponding dominant eigenvector
is characterized by an alpha parameter in the range 0 ≤ α ≤ π/4, and which
shows dependence on local scattering geometry (angle of incidence and bistatic
scattering angle) and surface dielectric constant. We now turn to consider a
different class of interactions: volume scattering, where scattering entropies
and hence depolarisation can be much higher.
Hence we can in general have four non-zero eigenvalues, but the eigenvectors
are limited in their structure to C2—a two-dimensional complex subspace.
Hence we can use the scattering sphere concept (see Section 2.4.3.2) to represent
variations in the eigenvector information. We choose to consider three special
cases: a single particle for reference, a low concentration of 0.5%, and a high
concentration of 16%. The first parameter of interest is the phase function. This
378 Applications of polarimetry and interferometry
Phase function
103
1 particle
5 particle
102 160 particle
101
M11
100
10–1
is just the m11 element of the Stokes scattering matrix, normalized so that
π
1
m11 (ψ) sin ψd ψ = 1 (9.42)
2
0
Equivalently, m11 represents one half the trace of the coherency matrix (or sum
of the eigenvalues of [T ]), which physically represents the total scattered power
of the signal. Figure 9.35 shows how this function varies for three different par-
ticle concentrations (scattering angle = 0 is forward scatter in this notation). We
immediately see that the particles scatter more in the forward than the back-
ward direction (a typical feature of wavelength size particles), and that as more
particles are added there is coherent addition in the forward direction, and also
the appearance of a small enhanced backscatter peak in the backscatter direc-
tion (scattering angle equal to 180◦ ). We see that the phase function becomes
smoother, compared to the single particle case, as we add more particles to the
volume. This is the effect of multiple scattering.
We are now particularly interested in the depolarisation properties of this
process. The single particle case is trivial to consider, as it yields a single non-
zero eigenvalue for all scattering angles—the eigenvector corresponding to
which is obtained from the scattering matrix [S] for the particle (and which can
be obtained exactly using standard Mie scattering theory; see Section 3.3.2). Of
more interest are the low and high concentration results. Figures 9.36 and 9.37
show the normalized coherency spectra as a function of scattering angle for
these two cases.
In the low concentration example we see a dominant eigenvalue, with the
secondary eigenvalues less than 1% of the scattered power, except for a couple
of peaks around 60◦ and 100◦ . In forward scatter the entropy is zero (a coherent
wave propagating through the volume), and in backscatter one of the minor
eigenvalues goes to zero (because of reciprocity), while the other two are equal
9.5 Applications of polarimetry and interferometry 379
lambda 1
lambda 2
lambda 3
10–1 lambda 4
Relative amplitude
10–2
10–3
10–4
0 20 40 60 80 100 120 140 160 180
Fig. 9.36 Normalized coherency spectra for
Scattering angle low concentration of particles
10–1
Relative amplitude
10–2
lambda 1
lambda 2
lambda 3
10–3 lambda 4
10–4
0 20 40 60 80 100 120 140 160 180
Fig. 9.37 Normalized coherency spectra for
Scattering angle high concentration of particles
and lie around the 1% mark. The denser concentration starts to lose structure,
as multiple scattering dominates. Here we see a smoother variation of eigen-
values, still with zero entropy in forward scatter, but much higher entropy in
the backscatter hemisphere. Note again that for exact backscatter one of the
eigenvalues still falls to zero as a result of reciprocity. These results can be
expressed in terms of the scattering entropy, as shown in Figure 9.38. Here we
show the variation of entropy for the three cases. The single particle case yields
zero entropy for all scattering angles, while the denser concentration leads
to more depolarisation and increased entropy, especially in the backscattering
hemisphere.
380 Applications of polarimetry and interferometry
Bistatic entropy
0.9
0.8
0.7
0.6
160 particle
Entropy
0.5
0.4
0.3
0.2 5 particle
0.1
0 1 particle
–0.1
0 20 40 60 80 100 120 140 160 180
Fig. 9.38 Scattering entropy of a cloud of
spheres versus scattering angle for three par-
ticle concentrations Scattering angle
We see that the entropy peaks around 0.8 for the denser concentration but
then reduces slightly for backscatter. However, if we ignore the fourth eigen-
value in the backscatter direction (because reciprocity, rather than scattering
symmetries, forces it to zero) and calculate the monostatic entropy, we find that
the backscatter entropy rises from 0.75 to 0.94. This is close to the maximum
entropy obtained from single scattering by a cloud of prolate particle shapes,
even though we are dealing with a cloud of spheres. In this way we see that
multiple scattering can be an important source of depolarisation.
Turning now to the dominant eigenvector, we start by noting that it has only
two non-zero elements: the first and second Pauli components. Figure 9.39
shows how these two vary in amplitude for each of the three cases considered:
single particle, five particles, and 160 particles. We see that in forward scattering
we have the identity matrix (first Pauli component) as consistent with forward
scattering in the FSA system. In backscatter, by contrast, we have the second
Pauli component with the first falling to zero. This corresponds to a backscat-
tering matrix with equal amplitudes but 180-degree phase shift between HH
and VV, as expected in the FSA system. Between these two extremes we see
that the eigenvectors change significantly. Indeed, these changes not only occur
in amplitude but in phase. As the eigenvector is an element of C2—the two-
dimensional complex space—we can map it in both amplitude and phase on the
surface of the scattering sphere. This allows us to visualize changes in the com-
plex nature of the eigenvector between forward and back scattering. Figure 9.40
shows the dominant eigenvector variation on the scattering sphere for the three
cases considered. In forward scattering they all begin on the equator at zero
phase. As scattering angle changes we then see combined amplitude and phase
variations trace a complicated loci across the sphere, ending for backscatter
on the equator again at 180◦ . For the single-particle case (the black line) these
are just the polarimetric phase and amplitude variations due to Mie scattering
by a lossless dielectric sphere. In the low concentration case we see that the
9.5 Applications of polarimetry and interferometry 381
0.9 0.9
0.8 0.8
0.7 0.7
Component magnitude
Component magnitude
0.6 0.6
0.5 0.5
0.4 0.4
0.3 0.3
0.6 N=5
150 30
0.4
N = 160
0.2
180 0
210 330
50
40
30
20
10
m
56
Vegetation 9.
0=
sample
R
EMSL focal point
38 cm Y
Turntable
Normalized eigenvalues
1
0.9
0.8
0.6
lambda 3
0.5 lambda 2
lambda 1
0.4
0.3
0.2
0.1
0
2 3 4 5 6 7 8 9
Fig. 9.44 Variation of normalized eigenvalue
spectra with frequency for the maize sample Frequency (GHz)
1
Pauli components of eigenvector
Pauli 1
0.8
Pauli 2
Pauli 3
0.6
0.4
0.2
0
2 3 4 5 6 7 8 9
Fig. 9.45 Variation of Pauli components of
the maximum eigenvector (maize sample) Frequency (GHz)
to around 60–70% of the signal energy. The depolarisation is quite large across
the spectrum, and shows some anisotropy at low frequencies, which reduces as
the frequency increases.
Note that here we have increased the averaging window to 320 MHz (in
addition to averaging over the 72 azimuth samples) in order to reduce the speckle
noise on the estimates. The Pauli components of the maximum eigenvector vary
with frequency, as shown in Figure 9.45. Here we see that at low frequencies
the second Pauli component is significant, but that as frequency increases the
9.5 Applications of polarimetry and interferometry 385
Scattering entropy
1
0.9
Entropy
0.8
0.7
0.6
2 3 4 5 6 7 8 9
Frequency (GHz)
Alpha parameter
40
Alpha (degrees)
30
20
10
0
2 3 4 5 6 7 8 9 Fig. 9.46 Entropy (upper) and alpha (lower)
Frequency (GHz) variation with frequency for maize sample
Entropy/Alpha Diagram
90
80
70
60
50 Ap = 20dB
Alpha
40
Ap = 6dB
30
20
10
eigenvector tends to the first Pauli component. These results all indicate a high
level of wave depolarisation by the vegetation sample. This is confirmed in
Figure 9.46, which shows the entropy/alpha variations corresponding to Figures
9.44 and 9.45. Here we see high entropy, around 0.8 at low frequency and falling
slightly to 0.7 at higher frequencies. The dominant eigenvector alpha falls from
around 20◦ at low frequencies to nearly zero at high frequencies. However,
386 Applications of polarimetry and interferometry
Entropy/Alpha diagram
42
Ap = 10dB
40
38
36
Ap = 8dB
34
Alpha
32
30
28
26
24
given the relatively high entropy we can expect the maximum eigenvector to
be corrupted by depolarisation (as we saw in the cloud of spheres in Figure 9.38,
for example). Hence we again need some way of compensating the eigenvector
information for the depolarisation that is occurring. One way to do this is
to employ the average alpha, formed by the sum of the alpha values for each
eigenvector weighted by the normalized eigenvalues (see Section 2.4.2.4). This
leads to a representation on the entropy/alpha plane as shown in Figure 9.47,
with a zoom on the data points shown in Figure 9.48. Also superimposed on
these diagrams are the entropy/alpha loci for a cloud of anisotropic particles,
as developed in Section 3.4.1. We show results for prolate particle clouds of
varying orientation structure, and the same for oblate particle clouds. One way
to interpret the maize data points is in terms of the response for an equivalent
cloud of such particles. In this regard we see that at low frequencies the response
is equivalent to a cloud of strongly prolate particles (with a shape anisotropy
factor Ap around 10 dB), but with a non-random orientation distribution; while
as frequency increases two things happen: initially (up to 6 GHz) the cloud
becomes more random, while maintaining the same particle anisotropy. From 6–
9 GHz, however, the behaviour changes, the effective particle shape anisotropy
(Ap ) reduces (which causes a drop in the entropy), and the distribution stays
random (the lower curve of the entropy/alpha diagram represents azimuthal
symmetry in the volume).
However, this interpretation—in terms of the response of an equivalent cloud
of anisotropic particles—involves many simplifying assumptions, one of the
most important being that we see only volume scattering from the maize and
that the underlying surface plays no part in the scattered return. This is not
necessarily a good assumption, especially at low frequencies. We can see,
for example, that there is a change in scattering behaviour around 6 GHz, and
cannot dismiss the idea that for frequencies lower than this the surface response
is playing an important role.
9.5 Applications of polarimetry and interferometry 387
To resolve these issues we could try using one of the many model-based
decomposition methods based on mixed surface and volume scattering, but
these require us to make even more stringent assumptions about the nature of
the volume scattering. (The Freeman–Durden, for example, assumes a cloud
of dipole particles.) These assumptions are not supported for this example,
and so would not necessarily help us resolve issues about the true ratio of
surface-to-volume scattering. Instead we choose to look at the role that radar
interferometry can play in helping resolve this by adding additional information
about the structure of such complex scattering problems.
pq pq
s1 = Eθ (f ) s2 = Eθ+θ (f ) (9.43)
The wide-band interferogram is then formed from the product of common spec-
tral band filtered and phase offset signals, as shown in equation (9.44), where
the phase offset is required because the chamber focus lies below the surface
of the sample container:
f θ 4π f θ
s1 .s2∗ e−i c sin θ zo
pq pq
= Eθ (f ) · conj Eθ+θ f − (9.44)
tan θ
+ ∗,
s1 s2
γ̃pq (f ) = + ,+ , (9.45)
s1 s1∗ s2 s2∗
Figure 9.49 shows how this coherence amplitude and phase vary for three
polarisations as a function of frequency for the smallest baseline (0.25◦ ). We
see that the coherence falls as frequency increases. This is caused by volume
388 Applications of polarimetry and interferometry
0.8
Coherence
0.6
VV
0.4
VH
0.2 SINC
2 3 4 5 6 7 8 9
Frequency (GHz)
2
1.5
1
0.5
Fig. 9.49 Variation of interferometric coher-
ence amplitude (upper) and phase (lower) for 0
2 3 4 5 6 7 8 9
0.25-degree baseline and for polarisations VV
and VH Frequency (GHz)
9.5 Applications of polarimetry and interferometry 389
0.6 SINC
0.4
0.2
0
2 3 4 5 6 7 8 9
Frequency (GHz)
1
0
–1
–2
–3 Fig. 9.50 Variation of interferometric coher-
2 3 4 5 6 7 8 9 ence amplitude (upper) and phase (lower) for
0.5-degree baseline, and for polarisations VV
Frequency (GHz) and VH
To confirm this we can use these complex coherences to estimate the vertical
structure function using coherence tomography (Cloude, 2007a; Zhou, 2006).
In this case we know from the chamber geometry the true phase of the sur-
face position and the height of the vegetation, and so can use these directly to
reconstruct the profile using the baseline geometry. Figure 9.52 shows how the
interferometric wavenumber β z varies across the spectrum for the two baselines.
This can then be used with the known vegetation height (1.8 m) to calculate
the normalization parameter kv , as shown in Figure 9.53. We see that it varies
from around 0.4 for the small baseline at low frequencies to more than 4.0 for
the larger baseline at high frequency. These values provide good sensitivity to
changes in structure function in the vegetation layer (see Section 8.4). These
can then be used with the known surface phase to reconstruct a truncated Leg-
endre series expansion of the structure function. With one baseline we obtain
a second-order expansion, while by combining data from two baselines we
obtain a fourth-order reconstruction. Figures 9.54 and 9.55 show the recon-
structed vertical profiles through the vegetation as a function of frequency for
the single and dual baseline data. We note the following important features:
1. The dual baseline data has higher vertical resolution than the single.
This is due to the presence of higher-order terms in the Legendre series
reconstruction.
2. Both single and dual baseline datasets show a strong surface component
at low frequencies (below 6 GHz), with a more dominant volume con-
tribution for frequencies above this. However, the volume component
is not a simple exponential function. It shows a small response from the
top with a peak near the centre of the layer. This therefore represents a
390 Applications of polarimetry and interferometry
Coherence
0.6 SINC
0.4
0.2
0
2 3 4 5 6 7 8 9
Frequency (GHz)
1
0
–1
–2
Fig. 9.51 Variation of interferometric coher- –3
ence amplitude (upper) and phase (lower) for 2 3 4 5 6 7 8 9
0.5-degree baseline, and for two polarisations
HH-VV and HV Frequency (GHz)
Interferometric wavenumber
5
4.5
0.1 degree baseline
4 0.2 degree baseline
3.5
Vertical wavenumber
2.5
1.5
0.5
Wavenumber/height product
4.5
4
0.1 degree baseline
3.5 0.2 degree baseline
2.5
Kv
1.5
0.5
Fig. 9.53 Normalized wavenumber/height
0 product factor kv as a function of frequency
1 2 3 4 5 6 7 8 9 10
for the two baselines used in the EMSL
Frequency (GHz) chamber
HH Channel
2 1
Height (m)
1 0.5
0 0
2 3 4 5 6 7 8 9
Frequency (GHz)
HV Channel
2 1
Height (m)
1 0.5
0 0
2 3 4 5 6 7 8 9
Frequency (GHz)
VV Channel
2 1
Height (m)
1 0.5
Fig. 9.54 Reconstructed single-baseline ver-
0 0 tical profiles for the 1.8-m vegetation layer
2 3 4 5 6 7 8 9
in HH (upper), HV (centre), and VV (lower)
Frequency (GHz) polarisations
Figure 9.56 for the single (upper) and dual (lower) baseline data. Here we
see a strong surface component across most of the spectrum, confirming
that the structure function in this Pauli channel remains localized around
the surface.
In conclusion, we have seen that the presence of vegetation on a surface causes
complexity in backscatter, with an increase in the scattering entropy and hence
in the level of wave depolarisation. This causes a drop in polarimetric coherence
and hence an erosion of our ability to exploit polarimetric phase information.
392 Applications of polarimetry and interferometry
HH Channel
2 1
Height (m)
1 0.5
0 0
2 3 4 5 6 7 8 9
Frequency (GHz)
HV Channel
2 1
Height (m)
1 0.5
0 0
2 3 4 5 6 7 8 9
Frequency (GHz)
VV Channel
2 1
Height (m)
1 0.5
Fig. 9.55 Reconstructed dual-baseline verti-
0 0
cal profiles for the 1.8-m vegetation layer in
2 3 4 5 6 7 8 9
HH (upper), HV (centre), and VV (lower)
polarisations Frequency (GHz)
HH-VV channel
2 1
0.8
1.5
0.6
1
0.4
0.5
0.2
0 0
2 3 4 5 6 7 8 9
HH-VV channel
1
2
0.8
Height (m)
1.5
0.6
1 0.4
0.5 0.2
Fig. 9.56 Reconstructed vertical profiles for
the 1.8-m vegetation layer in the HH-VV 0 0
channel for single (upper), and dual (lower) 2 3 4 5 6 7 8 9
baselines Frequency (GHz)
Scattering entropy
1
46
0.95
45 0.9
Angle of incidence
44 0.85
43 0.8
0.75
42
0.7
41 0.65
40 0.6
39 0.55
0.5
500 1000 1500 2000 2500
Fig. 9.58 Entropy image of MAWAS P-band
data Azimuth (m)
P band interferogram
HH interferometric coherence
1
50 0.9
0.8
Angle of incidence (degrees)
45 0.7
0.6
40
0.5
0.4
35
0.3
30 0.2
0.1
25 0
0 1000 2000 3000 4000 5000
Fig. 9.60 Interferometric coherence of the
Azimuth (m) MAWAS-E test area for P-band HH channel
Parameter Value
Wavelength 0.236 m
Launch date Jan 24th 2006
Orbit height 691 km
Orbit Repeat 46 days
antenna located on the left. Some key system parameters are also shown on the
right-hand side.
We note four key points in particular. The first is that the transmitted band-
width in polarimetric mode is only 14 MHz, which is much smaller than the
airborne E-SAR system, and hence the range resolution is poorer (around 11 m
9.5 Applications of polarimetry and interferometry 397
HH image HH image
0 0 0 0
10 10
–5 –5
20 20
–10 –10
30 30
–15 –15
Range (m)
Range (m)
40 40
–20 –20
50 50
–25 60 –25
60
70 –30 70 –30
80 80
–35 –35
90 90
–40 –40 Fig. 9.63 Satellite SAR Images (HH left,
0 20 40 0 20 40
and HV right) for a trihedral corner reflector
Azimuth (m) Azimuth (m) before Faraday rotation correction
slant range for PALSAR compared to 1.5 m for E-SAR). The second point is
the relatively large PRF in polarimetric mode (around 4 kHz), which as men-
tioned in Section 9.3, limits the range coverage of the sensor. In this case the
range swath width is reduced to around 15 km in slant range, which for 21.5-
degree incidence (the default mode for the satellite) translates to around 40 km
of ground range. Thirdly, we note that the satellite repeats its orbit only every
46 days, so that repeat-pass interferometry can be implemented only with this
minimum temporal baseline. However, such long baselines lead to large levels
of temporal decorrelation, and hence are limiting for quantitative applications
such as POLInSAR height retrieval and coherence tomography. Finally, we
note the high orbit of the satellite, by which propagation effects through the
ionosphere cannot be ignored.
To illustrate this we show data for calibration trihedral corner reflectors (see
Figure 1.21) deployed in Adelaide, South Australia. The data was collected on
10 June 2006 at 13:50 UT. Figure 9.63 shows SAR images in HH (left) and
HV (right) channels. This data has been calibrated using the scheme of Section
9.3.2, but has not been corrected for Faraday effects. From scattering theory
we expect the HV backscatter to be zero for the trihedral, but we see it is only
around 20 dB below the copolar signal. This is due to a small rotation of the
plane of polarisation through the ionosphere. The level of Faraday rotation can
be estimated using equation (4.83). The estimated level is around +3-degree
one-way rotation. When this is removed from each pixel of the image we
obtain the corrected imagery shown in Figure 9.64. Here we see much better
isolation and residual cross-talk levels below –30 dB. This nicely illustrates
how vector-wave propagation models can be used to improve data quality and
system calibration for spaceborne radar systems.
One important application of such radars is in the identification and char-
acterization of coherent (polarised, in our context) scattering points in radar
imagery (Ferro-Famil, 2003; Schneider, 2006). Figure 9.65 is an example, in
398 Applications of polarimetry and interferometry
HH image HV image
0 0 0 0
10 10
–5 –5
20 20
–10 –10
30 30
–15 –15
Range (m)
Range (m)
40 40
–20 –20
50 50
60 –25 60 –25
70 –30 70 –30
80 80
–35 –35
90 90
Fig. 9.64 Satellite SAR images (HH left, and –40 –40
0 20 40 0 20 40
HV right) for a trihedral corner reflector after
Faraday rotation correction Azimuth (m) Azimuth (m)
HH RCS image
50
100
150
200
Slant range (m)
250
300
350
400
450
500
550
Fig. 9.65 HH image of ALOS-PALSAR cor- 500 1000 1500 2000 2500 3000
ner reflector scene (Adelaide, South Aus-
tralia) Azimuth (m)
which is shown an expanded view of the corner reflector scene. (The three cor-
ner reflectors are seen—that at far left being used in Figures 9.63 and 9.64.) We
then use a 3 × 5 (range x azimuth) window centred on each pixel to estimate the
local entropy from the eigenvalues of the coherency matrix. Figure 9.66 shows
the distribution of entropy/alpha values for the whole scene. We note that the
bulk of background pixels lie with high entropy, but there are several distinct
bands of low entropy distributions with various alpha values. The trihedrals are
seen lying close to the origin (small alpha and low entropy). However, there
9.5 Applications of polarimetry and interferometry 399
Entropy/alpha diagram
90
80
70
60
50
Alpha
40
30
20
10
0
0 0.2 0.4 0.6 0.8 1 Fig. 9.66 Distribution of entropy/alpha val-
Entropy ues for all pixels in Figure 9.69
are clearly coherent points (low entropy) with alpha values greater than 45◦ ,
indicating second-order dihedral scattering.
The differences in alpha then relate directly to the boundary conditions for that
scattering process. In particular they relate primarily to the dielectric constant
of the reflection. Furthermore, given the small angle of incidence of the data
(typically a 24-degree angle of incidence in mid swath) and its small variation
across the limited range swath, this dependence is primarily determined by one
of the two surfaces involved, and not both (see Figure 3.15). For dry materials
(low εr ) we can even approach the Brewster angle for reflection on one of the
surfaces, so that we can observe either 0- or 180-degree phase shift between
HH and VV, depending on which side of the Brewster point we operate. A phase
shift of 180◦ will be seen for dielectric constants greater than 5 (the Brewster
angle corresponding to which is 66◦ , matching the 24-degree local incidence
angle of ALOS). Hence the alpha angle for dihedral scattering at small angles
of incidence can be less than π /4. We can quantify this for the geometry of
PALSAR by calculating the variation of the apparent alpha of the boundary
conditions as a function of the dielectric constant.
Figure 9.67 shows an example. Here we show the variation of alpha with
dielectric constant of the vertical surface (for εr = 10 for the horizontal). We
show results for three angles of incidence—23◦ , 24◦ , and 25◦ (covering the
swath ofALOS-PALSAR)—and see only a small change. We see that increasing
alpha is directly linked to increasing dielectric constant of the vertical surface,
and that we have high sensitivity to changes in dielectric constant. This provides
us with a way to estimate εr of such dihedrals directly from spaceborne radar
data. For example, we can use Figure 9.67 to relate alpha directly to dielectric
constant using a suitable curve fit. One key feature of the fit must be that when
α tends to π/2 so εr tends to infinity (a metallic dihedral has an alpha of π/2),
and hence the function must have a pole at π/2. We then obtain the following
400 Applications of polarimetry and interferometry
This relation then enables us to estimate the dielectric constant of the vertical
surface reflection in dihedral scattering directly from the alpha angle of the
dominant eigenvector.
These few examples help illustrate the potential of spaceborne radar
polarimetry. The successful fusion of interferometry with polarimetry from
space will have to await the development of single-pass spaceborne systems,
or at least short temporal baseline low-frequency repeat pass sensors. At the
time of writing, no such system is yet operational, but plans for the launch of
Tandem-X in 2009 will see deployment of a spaceborne single-pass POLInSAR
system at X-Band (Krieger, 2007). This will not only open new applications in
the fusion of interferometry with polarimetry, but also provide new stimulus to
continued study of our understanding and exploitation of the ‘memory’ effect
imprinted on polarised electromagnetic waves scattered by random media.
70
60
Alpha parameter
23 degrees
50
25 degrees
40
30
As such, this operation replaces every element of [A] by its ij element multi-
plied by the whole matrix [B]. This form is particularly useful for converting
402 Introduction to matrix algebra
where [I ] is the m x m identity matrix, defined such that [I ]·[A] = [A]·[I ] = [A].
[A]2 [A]n
exp([A]) = [I ] + [A] + + ··· + + ··· (A1.5)
2! n!
This function is widely used in polarisation algebra (see Appendix 2 for more
details). Finally, we note that for two different square matrices of the same
dimension [A] and [B] (both m × m), the products [A] · [B] and [B] · [A] are
both well defined but are not necessarily equal. Hence matrix multiplication
(unlike ‘ordinary’ multiplication) is non-commutative. The ability of matrices
to represent differences in the order of multiplications is a key attraction for
their use in problems involving rotations and transformations as encountered
in polarimetry.
The multiplicative scale factor in front of the inverse matrix a11 a22 –a12 a21 is
called the determinant of the matrix |A| or det([A]), and evidently problems
arise with the existence of the inverse when this determinant goes to zero. The
concept of determinant is important, and can be extended to arbitrary matrix
dimensions. In this case we define the determinant as a scalar obtained from an
n × n matrix [A], as shown in equation (A1.7):
a11 a12 . . . a1n
a21 a22 · · · a2n ! n !n
|A| = . .. .. .. = a C
ij ij = aij (−1)i+j Mij (A1.7)
.. . .
. j=1
j=1
an1 an2 . . . ann
Here Cij is called the cofactor associated with the element aij , and is in turn
related to a signed version of Mij , called the minor, associated with the element
aij . The minor is itself a determinant, but importantly for a matrix of reduced
dimension (n−1)—the matrix formed by eliminating the ith row and jth column
of [A]. In this way we can reduce every determinant by reduction of dimension
to calculation of a series of 2 x 2 sub-matrices. The summation on the right-
hand side of equation (A1.7) is called the Laplace expansion of determinant
by the ith row (where i can be chosen arbitrarily). Similar expressions can be
used for a Laplace expansion by the jth column, again involving minors and
cofactors.
For arbitrary dimensions the inverse can be now be formally written as
follows:
adj([A])
[A]−1 = adj([A]) = [Cij ]T (A1.8)
det([A])
also the cyclic property of the trace for triple products so that Tr(ABC) =
Tr(BCA) = Tr(CAB). One important consequence of this result is that the
trace is invariant to unitary similarity transformations of a matrix [A], since
Tr(U∗TAU) = Tr(AUU∗T ) = Tr(A). Hence the trace is equal to the sum of
the eigenvalues of [A], and represents, in polarimetry, an important invariant
quantity identified as the total scattered power by an object.
Note that sometimes in the literature the transpose symbol ‘T’ is omitted for
notational convenience, and that only the conjugate sign is used to indicate
both operations implicitly. There is then, of course, the possibility of ambiguity
of notation. To counter this, the ‘dagger’ symbol † is often used to represent
the combined transpose and conjugate operations, so [A]† = [A]∗T . Again, an
important class of matrices arises when [A]† = [A]. These are termed Hermitian
matrices, and arise a great deal in the development of polarised wave scatter-
ing and propagation. Cleary, such matrices must have real elements along the
diagonal and complex conjugate elements on matching off-diagonal locations.
If [A]† = −[A], then [A] is termed skew or anti-Hermitian, and again must
have zeros along the diagonal.
Another way to think about this relationship is that the matrix [A] is transformed
into the matrix [B] by operation of [P]. This is termed a similarity transfor-
mation of [A] by [P]. Often the matrix [P] is unitary, in which case equation
(A1.11) becomes a unitary similarity transformation and the matrices [A] and
[B] are unitarily similar. Similar matrices share many properties in common.
For example, the determinants are equal, so det([A]) = det([B]). More impor-
tantly, their eigenvalues are equal (although their eigenvectors are different).
An important variation of this scheme is the congruent transformation. In this
case, [A] and [B] are related again by a third matrix [P], but now in the form:
A classical problem in matrix algebra is to find the x vector that maximizes the
scalar. To do this we need to first normalize by the magnitude of the vector to
obtain Rayleigh’s quotient R, as shown in equation (A1.14):
x∗T [H ]x
R x = (A1.14)
x∗T x
[A]xN = 0 (A1.15)
A second important set are the eigenvectors e. A square matrix [A] of dimension
n × n has n such eigenvectors, defined by the following equation:
[A] e = λe
⇒ [A] e − λe = 0 (A1.16)
⇒ det([A] − λ[I ]) = 0
where λ is a scalar called the eigenvalue. In this sense we can consider [A]
to operate on a vector x, and equation (A1.16) states that for some special
vectors, e, this operation will leave the vector x unchanged, apart from multi-
plication by a scalar. This physical interpretation is useful in many applications
in polarimetry, although equation (A1.16) is also an important general mathe-
matical idea. The eigenvalues can be found from the zeros of the determinant
(generally an nth order polynomial with n complex roots), as shown in equation
(A1.16). It is of special interest to be able to express a general matrix [A] as
a function of its eigenvectors and eigenvalues. An important theorem, called
Schur’s theorem, gives us a systematic way to achive this. Schur’s theorem
states that for any square matrix [A] there exists a unitary matrix [U ], such
that [U ]−1 [A][U ] = [B] is upper triangular. The eigenvalues of [A] must be
shared by the similar matrix [B] and appear along its main diagonal. The unitary
matrix [U ] is then composed of the eigenvectors of [A] as its columns. Hence
we can write a general matrix [A] in terms of its eigenvalues and eigenvectors,
as shown in equation (A1.17):
λ1 δ12 · · · δ1n
0 λ2 · · · δ2n
∗T
[A] = [U ] . . .. .. .. · [U ] [U ] = e1 e2 . . . en
.. . . .
0 0 · · · λn
(A1.17)
An important class of matrices has an even simpler form, when the off-diagonal
δ terms are all zero and the matrix [B] is diagonal. Symmetric and Hermitian
matrices fall into this category. The proof for Hermitian matrices is very sim-
ple, since if [A] is Hermitian then so is [B], and the only way [B] can be
Hermitian (equal to its conjugate transpose) is if it is diagonal, with real eigen-
values. Hence we have proved that Hermitian matrices have real eigenvalues
and orthogonal eigenvectors—a result of great importance in polarimetry. In
general, the required condition that [B] is diagonal is that the matrix [A] be nor-
mal, which by definition means that it commutes with its conjugate transpose;
that is, [A][A]∗T − [A]∗T [A] = 0. Unitary matrices, for example, also satisfy
A1.10 Norms, condition number, least squares and the SVD 407
this condition (when the conjugate transpose equals the inverse), and so can
also be expressed in diagonal form.
Eigenvalue decompositions are also of interest for solving optimization prob-
lems. Taking, for example, Rayleigh’s quotient again (equation (A1.14)), we
return to the problem of how to maximize this ratio. To solve this we can employ
the Lagrange multiplier method of constrained optimization. We first set up a
Lagrangian function L as follows:
and then differentiate with respect to the variable x to find the stationary points
as zeros of the derivative, as shown in equation (A1.19):
dL
= [H ]x − λx = 0 (A1.19)
dx
This implies that the optimum solution for x is the eigenvector of [H ] corre-
sponding to the maximum eigenvalue λmax . This simple example illustrates how
eigenvalue decompositions and optimization theory can be formally linked.
Finally, we consider an extension of this concept to the singular value decom-
position, or SVD. The Schur decomposition employs a single unitary matrix
[U ]. If, on the other hand, we employ two unitary matrices [U ] and [V ], then
any matrix (even those that are not square) can be expressed as a function of a
purely diagonal matrix [D], as follows:
The diagonal elements of [D] are termed the singular values of the matrix [A]
and [U ], the left singular vectors, while [V ] is the set of right singular vectors.
In the general case when [A] has dimensions m × n, [U ] is m × m, [D] is m × n,
and [V ] is n×n. However, only p elements of [D] are non-zero. The dimensions
p and m−p define two important subspaces that have important applications,
as we now consider.
[A] x = b (A1.21)
There are two important subspaces to this problem. The range of the matrix
[A] is the space of all possible vectors b for which the equation is solvable.
408 Introduction to matrix algebra
The dimension of the range is called the rank. Secondly, the set of vectors x
which satisfy the homogeneous equation (b = 0) define the null space of [A].
If there is no null space, then [A] is of full rank. The SVD provides a useful
perspective on these subspaces. The columns of [U ] in equation (A1.20) that
are associated with non-zero singular values form an orthonormal basis for the
range of [A], while the columns of [V ] associated with the zero singular values
form an orthonormal basis for the null space.
These concepts are particularly important in the solution of least squares
problems. In these cases we are usually interested in finding solution vector x
that solves an overdetermined set of equations. Given the possibility that an
exact solution may not exist, we ask instead to find x that has the minimum
residual; that is, one that minimizes the function L, defined as:
2
L(x) = [A] x − b (A1.22)
The classical way to solve this is to expand L and differentiate with respect to
x and set to zero. This yields the so-called normal equations, as follows:
Ax − b2 = Ax − b T Ax − b
= xT AT Ax − 2xT AT b + bT b
∂ (A1.23)
= 0 ⇒ AT Ax − AT b = 0
∂x
−1 T
⇒ x = AT A A b
There are two issues to be faced with such a solution. Firstly, does the inverse
matrix exist at all; and secondly, how sensitive is the solution x to perturbations
or small errors in the vector b? If there is any serious amplification of such errors
we say the equations are ill-conditioned. The SVD provides an important insight
into both these situations, as we now consider.
The generalized or pseudo-inverse of [A], designated [A]+ , can be defined
from the SVD as follows. Writing the function L in terms of the SVD we obtain
the following simplified expansion:
2 2
L = UDV T x − b = DV T x − U T b
.
UTb = z ! p
!m (A1.24)
⇒ L = Dy − z = (s y
i i − z i )2
+ zi2
VTx = y i=1 i=p+1
zi
yi = , x = V y ⇒ x = VD+ U T b = A+ b (A1.25)
si
Equation (A1.25) then provides a formal solution to linear least squares prob-
lems, even in the case when the inverse of ATA in the normal equations does
not exist. The generalized or pseudo-inverse of [A] is then defined as shown
in equation (A1.25). When the singular values of [A] have a clear cut-off the
above formulation is clear, but more often the situation is that the singular val-
ues fall off slowly and never actually go to zero. In this case we face another
challenge. Here the matrix is technically of full rank, but practically we can
obtain an ill-conditioned system. Again we can use the SVD to quantify this
via the concept of condition number of the matrix [A] or CN(A) as follows.
We are interested in how any fractional errors in the input vector b are mapped
into errors in the solution vector x. To obtain this we first define the norm of a
matrixA, which is a scalar defined in equation (A1.27):
Ax ? ? ? ?
A = max ⇒ ?Ax? ≤ A ?x? (A1.27)
x =0 x
This norm has one very important property: the norm of a product is less than
or equal to the product of the norms. We can find an expression for the norm
of [A] by using the SVD and first generating the squared norm, as shown in
equation (A1.28):
xT AT Ax
A2 = (A1.28)
xT x
Now the matrix ATA is always symmetric, and so this Raleigh quotient is
maximized by the maximum eigenvalue of ATA (see equation (A1.19)), λmax .
However, ATA = VD2 VT , and so this eigenvalue is equal to the squared modulus
of the maximum singular value of [A]. Hence we finally have an expression for
the norm of matrix [A] as shown in equation (A1.29):
Here we see that the amplification factor depends on the ratio of maximum
to minimum singular values of [A]. It is easy to show that if [A] is unitary,
for example, then CN = 1, and the errors are not amplified at all. In general,
however, we find that CN > 1, and some care must be taken to ensure that
errors are not amplified too much. Note from equation (A1.30) that we have a
way of controlling this amplification by just setting any small singular values
to zero in the pseudo-inverse, so increasing smin and reducing the CN until
an acceptable ratio is achieved. The price to pay for this, however, is a rank
reduction of [A].
Unitary and rotation
groups A2
In Appendix 1 we defined some very basic properties of matrices and matrix
algebra as used in a description of polarised waves. Here we extend the discus-
sion to examine several more advanced ideas, largely based on group theory,
which are widely used in a description of the propagation and scattering of
polarised electromagnetic waves, forming the subject of polarisation algebra.
There are three classical routes by which the transition from scalar to higher
dimensional forms can be achieved mathematically (Murnaghan, 1962; Misner,
1973; Goldstein, 1980; Cornwell, 1984; Penrose, 1984; Rosen, 1995; Georgi,
1999). These are:
• Scalars_Vectors_Tensors
• Scalars_Complex Numbers_Quaternions_Bi-Quaternions
• Scalars_Spinors_Twistors
Each of these has its strengths and weaknesses in terms of ease of formula-
tion, potential for quantitative analysis, and physical insight. However, two
important general themes arise from all approaches: firstly, the relationship
(sometimes conflict, as in relativistic quantum theory) between real and com-
plex formulations of a problem (which in our context relate to the role of phase
in multi-channel systems); and secondly, the unifying role played by group the-
ory, which provides not only a convenient unifying framework, but also aids
physical insight by exposing deep symmetries that can then be exploited to aid
analysis of complicated problems.
Among the many concepts used in abstract algebra, one of the most useful
is that of a group. We can introduce these ideas by identifying a hierarchy of
algebraic concepts as follows, each one building on the properties of the simpler
concepts to its left.
SET ⇒ GROUP ⇒ FIELD ⇒ VECTOR SPACE ⇒ ALGEBRA
A group G is then defined as a set of elements together with a composition
(generalized concept of a product) xy with x,y ∈ G, such that the following four
conditions hold:
a) Closure, xy ∈ G
b) Associative, x(yz) = (xy)z
c) Existence of the identity, xI = Ix = x
d) Existence of the inverse, xx−1 = I
If in addition we have the property shown in e), then the group is called Abelian
and such groups, as we shall see, have a simpler form and generate the basic
building blocks of a general classification theory.
e) Commutative for group multiplication, xy = yx
412 Unitary and rotation groups
[a, b] = ab − ba (A2.1)
r , defined
properties of the algebra are embodied in a set of structure constants cpq
from the commutation by
!n
ap , aq = r
cpq ar (A2.2)
r=1
The connection between Lie algebras and groups is often provided by the matrix
exponential function, so that we can define a general matrix element A as shown
in equation (A2.3):
a2 a3 an
A = exp(a) = I + a + + + ··· + ··· (A2.3)
2! 3! n!
By exponentiation we then arrive at the Pauli spin matrices σi = −2iai . The Table A2.I Commutation
commutation properties of the Pauli matrices can be conveniently represented matrix of SU(2)
as a matrix, the pqth element of which is 1, 0 or –1 according to equation (A2.5),
σ 1 2 3
as shown in Table A.2.I.
Turning now to higher dimensions, the algebra su(3) is likewise constructed 1 0 1 −1
from eight basis matrices, as shown in equation (A2.6). Just as su(2) leads to 2 −1 0 1
3 1 −1 0
the Pauli spin matrices as so-called generators for SU(2), so the corresponding
set for the group SU(3) are the Hermitian Gell–Mann matrices, obtained as
λk = −iak . Note that the scale factor in a8 is used to ensure that for all products
Tr(ai aj ) = −2δij .
Finally, the algebra su(4) can be represented by the set of fifteen matrices
shown in equation (A2.7). The corresponding generators for the group SU(4) are
called Dirac matrices. The pattern is now clearly developed for representation of
higher-dimensional unitary groups, although we see that the number of elements
414 Unitary and rotation groups
0 i 0 0 1 0 i 0 0
a1 = i 0 0 a2 = −1 0 0 a3 = 0 −i 0
0 0 0 0 0 0 0 0 0
0 0 i 0 0 1 0 0 0
a4 = 0 0 0 a5 = 0 0 0 a6 = 0 0 i (A2.6)
i 0 0 −1 0 0 0 i 0
0 0 0 i 0 0
1
a7 = 0 0 1 a8 = √ 0 i 0
0 −1 0 3 0 0 −2i
0 i 0 0 0 0 i 0 0 0 0 i
i 0 0 1 0 −1 0
0 0 0 0 0
a1 = a2 = a3 =
0 0 0 −1 i 0 0 0 0 1 0 0
0 0 1 0 26 0 −1 0 0 24 i 0 0 0 46
0 i 0 0 i 0 0 0 0 0 0 1
i 0 0 0
0 0 0 i 0 0 0 i
a4 = a5 = a6 =
0 0 0 1 0 0 −i 0 0 i 0 0
0 0 −1 0 35 0 0 0 −i 14
−1 0 0 0 16
0 0 −1 0 0 0 i 0 0 0 0 −1
0 i 0 −1 0 0
0 0 0 0 0 i
a7 = a8 = a9 =
1 0 0 0 i 0 0 0 0 i 0 0
0 i 0 0 12 0 1 0 0 15 1 0 0 0 34
i 0 0 0 0 1 0 0 0 0 0 i
0 −i 0 −1 0 0 0
0 0 0 0 1
a10 = a11 = a12 =
0 0 i 0 0 0 0 i 0 −1 0 0
0 0 0 −i 35 0 0 i 0 23 i 0 0 0 13
0 0 1 0 0 −1 0 0 i 0 0 0
0 i 1 0 0 −i 0
0 0 0 0 0
a13 = a14 = a15 =
−1 0 0 0 0 0 0 i 0 0 −i 0
0 i 0 0 45 0 0 i 0 56 0 0 0 i 25
(A2.7)
A2.2 The real Lie algebra L = so(n) and the group SO(n) 415
is an n×n matrix, the jth row of which consists of the structure constants for the
commutation of x with the jth element of L. The structure constants themselves
form an n-dimensional representation of L called the adjoint representation. For
example, for su(2) we have the following 3 × 3 matrix:
0 0 0 0 0 −1 0 1 0
ad (a1 ) = 0 0 1 ad (a2 ) = 0 0 0 ad (a3 ) = −1 0 0
0 −1 0 1 0 0 0 0 0
(A2.14)
The Killing form is then formed from the trace of matrix products, as shown in
equation (A2.15):
−2 0 0
Bsu(2) = 0 −2 0 = −2δpq (A2.15)
0 0 −2
The Killing form may initially seem a rather contrived concept, but it is used
as the basic distinguishing feature of the Lie algebra, and is central to the
general classification scheme. As mentioned earlier, it is possible to classify
these algebras by identifying an important sub-algebra: the Cartan algebra H .
This is Abelian by definition, and hence has very simple structure, being asso-
ciated with a set of commuting or simultaneously diagonal matrices in the
representation. Physically it can be considered a generalization of absolute
phase in coherent signal analysis. When binary (two channel) operations are
applied between signal channels this phase tends to disappear (in interferom-
etry, for example), and we shall see that similar properties hold for the phase
transformations generated by matrices obtained from the Cartan sub-algebra.
The dimension of H is called the rank of L and, importantly, this is always
much smaller than the dimension of L itself. For su(n), for example, there are
n–1 independent diagonal matrices, and hence the rank of su(n) is n–1. Hence
su(2), su(3), and su(4) have rank 1, 2, and 3 respectively, while their dimensions
(from Table A2.II) are 3, 8, and 15.
The algebra L can now be expressed as the direct sum of H , with rank k, and
a remaining root subspace R, so that
L=H ⊕R (A2.16)
Conveniently, the roots that span the space R can themselves always be written
in terms of a subset of k simple roots rs (where k equals the dimension of H ),
A2.3 The Killing form and Cartan matrix 417
so that
!
k
r= αs rs (A2.17)
s=1
where the coefficients αs are generally complex. The classification can then
be generated by constructing an r × r matrix from the simple roots, called the
Cartan matrix A. The jkth element of A is obtained from the simple roots aj and
ak as
2B aj , ak
Ajk = (A2.18)
B aj , aj
where B(..) is the Killing form. Clearly, the diagonal elements of A are equal
to 2, but less obvious is that the off-diagonal elements are limited in value
to 0, −1, −2, or −3. The Cartan matrix can be constructed from knowledge
of the roots and vice versa; that is, we can construct the whole algebra from
knowledge of A. Hence this matrix is the ‘signature’ of the algebra, and allows
us to compactly describe higher dimensional groups.
Two examples will illustrate the method. We start with su(2), which has basis
a1 , a2 , and a3 (equation (A2.5)), and the Cartan sub-algebra, which is one-
dimensional, with h1 = a3 . From equation (A2.5) we then have the following
(quasi-)eigenvalue conditions for the non-zero roots:
There are consequently two roots α1 and −α1 with α(h1 ) = i. The one-
dimensional root subspaces are then defined from complex linear combinations
of a1 and a2 as λ (a1 + ia2 ) and µ (a1 − ia2 ), where λ and µ are arbitrary
complex numbers. The Cartan matrix for su(2) has only one element: A = 2.
Turning to the more complicated case of su(3), from equation (A2.6) the
Killing form is now
B ap , aq = −12δpq p, q = 1, 2, ...8 (A2.20)
There are therefore six roots, α1 , α2 , α3 , −α1 , α2 , and −α3 , which can all
be expressed as linear combinations of two simple roots: r1 = ah1 + bh2 .
The coefficients a and b can be obtained from the commutation properties of
the roots combined with the Killing form to generate a pair of simultaneous
equations of the following form:
where we have used the fact that B(h1 , h1 ) = B(h2 , h2 ) = 12 and B(h1 , h2 ) = 0.
From these we can calculate the following Killing forms and Cartan matrix:
1 1 1 1
B(α1 , α1 ) = B(α1 , α2 ) = − B(α1 , α2 ) = − B(α2 , α2 ) =
3 6 6 3
(A2.24)
2 −1
A=
−1 2
One important point is that we can now reverse this procedure and use the Cartan
matrix to generate the roots and hence the whole algebra. This is facilitated
through use of a geometrical construction called Dynkin diagrams. These will
finally lead us to the complete classification scheme.
form, mean that the set of non-zero vectors in this space is very restricted. In
this way we can construct the full set of rank N spaces by employing purely
geometrical methods.
A Dynkin diagram is constructed for each algebra L by associating a node
with each simple root and connecting nodes corresponding to roots aj and ak by
a number of lines given by Ajk Akj , where Ajk is the jkth element of the Cartan
matrix. Each node is also given a weight ωj = ωαj , αj , where ω is a constant
chosen such that the minimum weight is unity. These diagrams are used to
generate the whole classification scheme by starting with a root space of rank
1 and using an iterative scheme to generate root spaces of higher rank.
This scheme leads to classification of the four infinite sets of algebras, denoted
Ai , Bi , Ci , and Di with Dynkin diagrams shown in Figure A2.1. There are
also five exceptional algebras E6 , E7 , E8 , F4 , and G2 , with irregular Dynkin
diagrams as shown in Figure A2.2. The classical continuous groups associated
with the four infinite sets of Lie algebras through the matrix exponential function
1 2 i–1 i
1 1 1 1
Ai
2 2 2 1
Bi
1 1 1 2
Ci
1
1 1 1
i>2
Di
Fig. A2.1 Dynkin diagrams for the algebras
1 A, B, C and D
1 1 1 1 E6
1 1 1 1
E7
1
1 1 1 1 1 1
E8
F4
2 2 1
The Dynkin diagram and corresponding Cartan matrix are then of the form
shown in equation (A2.28):
1 1 1 2 −1 0
A = −1 2 −1 (A2.28)
1 2 3 0 −1 2
A suitable set of generators for the group SU(4) are given by ηk = −iak , where
ak are defined in equation (A2.7). These fifteen matrices have the following
commutation properties:
!
15
[ηa , ηb ] = 2iεabc ηc (A2.29)
c=1
η 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
1 0 −1 1 0 0 −1 1 0 0 −1 1 0 0 −1 1
2 1 0 −1 0 1 0 −1 0 1 0 −1 0 1 0 −1
3 −1 1 0 0 −1 1 0 0 −1 1 0 0 −1 1 0
4 0 0 0 0 0 0 0 1 1 1 1 −1 −1 −1 −1
5 0 −1 1 0 0 −1 1 1 1 0 0 −1 −1 0 0
6 1 0 −1 0 1 0 −1 1 0 1 0 −1 0 −1 0
7 −1 1 0 0 −1 1 0 1 0 0 1 −1 0 0 −1
8 0 0 0 −1 −1 −1 −1 0 0 0 0 1 1 1 1
9 0 −1 1 −1 −1 0 0 0 0 −1 1 1 −1 0 0
10 1 0 −1 −1 0 −1 0 0 1 0 −1 1 0 1 0
11 −1 1 0 −1 0 0 −1 0 −1 1 0 1 0 0 1
12 0 0 0 1 1 1 1 −1 −1 −1 −1 0 0 0 0
13 0 −1 1 1 1 0 0 −1 1 0 0 0 0 −1 1
14 1 0 −1 1 0 1 0 −1 0 −1 0 0 1 0 −1
15 −1 1 0 1 0 0 1 −1 0 0 −1 0 −1 1 0
422 Unitary and rotation groups
t1 = e1 ∧ e2 t2 = e2 ∧ e3 t3 = e3 ∧ e1
(A2.30)
t4 = e3 ∧ e4 t5 = e1 ∧ e4 t6 = e2 ∧ e4
Hence when [A] ∈ SU(4) it has unit determinant and therefore when all four
terms in the volume element are distinct their coefficient is 1, whereas the
coefficient is 0 for all combinations with repeated indices.
This result is important, because we can now consider generation of a scalar,
the inner product of two bivectors x and y, as shown in equation (A2.33):
5
x = 5 xi ti
⇒ x ∧ y = f (x, y)e1 ∧ e2 ∧ e3 ∧ e4
y = yj tj (A2.33)
f (x, y) = x1 y4 + x2 y5 + x3 y6 + x4 y1 + x5 y2 + x6 y3
A2.4 Dynkin diagrams: classification of unitary and rotation groups 423
This can be shown quite easily by explicit expansion using the basis vectors
defined in equation (A2.30). Note that there is a cyclic permutation of ordering
in y, and so to obtain a scalar product in C6 we consider not matrix products such
as WT W but must also include a permutation matrix P, as shown in equation
(A2.34), where I 3 is the 3 × 3 identity matrix:
0 I3
[P] = ⇒ [W ]T [P][W ] (A2.34)
I3 0
Since we are considering SU(4), the matrix product will then have 1 in positions
14,25 and so on, where four distinct basis vectors occur and 0 elsewhere. In
other words, the following matrix identity holds:
This result introduces the permutation matrix P, which has the useful property
that P2 = I 6 and is central to the next and final stage of our mapping procedure.
Recall that so far we can go from a 4 × 4 unitary matrix A to a 6 × 6 matrix W,
which is also unitary. (To prove this, use the fact that A maps to W, A∗T maps to
W∗T , and then finally, I 4 maps to I 6 , to show that W∗T W = I 6 .) However, we
seek a mapping from SU(4) to SO(6), which we know exists from the associated
Dynkin diagrams. We now therefore seek a similarity transformation to convert
W into a real orthogonal matrix; that is, we seek a 6 × 6 matrix Q such that
R = QWQ−1 is real orthogonal. We can construct such a matrix using P, as
follows.
We start by defining a symmetric matrix Q = QT , as shown in equation
(A2.36):
The reason for this choice becomes clear when we consider testing whether R
is orthogonal, as shown in equation (A2.37). We see that by using the properties
in equation (A2.36) we guarantee that R is orthogonal, as required.
T
[R]T [R] = QWQ−1 QWQ−1 = Q3 W T PWQ3 = Q3 PQ3 = I6 (A2.37)
Technically, R could still be complex orthogonal, but we can show that it is real
orthogonal by demonstrating that it is unitary as well as orthogonal; that is, that
R∗T R = I 6 . This follows from a similar expansion to that shown in equation
(A2.37). Hence R is real orthogonal as required.
The matrix Q can then be defined explicitly, as shown in equation (A2.38):
(1 − i) I3 iI3 I iI3 I iI3 0 I3
Q= ⇒ Q2 = −i 3 · 3 = =P
2 iI3 I3 iI3 I3 iI3 I3 I3 0
(A2.38)
For example, consider the element of U 4 given by a simple phase shift between
elements in C4, as shown on the left of equation (A2.41). This maps into an
‘equivalent’ 6 × 6 real orthogonal matrix O6 , as shown. The generator for this
U 4 matrix is –ia15 (see equation (A2.7)), and it maps into a rotation in the 2,5
plane. This is shown in equation (A2.7) as a subscript of the form [. . . ]25 .
Equation (A2.42) shows a second example: mapping a (real) 1,6 plane rota-
tion into a (complex) unitary matrix with generator –ia6 . This procedure can
be extended to all fifteen of the generators of SU(4). The plane rotations cor-
responding to each of the fifteen elements of su(4) are shown as subscripts in
equation (A2.7).
iφ 1 0 0 0 0 0
e 0 0 0 0 cos 2φ 0 0 − sin 2φ 0
0 e−iφ 0 0 0 0 1 0 0 0
U4 = ⇒ O6 =
0 0 e−iφ 0 0 0 0 1 0 0
0 0 0 eiφ 0 sin 2φ 0 0 cos 2φ 0
0 0 0 0 0 1
(A2.41)
cos φ 0 0 0 0 − sin φ
0 1 0 0 0 0
0 0 1 0 0 0
O6 = 0
0 0 1 0 0
0 0 0 0 1 0
sin φ 0 0 0 0 cos φ
cos φ2 0 0 sin φ2
0 cos φ2 i sin φ2 0
⇒ U4 = (A2.42)
0 i sin φ2 cos φ2 0
φ
− sin 2 0 0 cos φ2
Coherent stochastic
signal analysis A3
In any discussion of noise processes, prime consideration is usually given to
Gaussian random variables. Their importance stems from the Central Limit
Theorem, which briefly states that the sum of a large number of independent
and identically distributed random variables will be normally distributed. Due
to the generic nature of this theorem, Gaussian statistics are often encoun-
tered in random wave propagation and scattering problems. (Some aspects of
non-Gaussian statistics have been explored for polarised waves; see, for exam-
ple, Bates (1998).) In this Appendix we briefly summarize the main impact of
such stochastic models on coherent signal analysis, particularly on phase and
coherence statistics (Touzi, 1999; Lee, 1994b, 2008; Lopez-Martinez, 2005;
Ferro-Famil, 2008).
Signals generated by Gaussian random processes are characterized by sta-
tistical and not deterministic measures. Consequently, any single sample of
such a process essentially contains zero information, and it is only by obtaining
multiple samples and forming sums or integrals that the signal can be charac-
terized by its statistical moments. The value of x for any particular sample is
then independent of any previous values, and is taken from a normal distribu-
tion such that it is characterized by a probability density function (pdf) p(x), as
summarized in equation (A3.1):
∞
x = E(x) = x.p (x) dx = m
1 (x−m)2
−∞
p (x) = G(m, σ ) = √ e− 2σ 2 ⇒ ∞
σ 2π
x =
2 x2 .p (x) dx = σ 2 + m2
−∞
σx2 = E((x − m)2 ) = σ 2
(A3.1)
For Gaussian signals the process is fully characterized by the mean m and
standard deviation σ . Pure noise signals are often characterized by a zero mean
process, written in shorthand as G(0,σ ). Hence Gaussian noise has only one
free parameter, σ . Figure A3.1 shows an example of a signal composed of
256 samples taken from a G(0,1) random number generator. Such generators
provide the basis for modelling multi-channel polarimetric and interferometric
signals, and are a useful way to investigate the statistical properties of wave
depolarisation, as we show later in this Appendix. First, however, we need to
extend the simple scalar model of equation (A3.1) to complex signals, and then
to the multidimensional complex signal vectors encountered in polarimetry.
426 Coherent stochastic signal analysis
2.5
1.5
0.5
Signal level
0
–0.5
–1
–1.5
–2
–2.5
50 100 150 200 250
Fig. A3.1 256 noise signal samples gener-
ated from a G(0,1) process Sample number
For complex signals, with amplitude and phase, as encountered in wave prop-
agation and scattering, we must extend these ideas to account for both real and
imaginary components of the signal. This is summarized in equation (A3.2).
One of the most important relations in equation (A3.2) is that the expectation of
the product between real and imaginary parts is zero. This is just a consequence
of the fact that noise carries zero phase information, and hence the real and
imaginary parts are independent Gaussian random variables. The phase, there-
fore, has a uniform distribution in the range 0 to 2π , while the intensity, defined
as shown in equation (A3.2), has an exponential distribution. The amplitude
A or square root of the intensity has a Rayleigh distribution as shown. Both
have large variances, and so some care is required to minimize errors when
estimating parameters from real data.
σ
n = nI + inQ ∈ GC (0, σ ) ⇒ nI ,Q = G 0,
2
E(nI ) = E(n Q) = 0
E nI nQ = 0
⇒
σ2
E(n2 ) = E(n22 ) = ⇒ E n2 = σ 2
I
2
nI ⇓
/
1 − I2 E(I) = σ 2
p(I = nI + nQ ) = 2 e σ ⇒
2 2
nR σ var(I ) = σ4
√
π
√ 2A −A22 E(A) = σ
p(A = I ) = 2 e σ ⇒ 2
var(I ) = (4 − π )σ
2
σ
4
(A3.2)
Coherent stochastic signal analysis 427
1 standard deviation
CV = = (A3.3)
ENL mean
/
LL I L−1 −LI E(I ) = σ 2
p(I ) = e σ2 I ≥0⇒ (A3.4)
var(I ) = σL
4
(L − 1)!σ 2L
√
The ratio of the standard deviation to the mean is then reduced to 1/ L. This
observation forms the basis for the design of speckle filters in SAR imaging.
For example, the Lee filter (Lee, 1994a) proposes that we estimate the intensity
of a pixel Î using a mixture of the pixel value itself, I , and its local mean, I
(usually estimated locally using a small M × N window centred on the pixel),
based on a local Taylor expansion for the intensity of a pixel of the form shown
in equation (A3.5):
CV 2 − 1
I = I + k(I − I ) ⇒ k = L
(A3.5)
CV 2
If the area is homogeneous (called fully developed speckle in the SAR context)
then from equation (A3.5), k = 0, and the mean value is taken. However, if
the region is very heterogeneous (a point target, for example) then CV will be
much greater than 1, and so k = 1, and the local value is kept. In this way,
local statistics can be used to strike a balance between spatial and radiometric
resolution. Note that application of this approach to image data relies on two
key assumptions: ergodicity in the mean, and local wide-sense stationarity—
that space and time averages converge to the same mean value so that the spatial
averaging locally around a SAR pixel can be considered equivalent to obtaining
multiple samples of the same random process.
We have seen in Chapter 1 that polarimetry involves a two-dimensional com-
plex space C2. Hence we need to take one further step in our characterization of
noise by considering the statistical properties of signals in C2: pairs of complex
428 Coherent stochastic signal analysis
We have seen, for example, that the product s1 · s2 * arises in many applica-
tions. This product is important, because the conjugate sign implies the phase
difference between signals 1 and 2. Hence, while s1 and s2 individually have
random phase, the phase difference can still be deterministic. Noise processes
in C2 must therefore be further specified by the following added constraints:
E s1 s1∗ = σ 2 E(s1 s2∗ ) = 0 E s2 s2∗ = σ 2 (A3.7)
where again the zero expectation of the cross terms forces the phase difference
to be uniformly random. We can summarize these properties of noise signals in
C2 by generating a 2 × 2 covariance matrix from the expectation of the outer
product of a vector in C2 with its conjugate, as shown in equation (A3.8):
s1 ∗
[C]noise =E . s1 s2
s2
2
E(s1 s1∗ ) E(s1 s2∗ ) σ 0 2 1 0
= = = σ (A3.8)
E(s2 s1∗ ) E(s2 s2∗ ) 0 σ2 0 1
To generalize the above discussion we must consider signals where the cross
expectation is not zero. In this regard, one of the most useful relationships is
the Schwarz inequality, which can be formulated as shown in equation (A3.9):
2
b b b
s1 (x) .s2 (x) dx ≤
∗
|s1 (x)| dx
2
|s2 (x)|2 dx (A3.9)
a a a
This ratio of integrals is called the coherence γ between signals s1 and s2 . From
equation (A3.8) we see that the coherence is always zero for noise signals, while
for polarised EM waves it follows that we can always write Ex = kEy for some
complex constant k, and the coherence of polarised waves is always unity.
Furthermore, as the mean phase of s1 s2 * may not be zero, it is convenient to
define the complex coherence as shown in equation (A3.11):
+ ∗,
s1 s2
γ̃ = γ e iφ
= + ,+ , (A3.11)
s1 s1∗ s2 s2∗
Coherent stochastic signal analysis 429
90 1
120 60
0.8
0.6
150 P
30
0.4
0.2
180 0
210 330
240 300
Fig. A3.2 Unit circle in the complex coher-
270 ence plane
This has a magnitude between 0 and 1 and a phase from 0 to 2π, hence we
can represent the coherence as a point P inside the unit circle of the complex
coherence plane as shown in FigureA3.2. Noise sits at the origin of this diagram,
while coherent signals lie around the outer unit circle.
Coherence is a ratio of random variables and hence is a stochastic quantity,
so attention must be paid to its statistics. In the most general case where corre-
lation is allowed between the signals s1 and s2 , the probability density function
becomes a multivariate Gaussian of the form shown in equation (A3.12):
s 1 ∗T −1
u = 1 ⇒ p(u) = 2 e−u [C] u (A3.12)
s2 π det([C])
Note that if s1 = ks2 then det([C]) = 0, and the density function must be
replaced by a delta function at u = u0 . The single-look density function for the
phase of s1 s2 * now has the following form:
0 1
1 − γ2 1 − ψ 2 + ψ π − cos−1 ψ
P(φ) = 1.5
2π 1 − ψ 2
ψ = γ cos φ −π ≤φ ≤π (A3.14)
averaging can be used to reduce the variance of the estimates for any given γ .
We can therefore define the maximum likelihood estimate of [C], denoted [Z],
as shown in equation (A3.15):
1!
L
[Z] = uk u∗T
k (A3.15)
L
j=1
The matrix [Z] is then itself a random matrix with a probability distribution;
the complex Wishart distribution—a function of the number of samples L and
of the general form shown in equation (A3.16) (Lee, 1994b, 2008; Conradsen,
2003):
Here q is the dimension of the complex vector u (2 in this case, but 3 for
monostatic polarimetry, 4 for bistatic, and 6 for single baseline polarimetric
interferometry and so on). Equation (A3.4) is a special case of this distribution
for q = 1. This distribution then leads to the following pdf for the phase as a
function of the number of looks L:
L + 12 (1 − γ 2 )L ψ (1 − γ 2 )L 1 2
P(φ) = √ L+ 1 + F L, 1 : ; ψ
2 π (L) 1 − ψ 2 2 2π 2
ψ = γ cos (φ − φm ) (A3.17)
5
L
∗
s1i s2i /
i=1 0 ≤ |γ̃ | ≤ 1
γ̃ = " " (A3.18)
5
L 5
L 0 ≤ arg(γ̃ ) = φ̂m < 2π
∗
s1i s1i ∗
s2i s2i
i=1 i=1
The pdf of the sample coherence magnitude g = |γ̃ | for jointly Gaussian signals
can be derived analytically, and is a function of the coherence magnitude γ , the
number of integrated independent samples, L, and the hypergeometric function
F, as shown in equation (A3.19):
This shows a bias towards higher coherence values, especially for low coherence
with a small number of samples L. The variance of the estimate can also be
derived from equation (A3.22) using equations (A3.20) (k = 2) and (A3.21).
var(g) = E g 2 − E(g)2 (A3.22)
Useful as these expressions are, they are difficult to interpret without detailed
calculations. For this reason the Cramer–Rao (CR) lower bounds on variance
of phase and coherence have also been derived (see Seymour (1994) and yield
simpler expressions, as shown in equation (A3.23):
2
1 − γ2 1 − γ2
var(φ) > var(g) > (A3.23)
2Lγ 2 2L
Figure A3.3 shows examples of these CR bounds as a function of coherence
and increasing number of looks. Generally, the higher the coherence, the lower
the number of looks required to obtain a specified variance.
Our representation of coherence inside the unit circle of Figure A3.2 is there-
fore rather misleading. In fact, each point P has a minimum cloud of uncertainty
around it representing the Cramer–Rao bounds in radius (coherence) and polar
angle (phase) fluctuations. Note that this cloud will be elliptical in shape. For a
given number of looks L the phase variance is larger than the radial coherence
variance. Figure A3.4 shows a schematic representation of this concept. This
result underpins our distinction in Chapters 7 and 8 between coherence loci and
associated coherence regions.
L=4
80 L = 16
0.12
L = 64
70
0.1
60
Variance (degrees)
50 0.08
Variance
40 0.06
30
0.04
20
0.02
10
90 1
120 60
0.8
0.6
150 30
0.4
0.2
180 0
210 330
240 300
Fig. A3.4 Complex coherence as a stochastic
variable inside the unit circle 270
equation (A3.26):
e1 0 0
.. 2 3
u = [UMN ] [E] = 0 . 0 ei = λi Ga 0, 12 + iGb (0, 12 )
0 0 eMN
0 1 !L
L→∞
ˆ MN = uu∗T −−−−→ [MN ] (A3.26)
i=1
From this we can define a non-negative distance measure such that we assign
k to the class with the shortest distance d , defined from equation (A3.28) by
434 Coherent stochastic signal analysis
ignoring elements that do not depend on the class, as shown in equation (A3.29):
d k, Ci = ln |Ci | + Tr(Ci−1 kk ∗T ) (A3.29)
Note that since the Mueller matrix [M ] can be mapped 1–1 with the scattering
coherency matrix [T ], which itself is unitarily similar to [C], the metric in
equation (A3.30) is invariant to use of [C] or [T ]. In this way we can also
provide a statistical distance measure between experimental Mueller matrices.
Bibliography
Schou, J., Skriver, H., Nielsen, A. A. and Conradsen, K. (2003). CFAR edge
detector for polarimetric SAR images. IEEE Transactions on Geoscience
and Remote Sensing, 41 (1), 20–32.
Schuler, D. L., Lee, J. S., Kasilingam, D. and Nesti, G. (2002). Sur-
face roughness and slope measurements using polarimetric SAR data.
IEEE Transactions on Geoscience and Remote Sensing, 40 (3),
687–698.
Seymour, S. and Cumming, I. G. (1994). Maximum likelihood estimation
for SAR interferometry. Proceedings of IEEE Geoscience and Remote
Sensing Symposium, IGARSS’94, Pasadena, USA.
Sharma, J. J., Hajnsek, I. and Papathanassiou, K. P. (2007). Vertical profile
reconstruction with POLInSAR of a subpolar glacier. Proceedings of
IEEE Geoscience and Remote Sensing Symposium, IGARSS’07, 1147–
1150.
Simon, R. (1987). Mueller matrices and depolarization criteria. Journal of
Modern Optics, 34, 569–575.
Souyris, J. C., Imbo, P., Fjortoft, R., Mingot, S. and Lee, J. S. (2005). Compact
polarimetry based on symmetry properties of geophysical media: the
pi/4 mode. IEEE Transactions on Geoscience and Remote Sensing, 43
(3), 634–646.
Stebler, O., Meier, E. and Nueesch, D. (2002). Multi-baseline polarimetric
SAR interferometry: first experimental spaceborne and airborne results.
ISPRS Journal of Photogrammetry and Remote Sensing, 56 (3).
Stokes, G. G. (1852). On the composition and resolution of streams of polar-
ized light from different sources. Cambridge Philosophical Society, 9,
399.
Strang, G. (2004). Linear Algebra and its Applications. Fourth edition. Brooks
Cole.
Tabb, M. and Carande, R. (2001). Robust inversion of vegetation struc-
ture parameters from low frequency polarimetric interferometric SAR.
Proceedings of IEEE International Geoscience and Remote Sensing
Symposium (IGARSS 2001), Sydney, Australia, July 2001.
Tabb, M., Orrey, J., Flynn, T. and Carande, R. (2002a). Phase diversity:
a decomposition for vegetation parameter estimation using polarimet-
ric SAR interferometry. Proceedings of the Fourth European Synthetic
Aperture Radar Conference, EUSAR 2002, 721–724.
Tabb, M., Flynn, T. and Carande, R. (002b). Direct estimation of vegetation
parameters from covariance data in POLINSAR. Proceedings of IGARSS
2002, Toronto, Canada, 1908–1910.
Topp, G. C., Davis, J. L. andAnnan, A. P. (1980). Electromagnetic determination
of soil water content: measurements in coaxial transmission lines. Water
Resources Research, 16, 574–582.
Touzi, R., Lopes, A., Bruniquel, J. and Vachon, P. W. (1999). Coherence esti-
mation for SAR imagery. IEEE Transactions on Geoscience and Remote
Sensing, 37/1, 135–149.
Touzi, R. (2007). Target scattering decomposition in terms of roll-invariant tar-
get parameters. IEEE Transactions on Geoscience and Remote Sensing,
45 (1), 73–84.
448 Bibliography
along track interferometry (ATI), 219 for semi-infinite random volume, 258 contrast optimization, 193
ALOS-PALSAR, 396 SNR decorrelation, 221 COPOL nulls, 57
alpha parameter geometric decorrelation, 224 Cramer–Rao bounds, 431
and dihedral dielectric constant, 399 image co-registration errors, 223 critical baseline, 215
definition, 184 interferometric, 220 cross-polarisation, 8
for Bragg scattering, 128 polarimetric, 73
for dihedral scattering, 130 polarimetric interferometric, 235
mean alpha, 98 temporal decorrelation, 222
decomposition
anisotropy scattering parameter A, 97 coherence loci
eigenvalue decomposition of [J], 74
definition, 252 eigenvalue decomposition of [T], 87
for IWCM model, 280
of interferometric coherence, 233
baseline components, 213 for oriented volume scattering, 263 of N-dimensional coherency matrices, 92
for OVOG model, 282
baseline decorrelation, 224 of the Stokes vector, 78
for OVUG model, 283
biaxial material, 13 of the wave coherency matrix, 76
bidirectional reflectance distribution function for random volume scattering, 257 propagation distortions, 205
for RVOG, 274
(BRDF), 139 roll invariance, 179
for RVUG model, 283
birefringence the point reduction theorem, 185
for surface scattering, 255
for circular polarisations, 24 degree of polarisation, 76
coherence optimization
general definition, 14 depolarisation
bivariate Gaussian distribution, 429 estimation bias, 250 definition, 71
Bragg surface scattering, 126 constrained, 243 state vector, 92
for oriented volume scattering, 261
Brewster angle, 122 depolarisers
for random volume scattering, 255
azimuthal depolarisation, 96
for surface scattering, 254
characterisation of, 110
SVD interpretation, 242 isotropic, 91
C2 symmetry, 11 unconstrained, 241 reflection depolarisation, 97
calibration coherence region, 245
of POLInSAR systems, 361 dielectric constant
coherence tomography
of soil, 120
of POLSAR systems, 351 condition number for multi-baselines, 338 Polder–Van Santen/de Loor formula, 171
polarimetric, 350 dual baseline, 325
soil moisture, 121
Quegan algorithm, 352 multi-baseline reconstruction, 335
soil salinity, 120
Cameron decomposition, 182 single baseline, 323
Cartan matrix, 417 dielectric tensor, 12
single baseline condition number, 332
differential interferometry, 217
Cartan sub-algebra, 416 temporal decorrelation and SNR dihedral scattering
application to coherency matrix, 92 effects, 334 from dielectrics, 124
Chandrasekhar decomposition, 191 coherency matrix from metal structures, 54
chiral media and CFAR detection, 194
Dirac matrices, 413
admittance parameters, 23 and contrast optimization, 193
directional hemispherical reflectivity
D and L-rotatory materials, 24 for forward scattering, 165
spatial dispersion, 23 (DHR), 139
for general scattering, 86
specific rotatory power of, 25 discrete dipole approximation (DDA), 152
for waves, 73
wave equation for, 23 DLR E-SAR, 392
propagation effects, 202
Dynkin diagrams, 418
Cloude–Pottier decomposition, 192 compact polarimetry, 354
coefficient of variation (CV), 427 and Bragg scattering, 176
coherence and Rayleigh scattering by spheroids, 176
for two-layer surface/volume problems, 270 π/4 mode, 356 effective length of an antenna, 51
for circular polarisation, 136 compact POLInSAR, 362 effective number of looks (ENL), 427
for exponential profile, 231 condition number of a matrix, 409 eigenpropagation states, 14
452 Index