Slides - Graph Signal Processing: Fundamentals and Applications To Diffusion Processes
Slides - Graph Signal Processing: Fundamentals and Applications To Diffusion Processes
EUSIPCO16
Budapest, Hungary - August 29, 2016
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 1 / 118
Network Science analytics
PC AS FR RA
OG IC
CO MP SC MP
I Oil extraction (OG), Petroleum and coal products (PC), Construction (CO)
I Administrative services (AS), Professional services (MP)
I Credit intermediation (FR), Securities (SC), Real state (RA), Insurance (IC)
x2 x4
2 4
x1
1
3 5
x3 x5
e j2πkn/N
H
I Cyclic structure ⇒ Fourier transform ⇒ x̃ = F x Fkn = √
N
I Fourier transform ⇒ Projection on eigenvector space of cycle
I Random signal with mean E [x] = 0 and covariance Cx = E xxH
⇒ Eigenvector decomposition Cx = VΛVH
x1
σ16 1 σ12
I Covariance matrix Cx is a graph x6 x2
6 2
⇒ Not a very good graph, but still σ15 σ13
I V = {1, 2, 3, . . . , 9}
4 5 6
I E = {(1, 2), (2, 3), . . . , (8, 9), (1, 4), . . . , (6, 9)}
I W : (n, m) 7→ 1, for all (n, m) ∈ E
7 8 9
Σ11 Σ22
Σ12
p1 p2
I Weighted and undirected graphs (e.g., covariance)
Σ14 I V = {1, 2, 3, 4}
Σ13 Σ23 Σ24
I E = {(1, 1), (1, 2), . . . , (4, 4)} = V × V
p3 p4
I W : (n, m) 7→ σnm = σmn , for all (n, m)
Σ34
Σ33 Σ44
L=D−A
I L is positive semi-definite
⇒ xT Lx = 12 (i,j)∈E wij (xi − xj )2 ≥ 0, for all x
P
Part I: Fundamentals
Graph signals and the shift operator
Graph Fourier Transform (GFT)
Graph filters and network processes
Concluding remarks
5
x0 0.6
x1 0.7
0
x = x2 = 0.3
6 9
1 4
.. ..
2 3 . .
7 8 x9 0.7
I What if y = S2 x?
⇒ Like powers of
A: neighborhoods
⇒ yi found using
values within 2-hops
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 20 / 118
Graph Fourier Transform
Part I: Fundamentals
Graph signals and the shift operator
Graph Fourier Transform (GFT)
Graph filters and network processes
Concluding remarks
2π
I Let x be a temporal signal, its DFT is x̃ = FH x, with Fkn = √1 e +j N kn
N
⇒ Equivalent description, provides insights
⇒ Oftentimes, more parsimonious (bandlimited)
⇒ Facilitates the design of SP algorithms: e.g., filters
x̃ = V−1 x
x = Vx̃
I Additional structure
⇒ If S is normal, then V−1 = VH and x̃k = vH
k x =< vk , x >
⇒ Parseval holds, kxk2 = kx̃k2
P P P
T I T I T I
C P I C P I C P I
K T K T K T
J J J
500
1000
Genes
1500
2000
0.9 0.9
0.8 0.8
0.7 0.7
0.6 0.6
Amplitude
Signal
0.5 0.5
0.4 0.4
0.3 0.3
0.2 0.2
0.1 0.1
0 0
0 500 1000 1500 2000 2500 0 500 1000 1500 2000 2500
Gene Frequencies
0.01
0.009
0.008
Distinguishing Power
0.007
0.006
0.005
0.004
0.003
0.002
0.001
0
0 500 1000 1500 2000 2500
Frequency
15
error percentage
10
original
5
one frequency
p = 60
p = 75
p = 90
0
k=3 k=5 k=7
Part I: Fundamentals
Graph signals and the shift operator
Graph Fourier Transform (GFT)
Graph filters and network processes
Concluding remarks
H := Vdiag h̃ V−1 ,
h̃k = g (λk )
PL
I DEF1 says H = l=0 hl Sl
I Suppose H acts on a graph signal x to generate y = Hx
⇒ If we define x(l) := Sl x = Sx(l−1)
y is a linear
L
X combination of
y= hl x(l)
successive shifted
l=0
versions of x
I After introducing S, we stressed that y = Sx can be computed locally
⇒ x(l) can be found locally if x(l−1) is known
⇒ The output of the filter can be found in L local steps
h̃ = Ψh
1 1 ... 1
2π(1)(1) 2π(1)(N−1)
−j −j
1 e N ... e N
= FH
Ψ= .. .. ..
. . .
2π(N−1)(1) 2π(N−1)(N−1)
−j −j
1 e N ... e N
h̃ = FH h
ỹ = diag(FH h)FH x
ỹ = diag(Ψh)V−1 x
I The GFT for filters is different from the GFT for signals
⇒ Symmetry is lost, but both depend on spectrum of S
⇒ Many of the properties are not true for graphs
⇒ Several options to generalize operations
I Frequency or space?
PL
y = Vdiag(h̃)V−1 x vs. y= l
l=0 hl S x
region
temporal smo
⇒ Based on correlated activity
then cluste
I fMRI outputs a series of graph signals
time
⇒ x(t) ∈ R112 describing brain states
# entries in higher than a thres
I Does brain state variability correlate with learning? sum ( elementwise α-power in )
# changes in
# entries in higher than a thres
sum ( elementwise α-power in )
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications approximate entropy ( 47) / 118
Measuring brain state variability
I We propose three different measures capturing different time scales
⇒ Changes in micro, meso, and macro scales
I Micro: instantaneous changes higher than a threshold α
T
X kx(t) − x(t − 1)k2
m1 (x) = 1 >α
t=1
kx(t)k2
I Meso: Cluster brain states and count the changes in clusters
T
X
m2 (x) = 1 {c(t) 6= c(t − 1)}
t=1
⇒ where c(t) is the cluster to which x(t) belongs.
I Macro: Sample entropy. Measure of complexity of time series
P P !
t s6=t 1{kx̄3 (t) − x̄3 (s)k∞ > α}
m3 (x) = − log P P
t s6=t 1{kx̄2 (t) − x̄2 (s)k∞ > α}
Response
I High freq. components are attenuated
I β controls the level of attenuation Frequency
8 8 8
7 7 7
Learning Rate
Learning Rate
Learning Rate
6 6 6
5 5 5
4 4 4
3 3 3
2 2 2
1 1 1
0.52 0.53 0.54 0.55 0.56 0.57 0.58 0.59 0.05 0.1 0.15 0.2 0.25 0.3 0.03 0.04 0.05 0.06 0.07 0.08 0.09
Variability Variability Variability
Part I: Fundamentals
Graph signals and the shift operator
Graph Fourier Transform (GFT)
Graph filters and network processes
Concluding remarks
Part I: Fundamentals
Graph signals and the shift operator
Graph Fourier Transform (GFT)
Graph filters and network processes
Concluding remarks
I We can fix a point (present) and sample the evolution of the signal
I Both strategies coincide for time signals but not for general graphs
⇒ Give rise to selection and aggregation sampling
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 57 / 118
Selection sampling: Definition
I Number of samples P ≥ K
x̄ = Cx = CVK x̃K
⇒ (CVK ) submatrix of V
x = VK x̃K = VK (CVK )† x̄
I If C = ET
K , node i can recover x with info from K − 1 hops!
⇒ Node i has to be able to capture frequencies in K
⇒ The frequencies have to distinguishable
I Bandlimited signals: Signals that can be well estimated locally
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 61 / 118
Aggregation and selection sampling: Example
I Erdős-Rényi
p = 0.2, S = A,
K = 3,
non-smooth
I Three extensions:
⇒ Sampling in the presence of noise
⇒ Unknown frequency support
⇒ Space-shift sampling (hybrid)
−1
⇒ If P = K , then x̂(i) = VK (CΨi ) z̄i
(i) (i)
I Error covariances (Re ,R̃e ) in closed form ⇒ Noise covariances?
⇒ Colored, different models: white noise in zi , in x, or in x̃K
I Metric to optimize?
i−1
(i)−1
h
(i) (i) (i)
⇒ trace(Re ), λmax (Re ), log det(R̃e ), trace R̃e
I Erdős-Rényi
p = 0.15, 0.20, 0.25,
K = 3, non-smooth
I Beyond bandlimitedness
⇒ Smooth signals [Chen15]
⇒ Parsimonious in kernelized domain [Romero16]
I And more...
⇒ Low-complexity implementations [Tremblay16, Anis16]
⇒ Local implementations [Wang14, Segarra15]
⇒ Unknown spectral decomposition [Anis16]
Part I: Fundamentals
Graph signals and the shift operator
Graph Fourier Transform (GFT)
Graph filters and network processes
Concluding remarks
(2) Signal is the output of a LTI filter H excited with white noise w
Cx = Fdiag(p)FH
S = V Λ VH =⇒ Cx = V diag(p) VH
Proposition
Process x has shift invariant correlation matrix ⇔ it is the output of a linear
shift invariant filter ⇔ Covariance jointly diagonalizable with shift
I Shift and Filtering ⇒ How stationary signals look like (local invariance)
I Simultaneous Diagonalization ⇒ A PSD exists ⇒ p := diag VH Cx V
⇒ The PSD collects the eigenvalues of Cx and is nonnegative
Proposition
Let x be stationary in S and define the process x̃ := VH x. Then, it holds
that x̃ is uncorrelated with covariance matrix Cx̃ = E x̃x̃H = diag(p).
Theorem
If the process x is Gaussian, periodogram estimates have bias and
variance
I Bias ⇒ bpg := E [p̂pg ] − p = 0
2 2
I Variance ⇒ Σpg := E (p̂pg − p)(p̂pg − p)H = R diag (p)
2
Baseline ER
Smaller ER
1 Larger PSD
Small-world
0
1 3 5 7 9 11 13 15 17 19
Number of observations
Normalized MSE
1.5
Normalized MSE
1.5
1
1
0.5
0.5
0
0 1 3 5 7 9 11 13 15 17 19
Single window Ten local wind. Ten random wind. Number of windows
I The use of windows introduces bias but reduces total error (MSE)
I Local windows work better than random windows
⇒ Advantage of local windows is larger for local processes
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 84 / 118
Opinion source identification
0.45
< =0
0.4 < = 0.10
Source identification error
< = 0.15
0.35 < = 0.20
0.3
0.25
0.2
0.15
0.1
0.05
10 15 20 25 30
Number of observations
Normalized MSE
4 2.5
20
25 2 2
30
0
35 1.5
-2
40
1
45 -4
50 0.5
10 20 30 40 50 0 5 10 15 20
Frequency index Number of windows
I Extensions
⇒ Other parametric schemes
⇒ Space-time variation
Part I: Fundamentals
Graph signals and the shift operator
Graph Fourier Transform (GFT)
Graph filters and network processes
Concluding remarks
x = (I − αL)−1 w x = Ax + w
Inferred"
Input"
network"
A"priori"info,"desired"
topological"features"
STEP%1:% STEP%2:%
Inferred" Iden-fy%eigenvalues%to%
Iden-fy%the%eigenvectors% eigenvectors"
of%the%shi9% obtain%a%suitable%shi9%
1) Graph sparsification
I Goal: given Sf find sparser S with same eigenvectors
⇒ Find Sf = Vf Λf VfH and set V = Vf
⇒ Otentimes referred to as network deconvolution problem
2) Nodal relation assumed by a given transform
I GSP: decompose S = VΛVH and set VH as GFT
I SP: some transforms T known to work well on specific data
I Goal: given T, set VH = T and identify S ⇒ intuition on data relation
DCTs: i–iii
I Convex feasibility set ⇒ Search for the optimal solution may be easy
Ex: Identify the sparsest shift S∗0 that explains observed signal structure
⇒ Set the cost f (S, λ) = kSk0
N
X
S∗0 = argmin kSk0 s. to S = λk vk vkT , S ∈ S
S,λ
k=1
†
I Denoting by mT
i the i-th row of M := (I − WW )D c
⇒ Construct R := [m2 −m1 , . . . , mN−1 −m1 , mN , . . . , m|Dc | ]T
⇒ Denote by K the indices of the support of s∗0 = vec(S∗0 )
S∗1 and S∗0 coincide if the two following conditions are satisfied:
1) rank(RK ) = |K|; and
2) There exists a constant δ > 0 such that
ψR := kIKc (δ −2 RRT + IT −1 T
Kc IKc ) IK k∞ < 1.
PN
min kSk1 s. to Ŝ = k=1 λk v̂k v̂k T , S ∈ S, d(S, Ŝ) ≤
{S,λ,Ŝ}
Frequency
0.6
30 30
N
N
30
0.4 40 40 20
0.2 50 50 10
0 0.1 0.3 0.5 0.7 0.9 0.1 0.3 0.5 0.7 0.9 0
2 2.5 5 6 7 8 9
p p Rank
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 100 / 118
Inferring brain graphs from noisy templates
0.15
0.1
0.05
0
10 3 10 4 10 5 10 6
Number of Observations
0.6
0.4
0.2
0
16 18 20 22 24 26 28
Number of spectral templates
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 102 / 118
Performance comparisons
0.9
0.8
F-measure
0.2
10 1 10 2 10 3 10 4 10 5 10 6
Number of observations
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 103 / 118
Inferring direct relations
5 5 5 5
10 10 10 10
15 15 15 15
20 20 20 20
25 25 25 25
30 30 30 30
35 35 35 35
40 40 40 40
45 45 45 45
50 50 50 50
10 20 30 40 50 10 20 30 40 50 10 20 30 40 50 10 20 30 40 50
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 105 / 118
Wrapping up
Part I: Fundamentals
Graph signals and the shift operator
Graph Fourier Transform (GFT)
Graph filters and network processes
Concluding remarks
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 106 / 118
Concluding remarks
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 107 / 118
Concluding remarks
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 108 / 118
Looking ahead
I Thanks!
⇒ If you have questions, feel free to contact me by e-mail
[email protected] or any of the other authors.
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 109 / 118
References
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 110 / 118
References: our work
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 111 / 118
References: our work
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 112 / 118
References: our work
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 113 / 118
References: our work
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 114 / 118
References: GSP
General references
D. Shuman, S. Narang, P. Frossard, A. Ortega, and P. Vandergheynst, The emerging field of signal
processing on graphs: Extending highdimensional data analysis to networks and other irregular
domains, IEEE Signal Process. Mag., vol. 30, no. 3, pp. 83-98, Mar. 2013.
A. Sandryhaila and J. Moura, Discrete signal processing on graphs, IEEE Trans. Signal Process.,
vol. 61, no. 7, pp. 1644-1656, Apr. 2013.
A. Sandryhaila and J. Moura, Discrete signal processing on graphs: Frequency analysis, IEEE
Trans. Signal Process., vol. 62, no. 12, pp. 3042-3054, June 2014.
A. Sandryhaila and J. M. F. Moura, Big Data analysis with signal processing on graphs, IEEE
Signal Process. Mag., vol. 31, no. 5, pp. 80-90, 2014.
M. Rabbat and V. Gripon, Towards a spectral characterization of signals supported on small-world
networks, in IEEE Intl. Conf. Acoust., Speech and Signal Process. (ICASSP), May 2014, pp.
4793-4797.
A. Agaskar, Y.M. Lu, A spectral graph uncertainty principle, IEEE Trans. Info. Theory, vol. 59,
no. 7, pp. 4338-4356, 2013.
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 115 / 118
References: GSP
Filtering
D. I. Shuman, P. Vandergheynst, and P. Frossard, Distributed signal processing via Chebyshev
polynomial approximation, CoRR, vol. abs/1111.5239, 2011.
S. Safavi and U. Khan, Revisiting finite-time distributed algorithms via successive nulling of
eigenvalues, IEEE Signal Process. Lett., vol. 22, no. 1, pp. 54-57, Jan. 2015.
Sampling
A. Anis, A. Gadde, and A. Ortega, Towards a sampling theorem for signals on arbitrary graphs, in
IEEE Intl. Conf. Acoust., Speech and Signal Process. (ICASSP), May 2014, pp. 3864-3868.
I. Shomorony and A. S. Avestimehr, Sampling large data on graphs, arXiv preprint
arXiv:1411.3017, 2014.
S. Chen, R. Varma, A. Sandryhaila, and J. Kovacevic, Discrete signal processing on graphs:
Sampling theory, IEEE. Trans. Signal Process., vol. 63, no. 24, pp. 6510-6523, Dec. 2015.
M. Tsitsvero, Mikhail, S. Barbarossa, and P. Di Lorenzo. Signals on graphs: Uncertainty principle
and sampling, arXiv preprint arXiv:1507.08822, 2015.
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 116 / 118
References: GSP
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 117 / 118
References: GSP
Topology inference
V. Kalofolias, How to learn a graph from smooth signals, arXiv preprint arXiv:1601.02513, 2016.
X. Dong, D. Thanou, P. Frossard, and P. Vandergheynst, Learning laplacian matrix in smooth
graph signal representations, arXiv preprint arXiv:1406.7842, 2014.
B. Pasdeloup, V. Gripon, G. Mercier, D. Pastor, M. Rabbat, Characterization and inference of
weighted graph topologies from observations of diffused signals, arXiv preprint arXiv:1605.02569,
2016.
J. Mei and J. Moura, Signal processing on graphs: Estimating the structure of a graph, in IEEE
Intl. Conf. Acoust., Speech and Signal Process. (ICASSP), 2015, pp. 54955499.
E. Pavez and A. Ortega, Generalized Laplacian precision matrix esti- mation for graph signal
processing, in IEEE Intl. Conf. Acoust., Speech and Signal Process. (ICASSP), Shanghai, China,
Mar. 20-25, 2016.
Stationarity
N. Perraudin and P. Vandergheynst, Stationary signal processing on graphs, arXiv preprint
arXiv:1601.02522, 2016.
B. Girault, Stationary graph signals using an isometric graph translation,” Proc. IEEE European
Signal Process. Conf. (EUSIPCO), 2015.
Marques, Segarra, Ribeiro Graph SP: Fundamentals and Applications 118 / 118