Processing Signals Supported on...

34
Processing Signals Supported on Graphs Michael Rabbat

Transcript of Processing Signals Supported on...

Page 1: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Processing Signals Supported on Graphs

Michael Rabbat

Page 2: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Traditional Signal Processing

2

1-D (e.g., audio) 2-D (e.g., images)

Page 3: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Smoothness

Example: Audio signal

Smooth = (mostly) low frequency

3

Page 4: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Sparsity

Example: 2D Image and its Wavelet Transform

Sparsity = most wavelet coefficients are (nearly) zero (Note: zero = blue)

4

Page 5: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Implications of Smoothness & Sparsity

•  Signal processing tasks –  Signal measurement, acquisition à Estimation –  Signal storage, communication à Compression

•  Approximation Theory When/how can one signal be approximated well by another? –  Other signal is “cleaner” or “simpler” than the other –  Smoothness (focus on low frequency) –  Sparsity (focus on few high-energy coefficients)

5

Page 6: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Signals Supported on Graphs

Many applications: •  Sensor networks •  Smart grid •  Social networks •  Transportation •  Internet monitoring •  Economic networks •  …

6

32 34 36 38 40 42−124

−123

−122

−121

−120

−119

−118

−117

−116

−115

−114

−0.1

−0.05

0

0.05

0.1

Data Source: California Irrigation Management Information System http://www.cimis.water.ca.gov/

Page 7: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Questions

•  When and how can we approximate signals on graphs?

•  What is a “smooth” signal on a graph? •  What is a “Fourier” transform for signals on a graph? •  Which graphs have meaningful “Fourier” transforms? •  Which graphs have interesting smooth signals? •  When and how can smooth signals be helpful?

7

Page 8: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Outline

•  Introduction and motivation •  Approximating signal supported on graphs

–  Classical approximation theory –  Approximation theory for graphs

•  Field estimation in sensor networks

•  Based on joint work with Xiaofan Zhu –  X. Zhu and M. Rabbat, “Approximating signals

supported on graphs,” ICASSP 2012 –  X. Zhu and M. Rabbat, “Graph spectral

compressed sensing,” ICASSP 2012

8

Page 9: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

APPROXIMATING SIGNALS SUPPORTED ON GRAPHS

9

Page 10: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Classical Approximation Theory

Let

Fourier transform

Total variation Proposition:

10

f 2 L2([0, 1])

kfkV =

Z 1

0|f 0(t)|dt

bf(!) =Z 1

0f(t)e�i!tdt

| bf(!)| kfkV|!|

S. Mallat, A Wavelet Tour of Signal Processing, 3rd ed, Academic Press, 2009.

Small TV è energy mainly in low frequencies

Page 11: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Fourier Approximation

Fourier coefficient Fourier expansion

11

hf(u), ei2⇡mui =Z 1

0f(u)e�i2⇡mudu

f(t) =1X

m=�1hf(u), ei2⇡muiei2⇡mt

Page 12: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

M-term Linear Approximation

Only keep M lowest frequency coefficients (Force others to zero) M-term linear approximation:

M-term linear approximation error:

12

fM (t) =X

m:|m|<M/2

hf(u), ei2⇡muiei2⇡mt

✏l(M, f) = kf � fMk2

=X

m:|m|>M/2

|hf(u), ei2⇡mui|2

Page 13: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Approximation Error Scaling

Theorem: If then

13

kfkV < 1 ✏l(M, f) = O

✓kfkVM�1

Theorem: For any , if then .

s > 1/21X

m=0

|m|2s |hf, ei2⇡mui|2 < 1

✏l(M, f) = o(M�2s)

Page 14: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Signals on Graphs?

14

32 34 36 38 40 42−124

−123

−122

−121

−120

−119

−118

−117

−116

−115

−114

−0.1

−0.05

0

0.05

0.1

Page 15: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Quick Intro to Spectral Graph Theory

•  Set representation of a graph •  Adjacency Matrix with entries

•  Degree of node u:

•  Degree matrix is diagonal with entries

15

A

G = (V,E,w)

Au,v =

⇢wu,v if (u, v) 2 E0 otherwise

d(u) =X

v2V

wu,v

D Du,u = d(u)

Page 16: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Smoothness and the Graph Laplacian

•  Signal defined on vertices of where is the value at node

•  The graph Laplacian is

•  Define graph variation so that

16

L = D �A

Gx 2 R|V |

xv v

kxkG

kxk2G = x

TLx

=X

(u,v)2E

wi,j(xu � xv)2

Page 17: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Graph Fourier Transform (GFT)

Consider eigenvalue decomposition of with eigenvalues and corresponding ith eigenvector We’ll call the ith graph Fourier coefficient. Clearly,

17

L

L = U⇤U�1

0 = �1 �2 · · · �n n = |V |

bx(�i) = hx, uii

ui

x =nX

i=1

bx(�i)ui

Page 18: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

GFT Example

18

32 34 36 38 40 42−124

−123

−122

−121

−120

−119

−118

−117

−116

−115

−114

−0.1

−0.05

0

0.05

0.1

0 20 40 60 80 100 120 140−1000

0

1000

2000

3000

4000

5000

6000

x and G

bx(�k) vs k

Page 19: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Many Other Applications Using GFT

•  Machine Learning –  J. Shi and J. Malik, “Normalized cuts and image segmentation,”

IEEE Trans. on Pattern Analysis and Machine Intelligence, 2000. –  M. Belkin and P. Nyogi, “Using manifold structure for partially

labeled classification,” NIPS, 2002. –  X. Zhu, J. Kandola, J. Lafferty, and Z. Gharamani, “Nonparametric

transforms of graph kernels for semi-supervised learning,” NIPS, 2005.

–  A. Smola and R. Kondor, “Kernels and regularization on graphs,” COLT, 2003.

•  Computer graphics –  Z. Karni and C. Gotsman, “Spectral compression of mesh

geometry,” ACM Conf. on Computer Graphics and Interactive Techniques, 2000.

19

Page 20: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Why the Graph Laplacian Eigenbasis?

•  Consider a ring graph on n vertices –  Its Laplacian is circulant –  Circulant matrices diagonalized by DFT matrix

–  Eigenvalues

20

Uj,k = e2⇡ijk/n

�k = 2� 2 cos(2⇡k/n)

⇡ (2⇡k/n)2

Page 21: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Does this always make sense?

•  Consider a complete graph on n vertices –  Its Laplacian is circulant –  Circulant matrices diagonalized by DFT matrix

–  Eigenvalues

•  What does it mean to have a “smooth” signal on the complete graph?

21

Uj,k = e2⇡ijk/n

�1 = 0

�k = n k � 2

Page 22: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Smooth Signals on Graphs

Intuitively smooth on if small

Theorem: Let where is the ith eigenvector of . Then

22

x G kxkG = x

TLx

bx(�i) = hx, uii ui

L

|bx(�k)| kxkGp

�k

Page 23: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Approximating Signals on Graphs

Define M-term linear approximation of on as M-term linear approximation error Theorem:

23

x G

xM =MX

k=0

bx(�k)uk

✏l(M,x) =nX

k=M+1

|bx(�k)|2

✏l(M,x) kxk2G��1M

Page 24: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Asymptotics

Let be a graph with If then as

24

G |V | = 1

1X

k=0

k�k|bx(�k)|2 1

✏l(M,x) = o

✓1

M�M/2

◆M ! 1

Page 25: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Summary

•  GFT has many similarities to the Fourier transform –  Notion of smoothness –  Linear approximation error

•  Not all graphs support meaningful “smooth” signals –  Laplacian eigenvalues should grow

•  Can be used for “fitting” a graph to a signal or sequence of signals

25

Page 26: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

GRAPH SPECTRAL COMPRESSED SENSING

26

Page 27: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Field Estimation in Sensor Networks

Estimate sensor measurements at fusion center (FC) Performance metrics •  Distortion, MSE •  Bandwidth usage •  Energy usage

27

Page 28: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Compressed Sensing

•  Assume signal is sparse •  Measure few random

linear combinations

28

Candes & Tao, “Near-optimal signal recovery from random projections,” IEEE Trans Info Theory, 2006. D. Donoho, “Compressed sensing,” IEEE Trans Info Theory, 2006. W. Bajwa, J. Haupt, A. Sayeed, and R. Nowak, “Joint source-channel communication for distributed estimation in sensor networks,” IEEE Trans Info Theory, 2007

Page 29: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Using CS for Field Estimation

•  Pros: –  Require fewer overall measurements –  Each measurement is equally important –  Distortion performance nearly optimal

•  Cons: –  Requires synchronization across network –  Fewer total measurements, but every node transmits for

every measurement

29

Page 30: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Graph Spectral Compressed Sensing

•  Randomly sample a few

sensors •  Interpolate remaining

values wrt GFT basis

30

Page 31: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Reconstruction Guarantee

Suppose there are constants and such that If the number of measurements m obeys then, with probability , where

31

s S

✏l(M,x) SM

�s

m � C1M log(M/�)

1� �

kx� x̃k2 kx� xMk2 + C2SM�s

logdn/Me

x̃ = �†My

Page 32: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Performance Example

•  Using CIMIS data •  Comparing with

–  Gaussian random matrix: Bajwa, Haupt, Sayeed, and Nowak 2007 –  Sparse random projections: Wang, Garofalakis, Ramchandran 2007

32

0 20 40 60 80 100 120 14010−4

10−3

10−2

10−1

100

Number of measurements

Dis

torti

on(M

SE)

GSCS with oracle estimatorGaussian random matrixGSCS with BPSparse Random Projection

Page 33: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Summary

•  Graph structure can be useful for interpolation –  When signal is smooth –  (Graph should have interesting smooth signals)

•  Potential implications for –  Distributed measurement systems –  Network design –  Semi-supervised learning

33

Page 34: Processing Signals Supported on Graphsnetworks.ece.mcgill.ca/sites/default/files/Rabbat_GraphSP.pdf · transforms of graph kernels for semi-supervised learning,” NIPS, 2005. –

Discussion and Directions

•  From smoothness to sparsity

•  Connection to random walks –  Either G has interesting smooth signals –  Or it has a rapidly mixing Markov chain

•  Connection to gossip and network diffusion –  Stop early, randomly sample a few nodes, and interpolate?

•  Uncertainty principles for signals on graphs

34