• 검색 결과가 없습니다.

Review: Summary questions of the last lecture

N/A
N/A
Protected

Academic year: 2022

Share "Review: Summary questions of the last lecture"

Copied!
29
0
0

로드 중.... (전체 텍스트 보기)

전체 글

(1)

Review: Summary questions of the last lecture

What is the meaning of eigenvalues and eigenvectors of Laplacian?

What is the meaning of the multiplicity of eigenvalue 0 of Laplacian?

What is the meaning of the smoothness of an eigenvector?

How to get the eigen-spectrum of Laplacian of the complete graph?

What is the relation btw Laplacian and random walks on undirected graphs?

(2)

Review: Summary questions of the last lecture

What is the meaning of eigenvalues and eigenvectors of Laplacian?

→ Frequencies and its corresponding graph signals that a graph can have.

A signal 𝒇 can be written as graph Fourier series:

𝒇𝑇𝒖𝑘 = ෍

𝒊

𝑓መ𝑖𝒖𝑖𝑇𝒖𝑘 = መ𝑓𝑘 = 𝒖𝑘𝑇𝒇 𝒇 = ෍

𝒊

𝑓መ𝑖𝒖𝑖 ෠𝒇 =

𝑓መ𝟏

… 𝑓መ𝑵

=

𝒖1𝑇

… 𝒖𝑁𝑇

𝒇

Spatial domain: 𝒇

෠𝒇 = 𝑼𝑇𝒇

Decompose signal 𝑓

Spectral domain: ෠𝒇 Reconstruct signal 𝒇

Design GCN in spectral domain Design GCN in spectral domain

(3)

Review: Summary questions of the last lecture

What is the meaning of the multiplicity of eigenvalue 0 of Laplacian?

→ The number of connected components in a graph.

𝑣3

𝑣2

𝑣7

𝑣5 𝑣6

𝑣4 𝑣8 𝑣1

2 − 1 − 1 0 0 0 0 0

−1 2 − 1 0 0 0 0 0

−1 − 1 3 − 1 0 0 0 0

0 0 − 1 1 0 0 0 0

0 0 0 0 1 0 − 1 0

0 0 0 0 0 2 − 1 − 1 0 0 0 0 − 1 − 1 3 − 1 𝑳𝟏

𝑳𝟐

[Intuitive Proof]

Letting two eigenvectors be 𝒖1 = [ 1 1 1 1 0 0 0 0]𝑇, 𝒖2 = [ 0 0 0 0 1 1 1 1]𝑇. Then

𝑳𝒖1 = 0𝒖1, 𝑳𝒖2 = 0𝒖2. Thus

(0, 𝒖1) and (0, 𝒖2) are eigenpairs.

The multiplicity of eigenvalue 0 of 𝐿 equals to 2.

(4)

Review: Summary questions of the last lecture

What is the meaning of the smoothness of an eigenvector?

→ The smoothness of a eigenvector is its eigenvalue (frequency).

𝑆𝐺 𝒇 = 𝒇𝑇𝑳𝒇 = 𝒇𝑇𝑼𝚲𝑼𝑻𝒇 = 𝜶𝑻𝚲𝜶 = 𝜶 𝚲𝟐 = ෍

𝟏≤𝒊≤𝑵

𝜆𝒊𝛼𝒊𝟐, 𝜶 = 𝑼𝑻𝒇

Spectral coordinate (unique vector) of eigenvector 𝒖𝑘: 𝜶𝑘= 𝑼𝑇𝒖𝑘 = 𝒆𝑘. 𝑆𝐺 𝒖𝑘 = 𝒖𝑘𝑇𝑳𝒖𝑘 = 𝒖𝑘𝑇𝑼𝚲𝑼𝑇𝒖𝑘 = 𝒆𝑘𝑇𝚲𝒆𝑘 = 𝒆𝑘 𝚲𝟐 = ෍

1≤𝑖≤𝑁

𝜆𝑖𝑒𝑘,𝑖2 = 𝜆𝑘

(5)

Review: Summary questions of the last lecture

How to get the eigen-spectrum of Laplacian of the complete graph?

→ The 1-st eigen-pair is (0, 𝟏𝑁) and compute the second eigen-pair of which eigenvector is orthogonal to 1-st eigenvector. The remaining ones can be computed to be orthogonal to the previous ones.

If 𝒖 ≠ 𝟎 and 𝒖 ⊥ 𝟏𝑵 ⇒ σ𝑖 𝑢𝑖 = 0 . To get the other eigenvalues, we compute (𝑳𝑲𝑵𝒖)1and divide by 𝑢1 (letting 𝑢1≠ 0).

(𝑳𝑲𝑵𝒖)1 = (𝑁 − 1)𝑢1 − ෍

2≤𝑖≤𝑁

𝑢𝑖 = 𝑁𝑢1

→ (0, 𝟏𝑁), (𝑁, 1 − 1 0 … 0 𝑇), …

(6)

Review: Summary questions of the last lecture

What is the relation btw Laplacian and random walks on undirected graphs?

→ The random walks is a stochastic process with a transition probability 𝑝𝑖𝑗 =

𝑤𝑖𝑗

𝑑𝑖 between node 𝑖 and 𝑗 of a graph with a Laplacian 𝐿 = 𝐷 − 𝑊.

Transition matrix: 𝑷 = 𝑝𝑖𝑗 = 𝑫−1𝑾 (notice 𝑳𝑟𝑤 = 𝐈 − 𝑷).

Unique stationary distribution 𝝅 = 𝜋1, … , 𝜋𝑁 where 𝜋𝑖 = 𝑑𝑖

𝑣𝑜𝑙 𝑽 .

← 𝑣𝑜𝑙 𝑮 = 𝑣𝑜𝑙 𝑽 = 𝑣𝑜𝑙 𝑾 ≜ σ𝑖 𝑑𝑖 = σ𝑖𝑗 𝑤𝑖𝑗. 𝝅 = 𝟏𝑇𝑾

𝑣𝑜𝑙(𝑾) verifies 𝝅𝑷 = 𝝅 as 𝝅𝑷 = 𝟏𝑇𝑾𝑷

𝑣𝑜𝑙(𝑾) = 𝟏𝑇𝑫𝑷

𝑣𝑜𝑙(𝑾) = 𝟏𝑇𝑫𝑫−1𝑾

𝑣𝑜𝑙(𝑾) = 𝟏𝑇𝑾

𝑣𝑜𝑙(𝑾) = 𝝅.

(7)

Spectral Clustering

𝑐𝑢𝑡 𝑨, 𝑩 = 1

2 𝒇 𝑻 𝑳𝒇

𝑣3

𝑣2

𝑣5 𝑣6

𝑣4 𝑣8 𝑣1

𝒇 =

1

−1

−1 1

−1

(8)

Application of Graphs for ML: Clustering

(9)

Spectral Clustering: Cuts on graphs

(10)

Spectral Clustering: Cuts on graphs

Defining the cut objective we get the clustering!

(11)

Spectral Clustering: Cuts on graphs

MinCut: 𝑐𝑢𝑡 𝑨, 𝑩 = σ𝑖∈𝑨,𝑗∈𝑩 𝑤𝑖𝑗 Are we done?

(12)

Spectral Clustering: Balanced Cuts

MinCut

RatioCut

Let’s balance the cuts!

NormalizedCut

𝐶𝑢𝑡 𝑨, 𝑩 = σ𝑖∈𝑨,𝑗∈𝑩 𝑤𝑖𝑗 s.t. 𝑨 = 𝑩

𝑅𝐶𝑢𝑡 𝑨, 𝑩 = ෍

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 1

𝑨 + 1 𝑩

𝑁𝐶𝑢𝑡 𝑨, 𝑩 = ෍

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 1

𝑣𝑜𝑙(𝑨) + 1 𝑣𝑜𝑙(𝑩)

(13)

Spectral Clustering: Balanced Cuts

𝑅𝐶𝑢𝑡 𝑨, 𝑩, 𝑪, … = 𝐶𝑢𝑡 𝑨, 𝑩, 𝑪, … 1

𝑨 + 1

𝑩 + 1

𝑪 + … 𝑁𝐶𝑢𝑡 𝑨, 𝑩, 𝑪, … = 𝐶𝑢𝑡 𝑨, 𝑩, 𝑪, … 1

𝑣𝑜𝑙(𝑨) + 1

𝑣𝑜𝑙(𝑩) + 1

𝑣𝑜𝑙(𝑪) + … Easily generalizable to 𝑘 ≥ 2.

Can we compute this? 𝐶𝑢𝑡, 𝑅𝐶𝑢𝑡 and 𝑁𝐶𝑢𝑡 are NP hard.

Approximate! (Relaxation)

𝐶𝑢𝑡 𝑨, 𝑩 = σ𝑖∈𝑨,𝑗∈𝑩𝑤𝑖𝑗 s.t. 𝑨 = 𝑩 = 𝑪

(14)

Spectral Clustering: Relaxing Balanced Cuts

Laplacian formulation for simple balanced cuts for 2 sets

What is the cut value with this definition?

Graph function 𝒇 for cluster membership: 𝑓𝑖 = ቊ 1 𝑖𝑓 𝑣𝑖 ∈ 𝑨

−1 𝑖𝑓 𝑣𝑖 ∈ 𝑩 min𝑨,𝑩 𝐶𝑢𝑡 𝑨, 𝑩 subject to 𝑨 = 𝑩

𝐶𝑢𝑡 𝑨, 𝑩 = ෍

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 = 1

4 ෍

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 (𝑓𝑖 − 𝑓𝑗)2 = 1

2 𝒇𝑇𝑳𝒇 What is the relationship with the smoothness of a graph function?

(15)

Spectral Clustering: Relaxing Balanced Cuts

Optimization formulation for spectral clustering Constraints:

𝑨 = 𝑩 ↔ σ𝒊 𝑓𝑖 = 0 ↔ 𝒇𝑇𝟏𝑵 = 0 ↔ 𝒇 ⊥ 𝟏𝑵

𝐦𝐢𝐧𝒇 𝒇𝑇𝑳𝒇 subject to 𝑓𝑖 = ±1, 𝒇𝑇𝟏𝑵 = 0, 𝒇 = 𝑵 Objective:

𝐶𝑢𝑡 𝑨, 𝑩 = ෍

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 = 1

4 ෍

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 (𝑓𝑖 − 𝑓𝑗)2 = 1

2 𝒇𝑇𝑳𝒇

Still NP hard (∵ 𝑓𝑖 = ±1, nonconvex). → relax even further!

𝑓𝑖 = ±1 → 𝑓𝑖∈ 𝑅 or 0 ≤ 𝑓𝑖≤ 1

(16)

Spectral Clustering: Relaxing Balanced Cuts

Optimization formulation for spectral clustering

𝐦𝐢𝐧𝒇 𝒇𝑇𝑳𝒇 subject to 𝑓𝑖∈ 𝑅, 𝒇𝑇𝟏𝑵 = 0, 𝒇 = 𝑵

Rayleigh-Ritz theorem

If 𝜆1 ≤ … ≤ 𝜆𝑁 are the eigenvalues of real symmetric 𝑳, then 𝜆1 = 𝐦𝐢𝐧

𝒙≠𝟎

𝒙𝑇𝑳𝒙

𝒙𝑇𝒙 = 𝐦𝐢𝐧

𝒙𝑇𝒙=𝟏 𝒙𝑇𝑳𝒙 𝜆𝑁 = 𝐦𝐚𝐱

𝒙≠𝟎

𝒙𝑇𝑳𝒙

𝒙𝑇𝒙 = 𝐦𝐚𝐱

𝒙𝑇𝒙=𝟏 𝒙𝑇𝑳𝒙

𝒙𝑇𝑳𝒙

𝒙𝑇𝒙 is Rayleigh quotient How can we use it?

Proof:

Derivative yields 𝑳𝒙 = 𝜆𝒙

(17)

Spectral Clustering: Relaxing Balanced Cuts

Optimization formulation for spectral clustering

𝐦𝐢𝐧𝒇 𝒇𝑇𝑳𝒇 subject to 𝑓𝑖∈ 𝑅, 𝒇 ⊥ 𝟏𝑵, 𝒇 = 𝑵 Generalized Rayleigh-Ritz theorem (Courant-Fischer-Weyl) If 𝜆1 ≤ … ≤ 𝜆𝑁 are the eigenvalues of real symmetric 𝑳, and

𝒖1, … , 𝒖𝑁 the corresponding orthogonal eigenvectors, for 𝑘 = 1, … , 𝑁 − 1 𝜆𝑘+1 = 𝐦𝐢𝐧

𝒙≠𝟎, 𝒙⊥𝒖1,…,𝒖𝑘

𝒙𝑇𝑳𝒙

𝒙𝑇𝒙 = 𝐦𝐢𝐧

𝒙𝑇𝒙=𝟏, 𝒙⊥𝒖1,…,𝒖𝒌 𝒙𝑇𝑳𝒙 𝜆𝑁−𝑘 = 𝐦𝐚𝐱

𝒙≠𝟎, 𝒙⊥𝒖𝑁−𝑘+1,…,𝒖𝑁

𝒙𝑇𝑳𝒙

𝒙𝑇𝒙 = 𝐦𝐚𝐱

𝒙𝑇𝒙=𝟏, 𝒙⊥𝒖𝑁−𝑘+1,…,𝒖𝑁 𝒙𝑇𝑳𝒙

The solution becomes the second eigenvector which can be obtain by 𝑘 = 1.

(18)

Spectral Clustering: Relaxing Balanced Cuts

Optimization formulation for spectral clustering

𝐦𝐢𝐧𝒇 𝒇𝑇𝑳𝒇 subject to 𝑓𝑖∈ 𝑅, 𝒇 ⊥ 𝟏𝑵, 𝒇 = 𝑵

The solution

𝜆2 = 𝐦𝐢𝐧

𝒙𝑇𝒙=𝟏, 𝒙⊥𝒖1 𝒙𝑇𝑳𝒙, where 𝒖1 = 𝟏𝑵 for 𝜆1 = 0

→ second eigenvector 𝒙 of 𝑳

Since the elements in 𝒙 are not integer and 𝒙 = 𝟏, 𝒇 can be obtained by 𝑓𝑖 = ቊ 1 𝑖𝑓 𝑥𝑖 ≥ 0

−1 𝑖𝑓 𝑥𝑖 < 0 → 𝒇 = 𝑵 → 𝐴 = 𝐵 𝑓𝑖∈ {1, −1}

(19)

Spectral Clustering: Approximating RatioCut

RatioCut

Define graph function 𝒇 for cluster membership of RatioCut: 𝑓𝑖 = ൞

𝑩

𝑨 𝑖𝑓 𝑣𝑖 ∈ 𝑨

𝑨

𝑩 𝑖𝑓 𝑣𝑖 ∈ 𝑩

min𝑨,𝑩 𝑅𝐶𝑢𝑡 𝑨, 𝑩 = min

𝑨,𝑩

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 1

𝑨 + 1 𝑩

𝒇𝑇𝑳𝒇 = 1

2 ෍

𝑖,𝑗

𝑤𝑖𝑗 (𝑓𝑖 − 𝑓𝑗)2= 𝑨 + 𝑩 𝑅𝐶𝑢𝑡 𝑨, 𝑩 Since 𝑨 + 𝑩 is constant, min

𝑨,𝑩 𝑅𝐶𝑢𝑡 𝑨, 𝑩 = min

𝒇 𝒇𝑇𝑳𝒇, subject to 𝑓𝑖𝑩

𝑨 , − 𝑨

𝑩

𝑨 = 𝑩

(20)

Spectral Clustering: Approximating RatioCut

Optimization formulation for RatioCut (same with balanced mincut) 𝐦𝐢𝐧𝒇 𝒇𝑇𝑳𝒇 subject to, 𝑓𝑖 ∈ 𝑅, 𝒇 = 𝑵

𝑠. 𝑡. 𝑓𝑖 =

𝑩

𝑨 𝑖𝑓 𝑣𝑖 ∈ 𝑨

𝑨

𝑩 𝑖𝑓 𝑣𝑖 ∈ 𝑩

min𝑨,𝑩 𝑅𝐶𝑢𝑡 𝑨, 𝑩 = min

𝑨,𝑩 𝒇𝑇𝑳𝒇

𝒇 𝟐 = σ𝑖 𝑓𝑖2 = 𝑨 𝑩

𝑨 + 𝑩 𝑨

𝑩 = 𝑨 + 𝑩 = 𝑵 → not sufficient for 𝑓𝑖 𝑩

𝑨 , − 𝑨

𝑩

𝑨 = 𝑩 𝒇𝑇𝑳𝒇 ≠ 𝑨 + 𝑩 ෍

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 1

𝑨 + 1 𝑩

(21)

Spectral Clustering: Approximating RatioCut

Optimization formulation for RatioCut (same with balanced Mincut) 𝐦𝐢𝐧𝒇 𝒇𝑇𝑳𝒇 subject to 𝑓𝑖∈ 𝑅, 𝒇 ⊥ 𝟏𝑵, 𝒇 = 𝑵

𝑨 = 𝑩

𝑖

𝑓𝑖 = 0 ↔ 𝒇 ⊥ 𝟏𝑵

Optimization formulation for RatioCut (same with balanced mincut) 𝐦𝐢𝐧𝒇 𝒇𝑇𝑳𝒇 subject to 𝑓𝑖∈ 𝑅, 𝒇 = 𝑵

(22)

Spectral Clustering: Approximating RatioCut

The solution

𝜆2 = 𝐦𝐢𝐧

𝒙𝑇𝒙=𝟏, 𝒙⊥𝒖1 𝒙𝑇𝑳𝒙, where 𝒖1 = 𝟏𝑵 for 𝜆1 = 0.

→ second eigenvector 𝒙 of 𝑳

Since the elements in 𝒙 are not integer and 𝒙 = 𝟏, 𝒇 can be obtained by

𝑓𝑖 =

𝑩

𝑨 𝑖𝑓 𝑥𝑖 ≥ 0

𝑨

𝑩 𝑖𝑓 𝑥𝑖 < 0

→ 𝒇 = 𝑵

𝑨 = 𝑩 𝒇𝑇𝑳𝒇 ≠ 𝑨 + 𝑩 ෍

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 1

𝑨 + 1 𝑩

(23)

Spectral Clustering: Approximating NormalizedCut

NormalizedCut

Balancing the clusters by considering the degrees of nodes Define graph function 𝒇 for cluster membership of NCut: 𝑓𝑖 = ൞

𝑣𝑜𝑙(𝑩)

𝑣𝑜𝑙(𝑨) 𝑖𝑓 𝑣𝑖 ∈ 𝑨

𝑣𝑜𝑙(𝑨)

𝑣𝑜𝑙(𝑩) 𝑖𝑓 𝑣𝑖 ∈ 𝑩

min𝑨,𝑩 𝑁𝐶𝑢𝑡 𝑨, 𝑩 = min

𝑨,𝑩

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 1

𝑣𝑜𝑙(𝑨) + 1 𝑣𝑜𝑙(𝑩)

min𝐴,𝐵 𝒇𝑇𝑳𝒇 = 𝑣𝑜𝑙 𝒱 𝑁𝐶𝑢𝑡 𝑨, 𝑩 , 𝑓𝑖 𝑣𝑜𝑙(𝑩)

𝑣𝑜𝑙(𝑨) , − 𝑣𝑜𝑙(𝑨) 𝑣𝑜𝑙(𝑩) 𝒇𝑇𝑳𝒇 = ෍

𝑖,𝑗

𝑤𝑖𝑗 𝑣𝑜𝑙(𝑩)

𝑣𝑜𝑙(𝑨) + 𝑣𝑜𝑙(𝑨) 𝑣𝑜𝑙(𝑩)

2

= ෍

𝑖,𝑗

𝑤𝑖𝑗 𝑣𝑜𝑙 𝑩 + 𝑣𝑜𝑙(𝑨) 𝑣𝑜𝑙(𝑨) 𝑣𝑜𝑙(𝑩)

2

𝑣𝑜𝑙 𝑩 = 𝑣𝑜𝑙(𝑨)

(24)

Spectral Clustering: Approximating NormalizedCut

NormalizedCut

Define graph function 𝒇 for cluster membership of NCut: 𝑓𝑖 = ൞

𝑣𝑜𝑙(𝑩)

𝑣𝑜𝑙(𝑨) 𝑖𝑓 𝑣𝑖 ∈ 𝑨

𝑣𝑜𝑙(𝑨)

𝑣𝑜𝑙(𝑩) 𝑖𝑓 𝑣𝑖 ∈ 𝑩

min𝑨,𝑩 𝑁𝐶𝑢𝑡 𝑨, 𝑩 = min

𝑨,𝑩

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 1

𝑣𝑜𝑙(𝑨) + 1 𝑣𝑜𝑙(𝑩)

(𝑫𝒇)𝑇𝟏𝑁 = 0, 𝒇𝑇𝑫𝒇 = 𝑣𝑜𝑙 𝒱 Necessary condition for 𝑓𝑖𝑣𝑜𝑙(𝑩)

𝑣𝑜𝑙(𝑨) , − 𝑣𝑜𝑙(𝑨)

𝑣𝑜𝑙(𝑩)

(25)

Spectral Clustering: Approximating NormalizedCut

Optimization formulation for NormalizedCut

𝐦𝐢𝐧𝒇 𝒇𝑇𝑳𝒇 subject to 𝑓𝑖∈ 𝑅, 𝑫𝒇 ⊥ 𝟏𝑵, 𝒇𝑇𝑫𝒇 = 𝑣𝑜𝑙 𝒱 NormalizedCut

Define graph function 𝒇 for cluster membership of NCut: 𝑓𝑖 = ൞

𝑣𝑜𝑙(𝑩)

𝑣𝑜𝑙(𝑨) 𝑖𝑓 𝑣𝑖 ∈ 𝑨

𝑣𝑜𝑙(𝑨)

𝑣𝑜𝑙(𝑩) 𝑖𝑓 𝑣𝑖 ∈ 𝑩

min𝑨,𝑩 𝑁𝐶𝑢𝑡 𝑨, 𝑩 = min

𝑨,𝑩

𝑖∈𝑨,𝑗∈𝑩

𝑤𝑖𝑗 1

𝑣𝑜𝑙(𝑨) + 1 𝑣𝑜𝑙(𝑩)

𝒇𝑇𝑳𝒇 = 𝑣𝑜𝑙 𝒱 𝑁𝐶𝑢𝑡 𝑨, 𝑩 , (𝑫𝒇)𝑇𝟏𝑁 = 0, 𝒇𝑇𝑫𝒇 = 𝑣𝑜𝑙 𝒱 ,

(26)

Spectral Clustering: Approximating NormalizedCut

Optimization formulation for NormalizedCut

𝐦𝐢𝐧𝒇 𝒇𝑇𝑳𝒇 subject to 𝑓𝑖∈ 𝑅, 𝑫𝒇 ⊥ 𝟏𝑵, 𝒇𝑇𝑫𝒇 = 𝑣𝑜𝑙 𝒱

Can we apply Rayleigh-Ritz now? Define 𝒉 = 𝑫𝟏/𝟐𝒇 Optimization formulation for NormalizedCut

𝐦𝐢𝐧𝒉 𝒉𝑇𝑫−1/2𝑳𝑫−𝟏/𝟐𝒉 subject to ℎ𝑖∈ 𝑅, 𝒉 ⊥ 𝒖𝟏,𝑳𝒔𝒚𝒎, 𝒉𝑇𝒉 = 𝑣𝑜𝑙 𝒱

𝐦𝐢𝐧𝒉 𝒉𝑇𝑳𝒔𝒚𝒎𝒉 subject to ℎ𝑖∈ 𝑅, 𝒉 ⊥ 𝒖𝟏,𝑳𝒔𝒚𝒎, 𝒉 = 𝑣𝑜𝑙 𝒱

(27)

Spectral Clustering: Approximating NormalizedCut

Optimization formulation for NormalizedCut

Solution by Rayleigh-Ritz? 𝒉 = 𝒖𝟐,𝑳𝒔𝒚𝒎, 𝒇 = 𝑫−𝟏/𝟐𝒉 𝐦𝐢𝐧𝒉 𝒉𝑇𝑳𝒔𝒚𝒎𝒉 subject to ℎ𝑖∈ 𝑅, 𝒉 ⊥ 𝒖𝟏,𝑳𝒔𝒚𝒎, 𝒉 = 𝑣𝑜𝑙 𝒱

𝑓𝑖

𝑣𝑜𝑙(𝑩)

𝑣𝑜𝑙(𝑨) 𝑖𝑓 ℎ𝑖 ≥ 0

− 𝑣𝑜𝑙(𝑨)

𝑣𝑜𝑙(𝑩) 𝑖𝑓 ℎ𝑖 < 0

↔ 𝒇𝑇𝑫𝒇 = 𝑣𝑜𝑙 𝒱 → 𝑣𝑜𝑙 𝑨 = 𝑣𝑜𝑙(𝑩)

→ eigenvector of 𝑳𝒓𝒘

→ 𝐿𝑢 = 𝜆𝐷𝑢

(28)

Spectral Clustering: Bibliography

 Random Work: M. Meila et al. “A random walks view of spectral

segmentation”. In: International Conference on Artificial Intelligence and Statistics (2001)

 𝑳𝒔𝒚𝒎: Andrew Y Ng, Michael I Jordan, and Yair Weiss. “On spectral clustering: Analysis and an algorithm”. In: Neural Information

Processing Systems. 2001

 𝑳𝒓𝒎: J Shi and J Malik. “Normalized Cuts and Image Segmentation”. In:

IEEE Transactions on Pattern Analysis and Machine Intelligence 22 (2000), pp. 888–905

 Things can go wrong with the relaxation: Daniel A. Spielman and Shang H. Teng. “Spectral partitioning works: Planar graphs and finite element meshes”. In: Linear Algebra and Its Applications 421 (2007), pp. 284–

305

(29)

Summary questions of Lecture

 Explain the meaning of Spectral Clustering in one sentence. Why spectral?

 What is represented by the solution of Laplacian formulation relaxing Balanced Graph cut problem?

 What does the solution of MinCut problem mean?

 What does the solution of NormalizedCut problem mean?

참조

관련 문서

Mixing law models predictions of thermal conductivity as a function of porosity for water- saturated quartz sand compared with experimental measurement

→ The central differences lie in the type of filters (or transition matrices in random walk-sense) used, and whether explicitly retaining original features (or random

 Discuss the machine learning background of auto-regressive model to obtain the

→ Hinge loss penalizes the case that the classifier f(w, x) does not decide the correct class of labeled x with the score margin of y f(w, x)&gt;=1.?. Questions of

→ Graph U-Nets downsamples the graph using gPool, upsamples it using gUnpool, and transforms graph features with the simplified ChebNet. The scores are also used to gate

the

 Explain the key idea of ChebNet: Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering (Defferard et al. NIPS 2016).  Explain the key idea

• Gantt charts illustrate the start and finish dates of the terminal elements and summary elements of a project.. • Terminal elements and summary elements comprise the work