• 검색 결과가 없습니다.

Summary Questions of the last lecture

N/A
N/A
Protected

Academic year: 2022

Share "Summary Questions of the last lecture"

Copied!
23
0
0

로드 중.... (전체 텍스트 보기)

전체 글

(1)

J. Y. Choi. SNU

Summary Questions of the last lecture

1

 What is the key idea to mitigate the scalability problem of SSL for millions of data?

→ We approximate the graph signal by a linear combination of only

the 𝑘 smoothest eigenvectors with 𝑘 Fourier coefficients and then we the 𝑘 Fourier coefficients are used for optimization variables instead of the 𝑁 elements of the graph signal.

(2)

Summary Questions of the lecture

 What happens to 𝑳 when 𝑁 → ∞?

→ As 𝑁 → ∞, Laplacian (𝑳) smoothness normalized by 𝑁2 over a graph signal 𝒇 defined on data set {𝑥𝑖} sampled from 𝑝 𝑥 converges to a weighted smoothness operator on a continuous function 𝐹(𝑥) defined over a measure space 𝒳.

(3)

J. Y. Choi. SNU

Summary Questions of the lecture

3

 What does the weighted smoothness operator in Hilbert space define, instead of eigenvectors of Laplacian 𝐿?

→ It defines eigenfunctions, each of which represents a function having a unique frequency (eigenvalue) in the Hilbert space.

(4)

Summary Questions of the lecture

 How can we obtain numerical eigenfunctions with sampled large dada?

→ After rotating the coordinate of the sample space using PCA, we divide each coordinate into B number of bins, and determine the discrete probability distribution by counting the number of nodes (samples) that fall into each bin. Then, we solve the generalized

eigenvalue problem to obtain an approximate eigenfunction having a constant in each bin.

(5)

J. Y. Choi. SNU

Summary Questions of the lecture

5

 How can we determine eigenvectors from the numerical eigenfunctions?

→ Each element value of an eigenvector is assigned by the

eigenfunction value of the bin where the corresponding sample (node) falls into.

(6)

Outline of Lecture (2)

Graph Convolution Networks (GCN)

What are issues on GCN

Graph Filtering in GCN

Graph Pooling in GCN

Spectral Filtering in GCN

Spatial Filtering in GCN

Recent GCN papers

Original GNN (Scarselli et al. 2005)

Spectral Graph CNN (Bruna et al. ICLR 2014)

GCN (Kipf & Welling. ICLR 2017)

GraphSage (Hamilton et al. NIPS 2017)

GAT (Veličković et al. ICLR 2018)

MPNN (Glimer et al. ICML 2017)

Link Analysis

PageRank

Diffusion

Recent Diffusion Papers

Semi-supervised Learning (SSL) : conti.

SSL with Graph using Regularized Harmonic Functions

SSL with Graph using Soft Harmonic Functions

SSL with Graph using Manifold Regularization (out of sample extension)

SSL with Graph using Laplacian SVMs

SSL with Graph using Max-Margin Graph Cuts

Online SSL

SSL for large graph

Review: Convolution Neural Networks (CNN)

Feedforward Neural Networks

Convolution Integral (Temporal)

Convolution Sum (Temporal)

Circular Convolution Sum

Convolution Sum (Spatial)

Convolutional Neural Networks

(7)

J. Y. Choi. SNU

Brief Review on Convolution Neural Network

7

CNN

(8)

R: Feedforward Neural Networks

Convolution Neural Networks?

(9)

J. Y. Choi. SNU

R: Convolutional Neural Networks

Convolution Integral (Temporal)

6

𝐿 ∘ 𝑓 ∗ 𝑔 = 𝐹 𝑠 𝐺 𝑠 cf. cross correlation

(10)

R: Convolutional Neural Networks

Convolution Sum (Temporal)

(11)

J. Y. Choi. SNU

R: Convolutional Neural Networks

Circular Convolution Sum

11

𝑘 = 𝑓 ∗ 𝑔 = ෍

𝑖=0 𝑛−1

𝑓𝑖𝑔𝑘−𝑖 𝑔 ≜ (… , 𝑔𝑛−1, 𝑔0, 𝑔1, … , 𝑔𝑛−1, 𝑔0, 𝑔1, … , 𝑔𝑛−1, …) 𝑓 ≜ (𝑓0, 𝑓1, … , 𝑓𝑛−1)

ℎ ≜ (… , ℎ𝑛−1, ℎ0, ℎ1, … , ℎ𝑛−1, ℎ0, ℎ1, … , ℎ𝑛−1, …)

Circulant Matrix

Correlation Filter 𝑓 ∗ 𝑔 =

𝑔0 𝑔𝑛−1 … 𝑔2 𝑔1 𝑔1 𝑔0 𝑔𝑛−1 … 𝑔2

⋮ 𝑔𝑛−1

𝑔𝑛

𝑔1

⋮ 𝑔𝑛−1

𝑔0

⋱ 𝑔1

⋮ 𝑔𝑛−2

𝑔0

𝑓0 𝑓1

⋮ 𝑓𝑛−1

𝑓𝑛

(12)

R: Convolutional Neural Networks

Convolution Sum (Spatial)

(13)

J. Y. Choi. SNU

R: Convolutional Neural Networks

𝑜 = 𝑓 𝑊, 𝑥

13

(14)

Geometric deep learning via graph convolution …

GCN (𝒢)

𝒉𝒊(𝒍+𝟏) = ෍

𝒗𝒋∈𝑵(𝒗𝒊)

𝒇 𝒙𝒊, 𝒘𝒊𝒋, 𝒉𝒋(𝒍), 𝒙𝒋

𝑥3, ℎ3𝑙

𝑣2 𝑣8 𝑣1

𝑣3 𝑣4

𝑣5

𝑣6 𝑣7

𝑥1, ℎ1𝑙

𝑥4, ℎ4𝑙

𝑥5, ℎ5𝑙

𝑥4, ℎ4𝑙 𝑥5, ℎ5𝑙

𝑥5, ℎ5𝑙 𝑥5, ℎ5𝑙

(15)

J. Y. Choi. SNU

GCN: What are issues?

15

Issues:

 Laplacian smoothing(LS) on input data space

 LS under the fixed graph structure

(16)

GCN: What are issues?

Issues:

 Laplacian smoothing(LS) on input data space

 Inefficient due to redundancy by high dimension

→ Do LS in feature space

→ Feature embedding in GCN

 LS under the fixed graph structure

(17)

J. Y. Choi. SNU

GCN: What are issues?

17

Issues:

 Laplacian smoothing(LS) on input data space

 Inefficient due to redundancy by high dimension

→ Do LS in feature space

→ Feature embedding in GCN

 LS under the fixed graph structure

 Highly dependent on graph structure?

→ Update graph structure

→ How to do? → Attentional aggregation in GCN, Diffusion, etc.

(18)

GCN: Two Main Operations

Node feature representation Node embedding

𝒉𝒊(𝒍+𝟏) =

𝒗𝒋∈𝑵(𝒗𝒊)

𝒇 𝒍𝒊, 𝒘𝒊𝒋, 𝒉𝒋(𝒍), 𝒍𝒋

𝑨 ∈ 𝟎, 𝟏 𝒏×𝒏, 𝑿 ∈ ℝ𝒏×𝒅 𝑨 ∈ 𝟎, 𝟏 𝒏×𝒏, 𝑯𝒍 ∈ ℝ𝒏×𝒅𝒍 Graph filtering

(19)

J. Y. Choi. SNU

GCN: Two Main Operations

19

A smaller graph

Graph Pooling

𝑨 ∈ 𝟎, 𝟏 𝒏×𝒏, 𝑿 ∈ ℝ𝒏×𝒅 𝑨𝒑 ∈ 𝟎, 𝟏 𝒏𝒑×𝒏𝒑, 𝑯𝒑 ∈ ℝ𝒏𝒑×𝒅𝒑 Graph pooling

(20)

GCN: Types of Operations for Graph Filtering

Spatial filtering Spectral filtering

Original GNN (Scarselli et al.

2005)

(Kipf & Welling. GCN ICLR 2017) (VeličkovićGAT et al.

ICLR 2018)

GraphSage (Hamilton et al.

NIPS 2017)

(Glimer et al. MPNN ICML 2017)

Spectral Graph CNN (Bruna et al.

ICLR 2014)

ChebNet (Defferard et al.

NIPS 2016)

(21)

J. Y. Choi. SNU

GCN: Graph Filtering

21

Original GNN: Graph neural networks for ranking web pages, IEEE Web Intelligence (Scarselli et al. 2005)

𝒉𝑖: hidden features

𝒙𝑖: input features Spatial filtering

𝑥3, ℎ3𝑙

𝑣2 𝑣8 𝑣1

𝑣3 𝑣4

𝑣5

𝑣6 𝑣7

𝑥1, ℎ1𝑙

𝑥4, ℎ4𝑙

𝑥5, ℎ5𝑙

𝑥4, ℎ4𝑙 𝑥5, ℎ5𝑙

𝑥5, ℎ5𝑙 𝑥5, ℎ5𝑙

(22)

GCN: Graph Filtering

𝒉𝑖: hidden features

𝒙𝑖: input features

𝒉𝑖(𝑙+1) = ෍

𝑣𝑗∈𝑁(𝑣𝑖)

𝒇 𝒙𝑖, 𝒉𝑗(𝑙), 𝒙𝑗 , ∀ 𝑣𝑖∈ 𝑉.

𝒇 ⋅ : feedforward neural network.

𝑁(𝑣𝑖): neighbors of the node 𝑣𝑖. Spatial filtering

GNN

𝑥3, ℎ3𝑙

𝑣2 𝑣8 𝑣1

𝑣3 𝑣4

𝑣5

𝑣6 𝑣7

𝑥1, ℎ1𝑙

𝑥4, ℎ4𝑙

𝑥5, ℎ5𝑙

𝑥4, ℎ4𝑙 𝑥5, ℎ5𝑙

𝑥5, ℎ5𝑙 𝑥5, ℎ5𝑙

(23)

J. Y. Choi. SNU

Summary Questions of the lecture

23

 Explain the concept of convolution sum and its meaning in the spatial and temporal applications.

 What problems in the conventional Laplacian smoothness applications does the GCN try to solve?

 Explain the two main operations in GCN.

참조

관련 문서

_____ culture appears to be attractive (도시의) to the

1 John Owen, Justification by Faith Alone, in The Works of John Owen, ed. John Bolt, trans. Scott Clark, "Do This and Live: Christ's Active Obedience as the

In optics and lens design, the Abbe number, also known as the V-number or constringence of a transparent material, is a measure of the material's dispersion (variation

Mixing law models predictions of thermal conductivity as a function of porosity for water- saturated quartz sand compared with experimental measurement

molecules slightly outside the cut-off distance (buffer).. Algorithm for Integration.. at Buffalo, School of Eng. David Kofke).. at Buffalo, School of Eng. David Kofke).. Nature

→ The central differences lie in the type of filters (or transition matrices in random walk-sense) used, and whether explicitly retaining original features (or random

 Discuss the machine learning background of auto-regressive model to obtain the

→ Hinge loss penalizes the case that the classifier f(w, x) does not decide the correct class of labeled x with the score margin of y f(w, x)>=1.?. Questions of