Question: In this exercise, we will derive the graph convolutional networks shown in Sec. 10.5.2 from spectral graph signal processing perspective. The classic convolution on graphs

In this exercise, we will derive the graph convolutional networks shown in Sec. 10.5.2 from spectral graph signal processing perspective. The classic convolution on graphs can be computed by y=Ugθ(Λ)UTx where xRN is the input graph signal (i.e., a vector of scalars that denote feature values of all nodes at a certain feature dimension). Matrices U and Λ denote the eigen-decomposition of the normalized graph Laplacian: L=ID12AD12=UΛUT. gθ(Λ) is the function of eigenvalues and is often used as the filter function.

10.5.2 Graph convolutional networks An effective type of GNNs is called GCNs. Given an input graph A with

Input: A, adjacency matrix of the input graph of size n x n; X, the node attribute matrix of size n x d; w!

a. If we directly use the classic graph convolution formula where gθ(Λ)=diag(θ) and θ RN to compute the output y, what are the potential disadvantages?

b. Suppose we use K-polynomial filter gθ(Λ)=k=0KθkΛk. What are the benefits, compared to the above filter gθ(Λ)=diag(θ) ?

c. By applying the K th order Chebyshev polynomial approximation on the filter gθ(Λ)= k=0KθkΛk, the filter function can be written as gθ(Λ)k=0KθkTk(Λ~) with a rescaled Λ~=2λmaxΛI, where λmax denotes the largest eigenvalue of L, and θRK denotes the Chebyshev coefficients. The Chebyshev polynomials can be computed recursively by Tk(z)=2zTk1(z)Tk2(z) with T0(z)=1 and

10.5.2 Graph convolutional networks An effective type of GNNs is called GCNs. Given an input graph A with node attributes X, a GCN model uses a set of weight matrices WI (l= 1,.., L), one weight matrix for each layer, to produce the node embedding matrix Z whose rows are the embedding of the corresponding nodes. The algorithm is summarized in Fig. 10.38. The steps are described next. Preprocessing and initialization: We first add a self-edge to each node i. For a given node i, this will enable the GCN model to "remember" its own embedding while aggregating its neighboring nodes' embedding that will be described next. In terms of the adjacency matrix, adding a self edge to each node is equivalent to updating the adjacency matrix by an identity matrix I (Step 1), where I (i, i) = 1 and I (i, j) = 0 Vi # j (i, j = 1, ..., n). Then, in Step 2, we calculate the degree matrix D of the updated adjacency matrix A, where D(i, i) =E=1 A(i, j) and D(i, j)=0 Vi j (i, j = 1, ..., n). Using the degree matrix D, in Step 3, we calculate a normalized matrix as = D-/2 AD-1/2, which is also referred to as the normalized graph Laplacian of matrix A. Each element in the matrix is obtained by normalizing the corresponding element in A by the square root of the degrees of the source and the A(i, j) target nodes of the given element: (i, j): D(ii) Dj.j)* D.). In Step 4, the initial embedding is simply = set as the input node attribute matrix Z = X.

Step by Step Solution

3.42 Rating (155 Votes )

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock

a The computational complexity is ON which is very high This is mainly because we need to do the eig... View full answer

blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Data Mining Concepts And Techniques Questions!