Artificial Intelligence and Quantum Computing for Advanced Wireless Networks. Savo G. Glisic

Чтение книги онлайн.

Читать онлайн книгу Artificial Intelligence and Quantum Computing for Advanced Wireless Networks - Savo G. Glisic страница 93

Artificial Intelligence and Quantum Computing for Advanced Wireless Networks - Savo G. Glisic

Скачать книгу

stability, the recurrent function of SSE is defined as a weighted average of the historical states and new states, which takes the form

      5.2.2 ConvGNNs

      These networks are closely related to recurrent GNNs. Instead of iterating node states with contractive constraints, they address the cyclic mutual dependencies architecturally using a fixed number of layers with different weights in each layer, as illustrated in Figure 5.2a. This key distinction from recurrent GNNs is illustrated in Figures 5.2b and 5.2c. As graph convolutions are more efficient and convenient to composite with other neural networks, the popularity of ConvGNNs has been rapidly growing in recent years. These networks fall into two categories, spectral‐based and spatial‐based. Spectral‐based approaches define graph convolutions by introducing filters from the perspective of graph signal processing [44] where the graph convolutional operation is interpreted as removing noise from graph signals. Spatial‐based approaches inherit ideas from RecGNNs to define graph convolutions by information propagation. Since GCN [14] bridged the gap between spectral‐based approaches and spatial‐based approaches, spatial‐based methods have developed rapidly recently due to their attractive efficiency, flexibility, and generality.

      where ⊙ denotes the element‐wise product. If we denote a filter as gθ = diag (UTg), then the spectral graph convolution *G is simplified as

      (5.37)equation

      Spectral‐based ConvGNNs all follow this definition. The key difference lies in the choice of the filter gθ.

      Spectral Convolutional Neural Network (Spectral CNN) [12] assumes the filter images is a set of learnable parameters and considers graph signals with multiple channels. The graph convolutional layer of Spectral CNN is defined as

      (5.38)equation

      where k is the layer index, images is the input graph signal, H(0) = X, fk − 1 is the number of input channels and fk is the number of output channels, and images is a diagonal matrix filled with learnable parameters. Due to the eigen decomposition of the Laplacian matrix, spectral CNN faces three limitations: (i) any perturbation to a graph results in a change of eigenbasis; (ii) the learned filters are domain dependent, meaning they cannot be applied to a graph with a different structure; and (iii) eigen decomposition requires O(n3) computational complexity. In follow‐up works, ChebNet [45] and GCN [14, 46] reduced the computational complexity to O(m) by making several approximations and simplifications.

      Chebyshev Spectral CNN (ChebNet) [45] approximates the filter gθ by Chebyshev polynomials of the diagonal matrix of eigenvalues, that is, images, where images, and the values of images lie in [−1, 1]. The Chebyshev polynomials are defined recursively by Ti(x) = 2xTi − 1(x) − Ti − 2(x) with T0(x) = 1 and T1(x) = x. As a result, the convolution of a graph signal x with the defined filter gθ is

      (5.39)equation

      where images. As images, which can be proved by induction on i. ChebNet takes the form

      As an improvement over spectral CNN, the filters defined by ChebNet are localized in space, which means filters can extract local features independently of the graph size. The spectrum of ChebNet is mapped to [−1, 1] linearly. CayleyNet [47] further applies Cayley polynomials,

Скачать книгу