Artificial Intelligence and Quantum Computing for Advanced Wireless Networks. Savo G. Glisic

Чтение книги онлайн.

Читать онлайн книгу Artificial Intelligence and Quantum Computing for Advanced Wireless Networks - Savo G. Glisic страница 105

Artificial Intelligence and Quantum Computing for Advanced Wireless Networks - Savo G. Glisic

Скачать книгу

j = 1, …, n . If wij = 0, this means that the vertices vi and vj are not connected by an edge. As G is undirected, we require wij = wji . The degree of a vertex viV is defined as

equation

      Note that, in fact, this sum only runs over all vertices adjacent to vi , as for all other vertices vj the weight wij is 0. The degree matrix D is defined as the diagonal matrix with the degrees d1 , …, dn on the diagonal. Given a subset of vertices AV, we denote its complement V\A by images. We define the indicator vector IA = (f1, …, fn)′ ∈ n as the vector with entries fi = 1 if viA and fi = 0 otherwise. For convenience, we introduce the shorthand notation iA for the set of indices {iviA}, in particular when dealing with a sum like ∑i ∈ A wij . For two not necessarily disjoint sets A, BV we define

equation

      We consider two different ways of measuring the “size” of a subset ⊂A:

images

      Intuitively, ∣A∣ measures the size of A by its number of vertices, whereas vol(A) measures the size of A by summing over the weights of all edges attached to vertices in A. A subset AV of a graph is connected if any two vertices in A can be joined by a path such that all intermediate points also lie in A. A subset A is called a connected component if it is connected and if there are no connections between vertices in A and images. The nonempty sets A1 , …, Ak form a partition of the graph if AiAj = ∅ and A1∪ … ∪Ak = V.

      Similarity graphs: There are several popular constructions to transform a given set x1 , …, xn of data points with pairwise similarities sij or pairwise distances dij into a graph. When constructing similarity graphs, the goal is to model the local neighborhood relationships between the data points.

      The ε‐neighborhood graph: Here, we connect all points whose pairwise distances are smaller than ε. As the distances between all connected points are roughly of the same scale (at most ε), weighting the edges would not incorporate more information about the data to the graph. Hence, the ε‐neighborhood graph is usually considered an unweighted graph.

      k ‐nearest neighbor graphs: Here, the goal is to connect vertex vi with vertex vj if vj is among the k‐nearest neighbors of vi . However, this definition leads to a directed graph, as the neighborhood relationship is not symmetric. There are two ways of making this graph undirected. The first way is to simply ignore the directions of the edges, that is, we connect vi and vj with an undirected edge if vi is among the k‐nearest neighbors of vj or if vj is among the k‐nearest neighbors of vi . The resulting graph is what is usually called the k‐nearest neighbor graph. The second choice is to connect vertices vi and vj if both of the following are true: (i) vi is among the k‐nearest neighbors of vj and (ii) vj is among the k‐nearest neighbors of vi . The resulting graph is called the mutual k‐nearest neighbor graph. In both cases, after connecting the appropriate vertices, we weight the edges by the similarity of their endpoints.

      The fully connected graph: Here, we simply connect all points with positive similarity with each other, and we weight all edges by sij . As the graph should represent the local neighborhood relationships, this construction is useful only if the similarity function itself models local neighborhoods. An example of such a similarity function is the Gaussian similarity function s(xi, xj) = exp (−‖xi − xj2/(2σ2)), where the parameter σ controls the width of the neighborhoods. This parameter plays a similar role as the parameter ε in the case of the ε‐neighborhood graph.

      In the following, we always assume that G is an undirected, weighted graph with weight matrix W, where wij = wji ≥ 0. When using the eigenvectors of a matrix, we will not necessarily assume that they are normalized. For example, the constant vector 1 and a multiple a1 for some a ≠ 0 will be considered the same eigenvectors. Eigenvalues will always be ordered in ascending order, respecting multiplicities. By “the first k eigenvectors” we refer to the eigenvectors corresponding to the k smallest eigenvalues.

      The unnormalized graph Laplacian is defined as L = DW.

      The following proposition summarizes the most important facts needed for spectral clustering.

      (Properties of L) The matrix L satisfies the following properties:

      1 For every vector f ∈ ℝn we have

      2 L is symmetric and positive semidefinite.

      3 The smallest eigenvalue of L is 0, and the corresponding eigenvector is the constant one vector 1.

      4 L has n non‐negative, real‐valued eigenvalues 0 = λ1 ≤ λ2≤… ≤λn.

      Proof:

      Part (1): By the definition of di,

equation

      Part (2): The symmetry of L follows directly from the symmetry of W and D. The positive semidefiniteness is a direct consequence of Part (1), which shows that fLf ≥ 0 for all fn.

      Part (3): Self‐evident.

      Part (4) is a direct consequence of Parts (1)–(3).

      The normalized graph Laplacians: There are two matrices that are called normalized graph Laplacians in the literature. Both matrices are closely related to each other and are defined as

equation

      We denote the first matrix by Lsym as it is a

Скачать книгу