Computational Statistics in Data Science. Группа авторов

Чтение книги онлайн.

Читать онлайн книгу Computational Statistics in Data Science - Группа авторов страница 41

Computational Statistics in Data Science - Группа авторов

Скачать книгу

6 7 10 11 12 13 15 16 2 1 2 6 7 8 11 12 13 14 16 3 1 2 3 7 8 9 12 14 15 16 4 2 3 4 7 8 9 10 13 15 16 5 3 4 5 8 9 10 11 13 14 16 6 4 5 6 9 10 11 12 14 15 16

      The row names are indices of input matrices, and the second column shows indices of output matrices that are connected to the corresponding input matrix. There are 60 connections in total, meaning 60 different kernel matrices.

      The fourth layer (S4) is a Max Pooling layer that produces 16 feature matrices with size 5 times 5. The kernel size of this layer is 2 times 2, and the stride is 2. Therefore, each of the input matrices is reduced to 5 times 5. The fifth layer (C5) is the last convolutional layer in LeNet‐5. The 16 input matrices are fully connected to 120 output matrices. Since both the input matrices and kernel matrices are of size 5 times 5, the output matrices are of size 1 times 1. Therefore, the output is actually a 120‐dimensional vector. Each number in the vector is computed by applying 16 different kernel matrices on the 16 different input matrices and then combining the results and bias.

      The sixth and seventh layers are fully connected layers, which are introduced in the previous section. In the sixth layer (S6), 120 input neurons are fully connected to 84 output neurons. In the last layer, 84 neurons are fully connected to 10 output neurons, where the 10‐dimensional output vector contains predict scores of each class. For the classification task, cross‐entropy loss between the model output and the label is usually used to train the model.

      There are many other architectures of CNNs, such as AlexNet [10], VGG [11], and ResNet [12]. These neural networks demonstrated state‐of‐the‐art performances on many machine learning tasks, such as image classification, object detection, and speech processing.

      5.1 Introduction

      5.2 Objective Function

      Autoencoder is first introduced in Rumelhart et al. [16] as a model with the main goal of learning a compressed representation of the input in an unsupervised way. We are essentially creating a network that attempts to reconstruct inputs by learning the identity function. To do so, an autoencoder can be divided into two parts, bold-italic upper E colon double-struck upper R Superscript n Baseline right-arrow double-struck upper R Superscript p (encoder) and

Скачать книгу