Electronics in Advanced Research Industries. Alessandro Massaro
Чтение книги онлайн.
Читать онлайн книгу Electronics in Advanced Research Industries - Alessandro Massaro страница 26
Figure 1.17 (a) Simple ANN. (b) DL neural network.
Figure 1.18 (a) Feedback system minimizing calculation error in the training model. (b) Neural network model implementing unit step function.
The pseudocode of the ANN training process is as follows:
1. Train_ANN (fi, wi, oj) 2. For epochs = 1 to N Do 3. While (j ≤ m) Do 4. Randomly initialize wi = { w1 , w2 …, wn}; 5. Input oj = { o1 , o2 ,…, om} in the input layer forward propagate (fi· wi) though layers until is obtained the predicted result y; 6. Compute the error e = y - y2; 7. Back propagate e from the right to the left of the ANN network through layers; 8. Update wi; 9. End While 10. End For
The pseudocode highlights that there are two mechanisms in the ANN network: the forward propagation of the estimation of the predicted output y, and the back propagation of the error function as sketched in Figure 1.18b. The output is estimated by considering the summation of the input contributions and is defined as:
(1.10)
where f is the activation function. Some examples of activation functions are plotted in Figure 1.19, where the analytical forms are:
(1.11)
(1.12)
(1.13)
(1.14)
(1.15)
Figure 1.19 Basic mathematical functions defining activation functions.
Other mathematical activation functions are the following [68]:
(1.16)
(1.17)
(1.18)
(1.19)
(1.20)
(1.21)
(1.22)
(1.23)
(1.24)
(1.25)
(1.26)
(1.27)
(1.28)
(1.29)
(1.30)
(1.31)
The activation function represents a basic research element of considerable importance. The correct choice of the activation function defines the best implementation of the logic defining the outputs. The analytical model must therefore be appropriately weighted by the various variables and must be “calibrated” for the specific case study. Another important aspect is the ability of the activation function to self‐adapt [69] to the specific case study providing a certain flexibility [70]. Of particular interest is the possibility to consider a combination of activation functions (activation ensemble [71]). The approach to follow is therefore to define a flexible and modular activation function as is the case for the adaptive spline activation function [72].
Concerning the training models, the full dataset of the neural network is divided into a training set, validation set, and test set (Figure