Math models designed to imitate the brain.
ANN are commonly used to classify or to extract information and patterns from (RS) images (often missed when using traditional statistical tools)
A typical feedforward ANN consists of three or more inter-connected layers of nodes ― an input layer, one or more hidden intermediate layers (often just 1), and an output layer (Figure 8‑15). The arrows indicate the direction of information flow, feeding information forward from input to output. Note that there may be any number of nodes at each level of the network, and not all nodes need to be connected to every node in the next layer. For example, in Figure 8‑15 hidden node H2 only receives input from input nodes I1 and I3, and only provides output to node O1. This arrangement can be seen as a directed graph, or as a rather complex-looking function mapping.
The connections between the input layer and hidden layer can be described using a weight matrix, W, where the row/column entries wij are positive or negative real-valued weights, or 0 if no connection exists. Likewise, the connections between the hidden layer and the output layer can also be viewed as a weight matrix, Z say, again consisting of a set of weights, zjk. Positive weights in each case imply a reinforcement process associated with the source node or input, whilst negative weights correspond to inhibition.
Terminologies in ANN
1. Neurons (Nodes/Perceptrons):
2. Layers:
3. Weights and Biases:
4. Activation Function:
1. Feedforward Neural Network:
2. Recurrent Neural Network (RNN):
The dataset is normally split into 3:
-Training -> presented to the network
-Testing -> early stopping to minimize the risk of “overfitting”
-Validation -> to get some kind of error measurement (RMSE)
How many layers and nodes to use in ANN?