Artificial
Neural Networks (ANNs)
 19 different neural network architectures are applied.
 The networks differed from each other in the number
of hidden layers and in the number of neurons in a
hidden layer.
 Every node in a layer (input, hidden and output)
in these feedforward networks (multilayer preceptors)
is fully connected (by adjustable weights) to every
node in a subsequent layer.
 Thus, the total input to a node of the hidden layer
can be written as follows:
 W are the adjustable connection weights.
 Constant terms, so called biases are given by introducing
an extra input to each unit.
 Transfer functions are of the logsig type, thus
a unit has a realvalued output:
