I am trying to understand the error backpropagation algorithm to be used in Artificial Neural Networks.
As far as I understand, each neuron has as an output the result of a sigmoid function, which takes as parameter the sum of the product of each weight/input pair plus a "bias" value.
Once the total error of the network is calculated, the derivative of the error with respect to the various weights can be used to find "local minimum" of the error function. These should be the weights where the error is minimum.
However, points where the derivative is zero could also in theory be local maximums... How can I solve this problem?
Cheers :)