I'm implementing a neural network with backward propagation. The weights are initialized to (-0.5, 0.5). However, after the first time the inputs are sent forward and errors and propagated back, the weights are increased to be around 1000, sometimes even 2000. (between the input layer and hidden layer)
The topology of the network consists of 3 layers: 1 input layer, 1 hidden layer, and 1 output layer. The input layer has 95 nodes, hidden layer has 3 nodes and output layer has 2 nodes. The training data set has 40,000 entries, they are normalized with their z-scores.
After seeing such high numbers I doubted my implementation, but then again, with learning rate set to 1 on the first propagation, if each entry has around (output*error)=0.25, which is reasonable, then a weight change to be about 1000 seems plausible.
Anyway, are weights in a neural network suppose to be this high?
Gracias