convolution - NAN when I use ReLU activation function in convolutional neural network Lenet-5 -


i did programmed convolution neural network lenet-5. made modifications:

  1. i replaced activation function of output neurons in last layer rbf softmah.
  2. subsampling layers maxpooling layers.
  3. learning method backpropagation

as result, network working correctly.

after tried replace sigmoid output of each neuron in feature maps relu (rectifier linear unit). result, network began learn faster, if not choose low speed, nan value.

for small set of input data, simpler use lower speed of learning. when comes more 1,000 examples, network working, in end nan again.

why there nan when using relu? lenet architecture not relu?


Comments

Popular posts from this blog

javascript - Chart.js (Radar Chart) different scaleLineColor for each scaleLine -

apache - Error with PHP mail(): Multiple or malformed newlines found in additional_header -

java - Android – MapFragment overlay button shadow, just like MyLocation button -