gradient vanishing problem in Back propagation of Artificial Neural Network:
- in a back propagation,we use Relu (Rectified linear unit) activation function more then Sigmoid function but why?
-before I explain,i hope you know the derivation of sigmoid function is ranging between 0 to 0.25.
so in a multi hidden layer neural n...
anisurrahmansblog.hashnode.dev1 min read