0 votes
in Machine Learning by
Why is ReLU better and more often used than Sigmoid in Neural Networks?

1 Answer

0 votes
by

Imagine a network with random initialized weights ( or normalised ) and almost 50% of the network yields 0 activation because of the characteristic of ReLu ( output 0 for negative values of x ). This means a fewer neurons are firing ( sparse activation ) and the network is lighter. 

Related questions

0 votes
asked Jun 29, 2023 in DevOps by SakshiSharma
0 votes
asked Aug 4, 2021 in Ingression Deep Learning by Robindeniel
...