官术网_书友最值得收藏!

Activation functions

The activation function in each artificial neuron decides whether the incoming signals have reached the threshold and should output signals for the next level. It is crucial to set up the right activation function because of the gradient vanishing issue, which we will talk about later.

Another important feature of an activation function is that it should be differentiable. The network learns from the errors that are calculated at the output layer. A differentiable activation function is needed to perform backpropagation optimization while propagating backwards in the network to compute gradients of error (loss) with respect to weights, and then optimize weights accordingly, using gradient descent or any other optimization technique to reduce the error.

The following table lists a few common activation functions. We will dive into them a bit deeper, talk about the differences between them, and explain how to choose the right activation function:

主站蜘蛛池模板: 德昌县| 云浮市| 饶平县| 同心县| 余江县| 海口市| 昂仁县| 湾仔区| 甘德县| 疏附县| 庆安县| 长葛市| 清原| 富蕴县| 清原| 红河县| 和田市| 霍城县| 乐陵市| 新乡市| 黑河市| 嘉善县| 淮安市| 车致| 潞西市| 陆川县| 岳西县| 吴桥县| 吴桥县| 弥勒县| 上杭县| 乌拉特前旗| 拉孜县| 德钦县| 竹溪县| 双柏县| 饶阳县| 萍乡市| 资阳市| 右玉县| 义马市|