官术网_书友最值得收藏!

Activation functions

The activation function in each artificial neuron decides whether the incoming signals have reached the threshold and should output signals for the next level. It is crucial to set up the right activation function because of the gradient vanishing issue, which we will talk about later.

Another important feature of an activation function is that it should be differentiable. The network learns from the errors that are calculated at the output layer. A differentiable activation function is needed to perform backpropagation optimization while propagating backwards in the network to compute gradients of error (loss) with respect to weights, and then optimize weights accordingly, using gradient descent or any other optimization technique to reduce the error.

The following table lists a few common activation functions. We will dive into them a bit deeper, talk about the differences between them, and explain how to choose the right activation function:

主站蜘蛛池模板: 西安市| 平南县| 呼玛县| 尉氏县| 丹寨县| 固镇县| 长子县| 水富县| 湘潭县| 辽宁省| 康定县| 和顺县| 武定县| 安仁县| 金湖县| 梨树县| 邹城市| 盐池县| 庆云县| 铜梁县| 湘乡市| 衡阳市| 深圳市| 潼关县| 定结县| 张北县| 伊金霍洛旗| 博罗县| 阳春市| 安阳市| 嘉鱼县| 高州市| 泌阳县| 襄城县| 邵武市| 隆安县| 青岛市| 盐城市| 长阳| 涪陵区| 抚远县|