- Deep Learning with R for Beginners
- Mark Hodnett Joshua F. Wiley Yuxi (Hayden) Liu Pablo Maldonado
- 319字
- 2021-06-24 14:30:44
Activation functions
The activation function determines the mapping between input and a hidden layer. It defines the functional form for how a neuron gets activated. For example, a linear activation function could be defined as: f(x) = x, in which case the value for the neuron would be the raw input, x. A linear activation function is shown in the top panel of Figure 4.2. Linear activation functions are rarely used because in practice deep learning models would find it difficult to learn non-linear functional forms using linear activation functions. In previous chapters, we used the hyperbolic tangent as an activation function, namely f(x) = tanh(x). Hyperbolic tangent can work well in some cases, but a potential limitation is that at either low or high values, it saturates, as shown in the middle panel of the figure 4.2.
Perhaps the most popular activation function currently, and a good first choice (Nair, V., and Hinton, G. E. (2010)), is known as a rectifier. There are different kinds of rectifiers, but the most common is defined by the f(x) = max(0, x) function, which is known as relu. The relu activation is flat below zero and linear above zero; an example is shown in Figure 4.2.
The final type of activation function we will discuss is maxout (Goodfellow, Warde--Farley, Mirza, Courville, and Bengio (2013)). A maxout unit takes the maximum value of its input, although as usual, this is after weighting so it is not the case that the input variable with the highest value will always win. Maxout activation functions seem to work particularly well with dropout.
The relu activation is the most commonly-used activation function and it is the default option for the deep learning models in the rest of this book. The following graphs for some of the activation functions we have discussed:

- 數據庫基礎教程(SQL Server平臺)
- 數據要素安全流通
- PyTorch深度學習實戰:從新手小白到數據科學家
- 數據分析實戰:基于EXCEL和SPSS系列工具的實踐
- Modern Programming: Object Oriented Programming and Best Practices
- Python金融大數據分析(第2版)
- 卷積神經網絡的Python實現
- Oracle RAC 11g實戰指南
- 云計算服務保障體系
- Lean Mobile App Development
- 軟件成本度量國家標準實施指南:理論、方法與實踐
- 白話大數據與機器學習
- Power BI商業數據分析完全自學教程
- Construct 2 Game Development by Example
- TextMate How-to