- Reinforcement Learning with TensorFlow
- Sayon Dutta
- 214字
- 2021-08-27 18:51:52
The rectified linear unit function
The rectified linear unit, better known as ReLU, is the most widely used activation function:

The ReLU function has the advantage of being non linear. Thus, backpropagation is easy and can therefore stack multiple hidden layers activated by the ReLU function, where for x<=0, the function f(x) = 0 and for x>0, f(x)=x.

The main advantage of the ReLU function over other activation functions is that it does not activate all the neurons at the same time. This can be observed from the preceding graph of the ReLU function, where we see that if the input is negative it outputs zero and the neuron does not activate. This results in a sparse network, and fast and easy computation.

Looking at the preceding gradients graph of ReLU preceding, we can see the negative side of the graph shows a constant zero. Therefore, activations falling in that region will have zero gradients and therefore, weights will not get updated. This leads to inactivity of the nodes/neurons as they will not learn. To overcome this problem, we have Leaky ReLUs, which modify the function as:


This prevents the gradient from becoming zero in the negative side and the weight training continues, but slowly, owing to the low value of .
- 后稀缺:自動(dòng)化與未來工作
- 大數(shù)據(jù)管理系統(tǒng)
- 傳感器技術(shù)實(shí)驗(yàn)教程
- WordPress Theme Development Beginner's Guide(Third Edition)
- Enterprise PowerShell Scripting Bootcamp
- 基于Xilinx ISE的FPAG/CPLD設(shè)計(jì)與應(yīng)用
- Dreamweaver CS6精彩網(wǎng)頁制作與網(wǎng)站建設(shè)
- Photoshop行業(yè)應(yīng)用基礎(chǔ)
- INSTANT Puppet 3 Starter
- 在實(shí)戰(zhàn)中成長:C++開發(fā)之路
- Hands-On Dashboard Development with QlikView
- 筆記本電腦電路分析與故障診斷
- 機(jī)器學(xué)習(xí)案例分析(基于Python語言)
- 穿越計(jì)算機(jī)的迷霧
- 計(jì)算機(jī)硬件技術(shù)基礎(chǔ)(第2版)