官术网_书友最值得收藏!

  • Deep Learning Essentials
  • Wei Di Anurag Bhardwaj Jianing Wei
  • 126字
  • 2021-06-30 19:17:52

Leaky ReLU and maxout

A Leaky ReLU will have a small slope α on the negative side, such as 0.01. The slope α can also be made into a parameter of each neuron, such as in PReLU neurons (P stands for parametric). The problem with this activation function is the inconsistency of the effectiveness of such modifications to various problems.

Maxout is another attempt to solve the dead neuron problem in ReLU. It takes the form . From this form, we can see that both ReLU and leaky ReLU are just special cases of this form, that is, for ReLU, it's . Although it benefits from linearity and having no saturation, it has doubled the number of parameters for every single neuron.

主站蜘蛛池模板: 伊金霍洛旗| 深泽县| 宁陕县| 涞水县| 疏勒县| 方山县| 鞍山市| 轮台县| 资源县| 沙洋县| 黑龙江省| 怀仁县| 翁源县| 缙云县| 平罗县| 汨罗市| 邓州市| 扎赉特旗| 陆河县| 达州市| 武安市| 枝江市| 洱源县| 武宁县| 汨罗市| 长宁县| 互助| 高密市| 普兰县| 南皮县| 隆子县| 彩票| 本溪市| 唐河县| 珲春市| 平远县| 金溪县| 龙川县| 衡南县| 城固县| 天门市|