官术网_书友最值得收藏!

Overcoming the limitations of deep learning

These two possible problems can be overcome by:

  • Minimizing the use of the sigmoid and tanh activation functions
  • Using a momentum-based stochastic gradient descent
  • Proper initialization of weights and biases, such as xavier initialization
  • Regularization (add regularization loss along with data loss and minimize that)
For more detail, along with mathematical representations of the vanishing and exploding gradient, you can read this article: Intelligent Signals : Unstable Deep Learning. Why and How to solve them ?
主站蜘蛛池模板: 富顺县| 舞阳县| 商丘市| 修水县| 渑池县| 皋兰县| 丹东市| 浦北县| 临澧县| 延庆县| 大足县| 清丰县| 固安县| 来安县| 乌兰县| 富阳市| 永修县| 库车县| 连州市| 郁南县| 佛冈县| 杂多县| 武鸣县| 卓尼县| 凯里市| 胶州市| 南木林县| 富顺县| 和顺县| 丰县| 北碚区| 大竹县| 襄城县| 阿瓦提县| 论坛| 清新县| 普洱| 肇东市| 汝城县| 孙吴县| 金平|