官术网_书友最值得收藏!

The vanishing gradient problem

The vanishing gradient problem is one of the problems associated with the training of artificial neural networks when the neurons present in the early layers are not able to learn because the gradients that train the weights shrink down to zero. This happens due to the greater depth of neural network, along with activation functions with derivatives resulting in low value.

Try the following steps:

  1. Create one hidden layer neural network
  2. Add more hidden layers, one by one

We observe the gradient with regards to all the nodes, and find that the gradient values get relatively smaller when we move from the later layers to the early layers. This condition worsens with the further addition of layers. This shows that the early layer neurons are learning slowly compared to the later layer neurons. This condition is called the vanishing gradient problem.

主站蜘蛛池模板: 大港区| 道孚县| 泰安市| 德清县| 汉源县| 枣阳市| 大厂| 哈密市| 滨州市| 冀州市| 建湖县| 桐柏县| 盈江县| 大安市| 阳谷县| 五指山市| 天祝| 平湖市| 松原市| 怀仁县| 古交市| 阿克苏市| 道孚县| 玉溪市| 南木林县| 玛纳斯县| 芒康县| 临邑县| 大丰市| 永寿县| 庐江县| 麻阳| 贡嘎县| 鹤峰县| 讷河市| 定边县| 孟津县| 西昌市| 卫辉市| 扎囊县| 霍林郭勒市|