官术网_书友最值得收藏!

The vanishing gradient problem

The vanishing gradient problem is one of the problems associated with the training of artificial neural networks when the neurons present in the early layers are not able to learn because the gradients that train the weights shrink down to zero. This happens due to the greater depth of neural network, along with activation functions with derivatives resulting in low value.

Try the following steps:

  1. Create one hidden layer neural network
  2. Add more hidden layers, one by one

We observe the gradient with regards to all the nodes, and find that the gradient values get relatively smaller when we move from the later layers to the early layers. This condition worsens with the further addition of layers. This shows that the early layer neurons are learning slowly compared to the later layer neurons. This condition is called the vanishing gradient problem.

主站蜘蛛池模板: 万源市| 财经| 黄山市| 龙游县| 昌平区| 康马县| 仁怀市| 克山县| 从化市| 阿勒泰市| 巫山县| 皮山县| 定结县| 大荔县| 六盘水市| 平舆县| 大厂| 寿光市| 京山县| 开化县| 枣庄市| 竹北市| 汶上县| 五家渠市| 湖州市| 河西区| 卢龙县| 嘉义县| 新干县| 丰台区| 峡江县| 塘沽区| 河源市| 元谋县| 眉山市| 华宁县| 灌云县| 柳江县| 迁西县| 集贤县| 遂昌县|