官术网_书友最值得收藏!

Limitations of deep learning

Deep neural networks are black boxes of weights and biases trained over a large amount of data to find hidden patterns through inner representations; it would be impossible for humans, and even if it were possible, then scalability would be an issue. Every neural probably has a different weight. Thus, they will have different gradients.

Training happens during backpropagation. Thus, the direction of training is always from the later layers (output/right side) to the early layers (input/left side). This results in later layers learning very well as compared to the early layers. The deeper the network gets, the more the condition deteriorates. This give rise to two possible problems associated with deep learning, which are:

  • The vanishing gradient problem
  • The exploding gradient problem
主站蜘蛛池模板: 科技| 神农架林区| 顺昌县| 铜陵市| 延川县| 平谷区| 蒙山县| 南岸区| 资溪县| 万安县| 平邑县| 扬中市| 金湖县| 民勤县| 九寨沟县| 银川市| 西昌市| 江北区| 巴中市| 清丰县| 东乡县| 湖南省| 迁安市| 白玉县| 建湖县| 布拖县| 岢岚县| 嵩明县| 隆化县| 普定县| 大田县| 和田县| 栖霞市| 安国市| 凤冈县| 乌什县| 成都市| 留坝县| 台安县| 都昌县| 湛江市|