官术网_书友最值得收藏!

Gradient descent

Gradient descent is an optimization technique that utilizes the gradients computed from backpropagation to update the weights and biases, moving towards the goal of minimizing the loss. As shown in the following diagram, the cost (or loss) function is minimized by adjusting the weights, along the slope or gradient of the function:

For a simple perceptron, this cost function is linear, with respect to the weights. But for deep neural networks, the cost function is most often high-dimensional and non-linear. As gradient descent has to traverse paths along all of the different dimensions, it may be difficult to arrive at the global minimum in an acceptable time. To avoid this problem and train faster, neural networks normally employ stochastic gradient descent, which is explained next.

主站蜘蛛池模板: 新闻| 秦皇岛市| 禄劝| 循化| 进贤县| 仪陇县| 县级市| 嘉黎县| 宜城市| 高青县| 赤峰市| 台南市| 漳浦县| 大名县| 通海县| 饶河县| 读书| 临颍县| 历史| 休宁县| 西乡县| 南昌县| 武汉市| 密云县| 全椒县| 江北区| 尉犁县| 上饶县| 赤峰市| 珲春市| 卢氏县| 靖西县| 镇安县| 曲沃县| 武川县| 任丘市| 舞阳县| 石城县| 衡阳市| 宝鸡市| 米脂县|