官术网_书友最值得收藏!

Training neural networks with backpropagation

Calculating the activation of a neuron, the forward part, or what we call feed-forward propagation, is quite straightforward to process. The complexity we encounter now is training the errors back through the network. When we train the network now, we start at the last output layer and determine the total error, just as we did with a single perceptron, but now we need to sum up all errors across the output layer. Then we need to use this value to backpropagate the error back through the network, updating each of the weights based on their contribution to the total error. Understanding the contribution of a single weight in a network with thousands or millions of weights could be quite complicated, except thankfully for the help of differentiation and the chain rule. Before we get to the complicated math, we first need to discuss the Cost function and how we calculate errors in the next section.

While the math of backpropagation is complicated and may be intimidating, at some point, you will want or need to understand it well. However, for the purposes of this book, you can omit or just revisit this section as needed. All the networks we develop in later chapters will automatically handle backpropagation for us. Of course, you can't run away from the math either; it is everywhere in deep learning.
主站蜘蛛池模板: 敦化市| 慈溪市| 满城县| 林口县| 衡山县| 库伦旗| 台东市| 明溪县| 九龙坡区| 石屏县| 湟源县| 白城市| 盘锦市| 龙口市| 古丈县| 抚宁县| 邹平县| 滦南县| 太白县| 华坪县| 浮山县| 安溪县| 玛纳斯县| 保康县| 贺州市| 商水县| 瓮安县| 岳普湖县| 镇原县| 桃源县| 鲜城| 东兰县| 宜丰县| 斗六市| 怀宁县| 亚东县| 崇义县| 响水县| 会理县| 德阳市| 文登市|