官术网_书友最值得收藏!

Gradient descent

Up until now, we have covered the different kind of neurons based on the activation functions that are used. We have covered the ways to quantify inaccuracy in the output of a neuron using cost functions. Now, we need a mechanism to take that inaccuracy and remedy it.

The mechanism through which the network can learn to output values closer to the expected or desired output is called gradient descent. Gradient descent is a common approach in machine learning for finding the lowest cost possible.

To understand gradient descent, let's use the single neuron equation we have been using so far:

Here, the following applies:

  • x is the input
  • w is the weight of the input
  • b is the bias of the input

Gradient descent can be represented as follows:

Initially, the neuron starts by assigning random values for w and b. From that point onward, the neuron needs to adjust the values of w and b so that it lowers or decreases the error or cost (cross entropy).

Taking the derivative of the cross entropy (cost function) results in a step-by-step change in w and b in the direction of the lowest cost possible. In other words, gradient descent tries to find the finest line between the network output and expected output.

The weights are adjusted based on a parameter called the learning rate. The learning rate is the value that is adjusted to the weight of the neuron to get an output closer to the expected output.

Keep in mind that here, we have used only a single parameter; this is only to make things easier to comprehend. In reality, there are thousands upon millions of parameters that are taken into consideration to lower the cost.

主站蜘蛛池模板: 车致| 定安县| 禹州市| 简阳市| 卓资县| 宽甸| 高尔夫| 分宜县| 上杭县| 华容县| 根河市| 东明县| 青铜峡市| 仙游县| 乳源| 齐齐哈尔市| 子长县| 中西区| 桃源县| 赫章县| 湾仔区| 保亭| 滦平县| 保定市| 通江县| 攀枝花市| 嘉义市| 双桥区| 虞城县| 连云港市| 新郑市| 曲周县| 延边| 蓬溪县| 丹东市| 平塘县| 绍兴县| 福海县| 新津县| 普安县| 齐齐哈尔市|