官术网_书友最值得收藏!

Overcoming over-fitting using regularization

In the previous section, we established that a high weight magnitude is one of the reasons for over-fitting. In this section, we will look into ways to get around the problem of over-fitting, such as penalizing for high weight magnitude values.

Regularization gives a penalty for having a high magnitude of weights in model. L1 and L2 regularizations are among the most commonly used regularization techniques and work as follows:

L2 regularization minimizes the weighted sum of squares of weights at the specified layer of the neural network, in addition to minimizing the loss function (which is the sum of squared loss in the following formula):

Where  is the weightage associated with the regularization term and is a hyperparameter that needs to be tuned, y is the predicted value of , and  is the weight values across all the layers of the model.

L1 regularization minimizes the weighted sum of absolute values of weights at the specified layer of the neural network in addition to minimizing the loss function (which is the sum of the squared loss in the following formula):

.

This way, we ensure that weights do not get customized for extreme cases in the training dataset only (and thus, not generalizing on the test data).

主站蜘蛛池模板: 舟曲县| 临高县| 贺州市| 波密县| 云林县| 陇南市| 科尔| 曲靖市| 聂拉木县| 汉寿县| 定远县| 邢台市| 江陵县| 德清县| 桑日县| 眉山市| 方城县| 郸城县| 保定市| 崇阳县| 玛曲县| 双辽市| 炎陵县| 普兰店市| 新源县| 阿拉善右旗| 卓资县| 金秀| 偃师市| 竹北市| 浙江省| 榕江县| 陵川县| 新乐市| 永泰县| 金寨县| 武陟县| 宁海县| 荆州市| 嵊州市| 开鲁县|