官术网_书友最值得收藏!

Overcoming over-fitting using regularization

In the previous section, we established that a high weight magnitude is one of the reasons for over-fitting. In this section, we will look into ways to get around the problem of over-fitting, such as penalizing for high weight magnitude values.

Regularization gives a penalty for having a high magnitude of weights in model. L1 and L2 regularizations are among the most commonly used regularization techniques and work as follows:

L2 regularization minimizes the weighted sum of squares of weights at the specified layer of the neural network, in addition to minimizing the loss function (which is the sum of squared loss in the following formula):

Where  is the weightage associated with the regularization term and is a hyperparameter that needs to be tuned, y is the predicted value of , and  is the weight values across all the layers of the model.

L1 regularization minimizes the weighted sum of absolute values of weights at the specified layer of the neural network in addition to minimizing the loss function (which is the sum of the squared loss in the following formula):

.

This way, we ensure that weights do not get customized for extreme cases in the training dataset only (and thus, not generalizing on the test data).

主站蜘蛛池模板: 启东市| 周宁县| 德昌县| 湘乡市| 九龙坡区| 临湘市| 南宫市| 铜川市| 津市市| 尼玛县| 冀州市| 洪泽县| 新干县| 凤庆县| 洛浦县| 平泉县| 东至县| 山阴县| 友谊县| 伊川县| 罗甸县| 绥德县| 札达县| 木里| 泗洪县| 武宁县| 邳州市| 辽阳市| 吉木萨尔县| 黄冈市| 宁晋县| 江门市| 惠水县| 诏安县| 新余市| 宝应县| 西藏| 东城区| 交口县| 汉阴县| 平远县|