官术网_书友最值得收藏!

Ridge regression (L2)

Due to ridge regression, we need to make some changes to the loss function. The original loss function gets added by a shrinkage component:

Now, this modified loss function needs to be minimized to adjust the estimates or coefficients. Here, the lambda is tuning the parameter that regularizes the loss function. That is, it decides how much it should penalize the flexibility of the model. The flexibility of the model is dependent on the coefficients. If the coefficients of the model go up, the flexibility also goes up, which isn't a good sign for our model. Likewise, as the coefficients go down, the flexibility is restricted and the model starts to perform better. The shrinkage of each estimated parameter makes the model better here, and this is what ridge regression does. When lambda keeps going higher and higher, that is, λ → ∞, the penalty component rises, and the estimates start shrinking. However, when λ → 0, the penalty component decreases and starts to become an ordinary least square (OLS) for estimating unknown parameters in a linear regression.

主站蜘蛛池模板: 沈丘县| 新蔡县| 塘沽区| 奉化市| 鹿邑县| 塘沽区| 乐亭县| 泉州市| 新密市| 安陆市| 沈丘县| 广东省| 石泉县| 贵南县| 东光县| 博湖县| 台北市| 灵寿县| 明水县| 乌什县| 商洛市| 芦山县| 和平县| 广丰县| 东乌珠穆沁旗| 临海市| 司法| 宁南县| 大田县| 农安县| 扎囊县| 林甸县| 大英县| 桐庐县| 绥宁县| 临漳县| 平乐县| 宝山区| 永川市| 万州区| 博客|