官术网_书友最值得收藏!

Ridge regression (L2)

Due to ridge regression, we need to make some changes to the loss function. The original loss function gets added by a shrinkage component:

Now, this modified loss function needs to be minimized to adjust the estimates or coefficients. Here, the lambda is tuning the parameter that regularizes the loss function. That is, it decides how much it should penalize the flexibility of the model. The flexibility of the model is dependent on the coefficients. If the coefficients of the model go up, the flexibility also goes up, which isn't a good sign for our model. Likewise, as the coefficients go down, the flexibility is restricted and the model starts to perform better. The shrinkage of each estimated parameter makes the model better here, and this is what ridge regression does. When lambda keeps going higher and higher, that is, λ → ∞, the penalty component rises, and the estimates start shrinking. However, when λ → 0, the penalty component decreases and starts to become an ordinary least square (OLS) for estimating unknown parameters in a linear regression.

主站蜘蛛池模板: 兴山县| 电白县| 龙胜| 裕民县| 西乌| 乌海市| 洮南市| 健康| 江达县| 长寿区| 万州区| 南陵县| 桦川县| 和平区| 同德县| 囊谦县| 广东省| 镇雄县| 德阳市| 萍乡市| 大庆市| 青海省| 新余市| 来凤县| 武邑县| 沁阳市| 青河县| 正镶白旗| 博乐市| 吉安县| 博爱县| 会昌县| 石景山区| 鸡泽县| 广德县| 广东省| 汨罗市| 沿河| 长兴县| 江安县| 合阳县|