- Advanced Machine Learning with R
- Cory Lesmeister Dr. Sunil Kumar Chinnamgari
- 240字
- 2021-06-24 14:24:39
LASSO
LASSO applies the L1-norm instead of the L2-norm as in ridge regression, which is the sum of the absolute value of the feature weights and so minimizes RSS + λ(sum |Bj|). This shrinkage penalty will indeed force a feature weight to zero. This is a clear advantage over ridge regression, as it may improve the model interpretability.
The mathematics behind the reason that the L1-norm allows the weights/coefficients to become zero is beyond the scope of this book (refer to Tibsharini, 1996 for further details).
If LASSO is so great, then ridge regression must be obsolete in machine learning. Not so fast! In a situation of high collinearity or high pairwise correlations, LASSO may force a predictive feature to zero, hence you can lose the predictive ability; that is, if both feature A and B should be in your model, LASSO may shrink one of their coefficients to zero. The following quote sums up this issue nicely:
There is the possibility of achieving the best of both worlds and that leads us to the next topic, elastic net.
- Learning SQL Server Reporting Services 2012
- 計算機組裝與系統配置
- 計算機組裝·維護與故障排除
- 精選單片機設計與制作30例(第2版)
- 硬件產品經理成長手記(全彩)
- 硬件產品經理手冊:手把手構建智能硬件產品
- Learning Stencyl 3.x Game Development Beginner's Guide
- 計算機組裝與維修技術
- Apple Motion 5 Cookbook
- 電腦高級維修及故障排除實戰
- 無蘋果不生活:OS X Mountain Lion 隨身寶典
- FreeSWITCH Cookbook
- Arduino項目開發:智能生活
- FPGA實驗實訓教程
- 微服務實戰(Dubbox +Spring Boot+Docker)