官术网_书友最值得收藏!

Using regularization to overcome overfitting

In the previous chapter, we saw the diminishing returns from further training iterations on neural networks in terms of their predictive ability on holdout or test data (that is, data not used to train the model). This is because complex models may memorize some of the noise in the data rather than learning the general patterns. These models then perform much worse when predicting new data. There are some methods we can apply to make our model generalize, that is, fit the overall patterns. These are called regularization and aim to reduce testing errors so that the model performs well on new data.

The most common regularization technique used in deep learning is dropout. However, we will also discuss two other regularization techniques that have a basis in regression and deep learning. These two regularization techniques are L1 penalty, which is also known as Lasso, and L2 penalty, which is also known as Ridge.

主站蜘蛛池模板: 姜堰市| 五莲县| 洪洞县| 乐昌市| 广南县| 星座| 瑞金市| 东莞市| 酒泉市| 永靖县| 舟曲县| 江安县| 武陟县| 阜宁县| 库车县| 连平县| 吉安市| 香港 | 南城县| 沁源县| 聂拉木县| 湟源县| 博白县| 东阳市| 南开区| 兴国县| 翁源县| 北安市| 广河县| 兴隆县| 竹溪县| 曲阳县| 新宁县| 石楼县| 临沭县| 惠来县| 长沙市| 宜州市| 鄯善县| 嘉善县| 河南省|