官术网_书友最值得收藏!

Using regularization to overcome overfitting

In the previous chapter, we saw the diminishing returns from further training iterations on neural networks in terms of their predictive ability on holdout or test data (that is, data not used to train the model). This is because complex models may memorize some of the noise in the data rather than learning the general patterns. These models then perform much worse when predicting new data. There are some methods we can apply to make our model generalize, that is, fit the overall patterns. These are called regularization and aim to reduce testing errors so that the model performs well on new data.

The most common regularization technique used in deep learning is dropout. However, we will also discuss two other regularization techniques that have a basis in regression and deep learning. These two regularization techniques are L1 penalty, which is also known as Lasso, and L2 penalty, which is also known as Ridge.

主站蜘蛛池模板: 淮南市| 富阳市| 永寿县| 措勤县| 阿勒泰市| 永济市| 五原县| 高碑店市| 青岛市| 万安县| 金华市| 宁国市| 鄂托克前旗| 江门市| 县级市| 黄骅市| 务川| 祁连县| 嘉黎县| 阳朔县| 武夷山市| 台安县| 凌海市| 江达县| 安阳县| 呼伦贝尔市| 凤城市| 南川市| 会东县| 河南省| 轮台县| 宝丰县| 伽师县| 永安市| 松桃| 莒南县| 迭部县| 监利县| 黄陵县| 内丘县| 长海县|