- Mastering Predictive Analytics with scikit:learn and TensorFlow
- Alan Fontaine
- 218字
- 2021-07-23 16:42:24
Bagging
Bagging, also known as bootstrap aggregation, is a general purpose procedure for reducing variance in the machine learning model. It is based on the bootstrap sampling technique and is generally used with regression or classification trees, but in principle this bagging technique can be used with any model.
The following steps are involved in the bagging process:
- We choose the number of estimators or individual models to use. Let's consider this as parameter B.
- We take sample datasets from B with replacement using the bootstrap sampling from the training set.
- For every one of these training datasets, we fit the machine learning model in each of the bootstrap samples. This way, we get individual predictors for the B parameter.
- We get the ensemble prediction by aggregating all of the individual predictions.
In the regression problem, the most common way to get the ensemble prediction would be to find the average of all of the individual predictions.
In the classification problem, the most common way to get the aggregated predictions is by doing a majority vote. The majority vote can be explained by an example. Let's say that we have 100 individual predictors and 80 of them vote for one particular category. Then, we choose that category as our aggregated prediction. This is what a majority vote means.
- 大數(shù)據(jù)導(dǎo)論:思維、技術(shù)與應(yīng)用
- 大數(shù)據(jù)專業(yè)英語
- 程序設(shè)計語言與編譯
- 最后一個人類
- 基于ARM 32位高速嵌入式微控制器
- 工業(yè)機器人應(yīng)用案例集錦
- Troubleshooting OpenVPN
- Red Hat Linux 9實務(wù)自學(xué)手冊
- 空間機械臂建模、規(guī)劃與控制
- Apache源代碼全景分析(第1卷):體系結(jié)構(gòu)與核心模塊
- 嵌入式Linux系統(tǒng)實用開發(fā)
- ASP.NET 2.0 Web開發(fā)入門指南
- 電氣控制及Micro800 PLC程序設(shè)計
- Embedded Linux Development using Yocto Projects(Second Edition)
- Mastering DynamoDB