官术网_书友最值得收藏!

Bagging

Bagging, also known as bootstrap aggregation, is a general purpose procedure for reducing variance in the machine learning model. It is based on the bootstrap sampling technique and is generally used with regression or classification trees, but in principle this bagging technique can be used with any model.

The following steps are involved in the bagging process:

  1. We choose the number of estimators or individual models to use. Let's consider this as parameter B.
  2. We take sample datasets from B with replacement using the bootstrap sampling from the training set.
  3. For every one of these training datasets, we fit the machine learning model in each of the bootstrap samples. This way, we get individual predictors for the B parameter.
  4. We get the ensemble prediction by aggregating all of the individual predictions.

In the regression problem, the most common way to get the ensemble prediction would be to find the average of all of the individual predictions.

In the classification problem, the most common way to get the aggregated predictions is by doing a majority vote. The majority vote can be explained by an example. Let's say that we have 100 individual predictors and 80 of them vote for one particular category. Then, we choose that category as our aggregated prediction. This is what a majority vote means.

主站蜘蛛池模板: 景宁| 肃北| 虞城县| 加查县| 霞浦县| 黑山县| 湟中县| 鹤山市| 宁波市| 汽车| 高台县| 天台县| 仁怀市| 蛟河市| 柯坪县| 内江市| 同仁县| 曲周县| 永修县| 黔南| 安化县| 西城区| 厦门市| 荆州市| 武威市| 天峻县| 玉门市| 邯郸县| 古交市| 曲阳县| 鞍山市| 铜陵市| 井冈山市| 栖霞市| 南和县| 北碚区| 灵石县| 自贡市| 原阳县| 呼和浩特市| 鄂托克前旗|