官术网_书友最值得收藏!

Bagging

Bagging, also known as bootstrap aggregation, is a general purpose procedure for reducing variance in the machine learning model. It is based on the bootstrap sampling technique and is generally used with regression or classification trees, but in principle this bagging technique can be used with any model.

The following steps are involved in the bagging process:

  1. We choose the number of estimators or individual models to use. Let's consider this as parameter B.
  2. We take sample datasets from B with replacement using the bootstrap sampling from the training set.
  3. For every one of these training datasets, we fit the machine learning model in each of the bootstrap samples. This way, we get individual predictors for the B parameter.
  4. We get the ensemble prediction by aggregating all of the individual predictions.

In the regression problem, the most common way to get the ensemble prediction would be to find the average of all of the individual predictions.

In the classification problem, the most common way to get the aggregated predictions is by doing a majority vote. The majority vote can be explained by an example. Let's say that we have 100 individual predictors and 80 of them vote for one particular category. Then, we choose that category as our aggregated prediction. This is what a majority vote means.

主站蜘蛛池模板: 福贡县| 丰台区| 水城县| 肇东市| 米脂县| 浑源县| 利辛县| 凤冈县| 潼南县| 改则县| 张北县| 都安| 文化| 龙海市| 格尔木市| 始兴县| 大荔县| 康乐县| 尚义县| 湟中县| 和平县| 萨迦县| 阳谷县| 密山市| 舟曲县| 卫辉市| 汶上县| 农安县| 财经| 永福县| 中阳县| 胶州市| 古浪县| 邵武市| 龙山县| 海晏县| 达州市| 新平| 湖口县| 红原县| 德令哈市|