官术网_书友最值得收藏!

Ensemble methods and their working

Ensemble methods are based on a very simple idea: instead of using a single model to make a prediction, we use many models and then use some method to aggregate the predictions. Having different models is like having different points of view, and it has been demonstrated that by aggregating models that offer a different point of view; predictions can be more accurate. These methods further improve generalization over a single model because they reduce the risk of selecting a poorly performing classifier:

In the preceding diagram, we can see that each object belongs to one of three classes: triangles, circles, and squares. In this simplified example, we have two features to separate or classify the objects into the different classes. As you can see here, we can use three different classifiers and all the three classifiers represent different approaches and have different kinds of decision boundaries.

Ensemble learning combines all those individual predictions into a single one. The predictions made from combining the three boundaries usually have better properties than the ones produced by the individual models. This is the simple idea behind ensemble methods, also called ensemble learning.

The most commonly used ensemble methods are as follows:

  • Bootstrap sampling
  • Bagging
  • Random forests
  • Boosting

Before giving a high-level explanation of these methods, we need to discuss a very important statistical technique known as bootstrap sampling.

主站蜘蛛池模板: 安乡县| 嘉鱼县| 云和县| 鄂托克旗| 阿鲁科尔沁旗| 灯塔市| 江安县| 龙泉市| 衡阳市| 景泰县| 南京市| 寿光市| 石阡县| 新巴尔虎右旗| 大方县| 扎囊县| 临潭县| 巧家县| 商河县| 岗巴县| 达孜县| 甘德县| 乌鲁木齐县| 万宁市| 沿河| 石狮市| 洞头县| 成安县| 富蕴县| 临泽县| 大理市| 吉首市| 锡林郭勒盟| 丹江口市| 当阳市| 阆中市| 普安县| 明溪县| 卓资县| 廉江市| 房产|