官术网_书友最值得收藏!

Ensemble methods and their working

Ensemble methods are based on a very simple idea: instead of using a single model to make a prediction, we use many models and then use some method to aggregate the predictions. Having different models is like having different points of view, and it has been demonstrated that by aggregating models that offer a different point of view; predictions can be more accurate. These methods further improve generalization over a single model because they reduce the risk of selecting a poorly performing classifier:

In the preceding diagram, we can see that each object belongs to one of three classes: triangles, circles, and squares. In this simplified example, we have two features to separate or classify the objects into the different classes. As you can see here, we can use three different classifiers and all the three classifiers represent different approaches and have different kinds of decision boundaries.

Ensemble learning combines all those individual predictions into a single one. The predictions made from combining the three boundaries usually have better properties than the ones produced by the individual models. This is the simple idea behind ensemble methods, also called ensemble learning.

The most commonly used ensemble methods are as follows:

  • Bootstrap sampling
  • Bagging
  • Random forests
  • Boosting

Before giving a high-level explanation of these methods, we need to discuss a very important statistical technique known as bootstrap sampling.

主站蜘蛛池模板: 万盛区| 渭南市| 曲水县| 二手房| 综艺| 湘乡市| 临海市| 清苑县| 曲周县| 桃江县| 永泰县| 潞城市| 称多县| 宜章县| 石嘴山市| 彭山县| 常熟市| 盘山县| 什邡市| 漳平市| 石林| 阜阳市| 广南县| 富锦市| 麻栗坡县| 游戏| 谷城县| 西昌市| 竹溪县| 昌宁县| 于田县| 苗栗县| 英山县| 嘉鱼县| 福鼎市| 葫芦岛市| 夏邑县| 吉安市| 泰州市| 法库县| 武安市|