官术网_书友最值得收藏!

What is a decision tree?

Decision trees are a family of non-parametric supervised learning methods. In the decision tree algorithm, we start with the complete dataset and split it into two partitions based on a simple rule. The splitting continues until a specified criterion is met. The nodes at which the split is made are called interior nodes and the final endpoints are called terminal or leaf nodes.

As an example, let us look at the following tree:

Here, we are assuming that the exoplanet data has only two properties: flux.1 and flux.2. First, we make a decision if flux.1 > 400 and then divide the data into two partitions. Then we divide the data again based on flux.2 feature, and that division decides whether the planet is an exoplanet or not. How did we decide that condition flux.1 > 400? We did not. This was just to demonstrate a decision tree. During the training phase, that's what the model learns – the parameters of conditions that divide the data into partitions.

For classification problems, the decision tree has leaf nodes that shows the result as the discrete classification of the data and for regression problems, the leaf nodes show the results as a predicted number. Decision trees, thus, are also popularly known as Classification and Regression Trees (CART).

主站蜘蛛池模板: 江油市| 信阳市| 贡嘎县| 德清县| 武冈市| 叶城县| 焉耆| 洪雅县| 灌阳县| 湖南省| 沾化县| 尚义县| 武威市| 海伦市| 玛曲县| 石渠县| 万荣县| 萨嘎县| 田东县| 仁怀市| 米易县| 启东市| 图木舒克市| 南部县| 肇源县| 哈密市| 贵溪市| 平昌县| 聂荣县| 井陉县| 文登市| 楚雄市| 怀安县| 高邮市| 沅江市| 平邑县| 天等县| 伊通| 高平市| 林西县| 玉田县|