官术网_书友最值得收藏!

Feature selection – say no to bad attributes

By this chapter, we will have a level of comfort when dealing with new datasets. We will have under our belt the abilities to understand and clean the data in front of us. Once we are able to work with the data given to us, we can start to make big decisions such as, at what point is a feature actually an attribute. Recall that by this distinction, feature versus attribute, the question really is, which columns are not helping my ML pipeline and therefore are hurting my pipeline and should be removed? This chapter focuses on techniques used to make the decision of which attributes to get rid of in our dataset. We will explore several statistical and iterative processes that will aid us in this decision.

Among these processes are:

  • Correlation coefficients
  • Identifying and removing multicollinearity
  • Chi-squared tests
  • Anova tests
  • Interpretation of p-values
  • Iterative feature selection
  • Using machine learning to measure entropy and information gain

All of these procedures will attempt to suggest the removal of features and will give different reasons for doing so. Ultimately, it will be up to us, the data scientists, to make the final call over which features will be allowed to remain and contribute to our machine learning algorithms.

主站蜘蛛池模板: 保德县| 渝北区| 龙川县| 余干县| 汉源县| 大宁县| 修武县| 大厂| 内乡县| 务川| 喀喇沁旗| 从化市| 西昌市| 滦南县| 招远市| 佛坪县| 玛多县| 昌黎县| 涞源县| 温泉县| 宿松县| 左贡县| 崇信县| 保亭| 永善县| 仲巴县| 修文县| 泸水县| 鹤峰县| 保靖县| 和龙市| 和静县| 芦山县| 禹州市| 平南县| 邹城市| 乌拉特前旗| 嘉定区| 河北区| 三台县| 克东县|