官术网_书友最值得收藏!

Dimensionality reduction

Dimensionality reduction is used to reduce the dimensionality of a dataset. It is really helpful in cases where the problem becomes intractable, when the number of variables increases. By using the term dimensionality, we are referring to the features. One of the basic reduction techniques is feature engineering.

Generally, we have many dimensionality reduction algorithms:

  • Low variance filter: Dropping variables that have low variance, compared to others.
  • High correlation filter: This identifies the variables with high correlation, by using pearson or polychoric, and selects one of them using the Variance Inflation Factor (VIF).
  • Backward feature elimination: This is done by computing the sum of square of error (SSE) after eliminating each variable n times.
  • Linear Discriminant Analysis (LDA): This reduces the number of dimensions, n, from the original to the number of classes?—?1 number of features.
  • Principal Component Analysis (PCA): This is a statistical procedure that transforms variables into a new set of variables (principle components).
主站蜘蛛池模板: 修武县| 宽甸| 南京市| 新干县| 白朗县| 山西省| 忻城县| 邯郸市| 赣州市| 岳池县| 门头沟区| 施甸县| 阿鲁科尔沁旗| 普格县| 定西市| 建平县| 博罗县| 玉树县| 手游| 嵊州市| 辉南县| 溧阳市| 延川县| 平山县| 东阳市| 璧山县| 彭山县| 武宣县| 分宜县| 桐乡市| 乐亭县| 武宣县| 张家港市| 旅游| 剑河县| 瑞昌市| 日喀则市| 射洪县| 马边| 叙永县| 伽师县|