- Mastering Java for Data Science
- Alexey Grigorev
- 165字
- 2021-07-02 23:44:31
Dimensionality reduction
Another group of unsupervised learning algorithms is dimensionality reduction algorithms. This group of algorithms compresses the dataset, keeping only the most useful information. If our dataset has too much information, it can be hard for a machine learning algorithm to use all of it at the same time. It may just take too long for the algorithm to process all the data and we would like to compress the data, so processing it takes less time.
There are multiple algorithms that can reduce the dimensionality of the data, including Principal Component Analysis (PCA), Locally linear embedding, and t-SNE. All these algorithms are examples of unsupervised dimensionality reduction techniques.
Not all dimensionality reduction algorithms are unsupervised; some of them can use labels to reduce the dimensionality better. For example, many feature selection algorithms rely on labels to see what features are useful and what are not.
We will talk more about this in Chapter 5, Unsupervised Learning - Clustering and Dimensionality Reduction.
- 大規模數據分析和建模:基于Spark與R
- Access 2016數據庫教程(微課版·第2版)
- Lean Mobile App Development
- 企業級數據與AI項目成功之道
- Flutter Projects
- 大數據分析:數據倉庫項目實戰
- Oracle 11g+ASP.NET數據庫系統開發案例教程
- 數據庫應用系統技術
- 實現領域驅動設計
- Spring Boot 2.0 Cookbook(Second Edition)
- 實用預測分析
- 從零進階!數據分析的統計基礎(第2版)
- 敏捷數據分析工具箱:深入解析ADW+OAC
- Discovering Business Intelligence Using MicroStrategy 9
- MySQL必知必會