- Mastering Java for Data Science
- Alexey Grigorev
- 165字
- 2021-07-02 23:44:31
Dimensionality reduction
Another group of unsupervised learning algorithms is dimensionality reduction algorithms. This group of algorithms compresses the dataset, keeping only the most useful information. If our dataset has too much information, it can be hard for a machine learning algorithm to use all of it at the same time. It may just take too long for the algorithm to process all the data and we would like to compress the data, so processing it takes less time.
There are multiple algorithms that can reduce the dimensionality of the data, including Principal Component Analysis (PCA), Locally linear embedding, and t-SNE. All these algorithms are examples of unsupervised dimensionality reduction techniques.
Not all dimensionality reduction algorithms are unsupervised; some of them can use labels to reduce the dimensionality better. For example, many feature selection algorithms rely on labels to see what features are useful and what are not.
We will talk more about this in Chapter 5, Unsupervised Learning - Clustering and Dimensionality Reduction.
- 數據浪潮
- GitHub Essentials
- 大規模數據分析和建模:基于Spark與R
- 輕松學大數據挖掘:算法、場景與數據產品
- Libgdx Cross/platform Game Development Cookbook
- 數據庫系統原理及應用教程(第4版)
- MySQL 8.x從入門到精通(視頻教學版)
- Web Services Testing with soapUI
- 數據庫應用系統技術
- 企業大數據處理:Spark、Druid、Flume與Kafka應用實踐
- 企業級大數據項目實戰:用戶搜索行為分析系統從0到1
- Artificial Intelligence for Big Data
- Learning Ansible
- 數據迷霧:洞察數據的價值與內涵
- R數據挖掘實戰