- Deep Learning By Example
- Ahmed Menshawy
- 157字
- 2021-06-24 18:52:36
Feature selection
The number of explanatory features (input variables) of a sample can be enormous wherein you get xi=(xi1, xi2, xi3, ... , xid) as a training sample (observation/example) and d is very large. An example of this can be a document classification task3, where you get 10,000 different words and the input variables will be the number of occurrences of different words.
This enormous number of input variables can be problematic and sometimes a curse because we have many input variables and few training samples to help us in the learning procedure. To avoid this curse of having an enormous number of input variables (curse of dimensionality), data scientists use dimensionality reduction techniques in order to select a subset from the input variables. For example, in the text classification task they can do the following:
- Extracting relevant inputs (for instance, mutual information measure)
- Principal component analysis (PCA)
- Grouping (cluster) similar words (this uses a similarity measure)
- 大數據導論:思維、技術與應用
- 基于LabWindows/CVI的虛擬儀器設計與應用
- TestStand工業自動化測試管理(典藏版)
- 機艙監測與主機遙控
- 水晶石精粹:3ds max & ZBrush三維數字靜幀藝術
- Machine Learning with Apache Spark Quick Start Guide
- Windows Server 2008 R2活動目錄內幕
- Godot Engine Game Development Projects
- 網絡管理工具實用詳解
- Mastering GitLab 12
- 自動化生產線安裝與調試(三菱FX系列)(第二版)
- 大數據案例精析
- 電動汽車驅動與控制技術
- Machine Learning in Java
- Windows 7來了