- Deep Learning By Example
- Ahmed Menshawy
- 157字
- 2021-06-24 18:52:36
Feature selection
The number of explanatory features (input variables) of a sample can be enormous wherein you get xi=(xi1, xi2, xi3, ... , xid) as a training sample (observation/example) and d is very large. An example of this can be a document classification task3, where you get 10,000 different words and the input variables will be the number of occurrences of different words.
This enormous number of input variables can be problematic and sometimes a curse because we have many input variables and few training samples to help us in the learning procedure. To avoid this curse of having an enormous number of input variables (curse of dimensionality), data scientists use dimensionality reduction techniques in order to select a subset from the input variables. For example, in the text classification task they can do the following:
- Extracting relevant inputs (for instance, mutual information measure)
- Principal component analysis (PCA)
- Grouping (cluster) similar words (this uses a similarity measure)
- 計算機原理
- PostgreSQL Administration Essentials
- JSF2和RichFaces4使用指南
- Docker on Amazon Web Services
- 單片機C語言程序設計完全自學手冊
- Building a BeagleBone Black Super Cluster
- Salesforce Advanced Administrator Certification Guide
- Silverlight 2完美征程
- 人工智能:智能人機交互
- Natural Language Processing and Computational Linguistics
- 傳感器原理及實用技術
- 計算機硬件技術基礎(第2版)
- EJB JPA數據庫持久層開發實踐詳解
- C#編程兵書
- JSP通用范例開發金典