舉報

會員
IBM SPSS Modeler Cookbook
Keith McCormick Dean Abbott Meta S. Brown Tom Khabaza Scott R.Mutchler 著
更新時間:2021-07-23 16:01:53
開會員,本書免費讀 >
最新章節:
Index
Thisisapracticalcookbookwithintermediate-advancedrecipesforSPSSModelerdataanalysts.Itisloadedwithstep-by-stepexamplesexplainingtheprocessfollowedbytheexperts.Ifyouhavehadsomehands-onexperiencewithIBMSPSSModelerandnowwanttogodeeperandtakemorecontroloveryourdataminingprocess,thisistheguideforyou.Itisidealforpractitionerswhowanttobreakintoadvancedanalytics.
最新章節
- Index
- Produce a project plan – ensuring a realistic timeline by Keith McCormick>
- Translating your business objective into a data mining objective by Dean Abbott>
- Assessing the situation by Meta Brown>
- Define business objectives by Tom Khabaza>
- Introduction>
品牌:中圖公司
上架時間:2021-07-23 15:04:47
出版社:Packt Publishing
本書數字版權由中圖公司提供,并由其授權上海閱文信息技術有限公司制作發行
- Index 更新時間:2021-07-23 16:01:53
- Produce a project plan – ensuring a realistic timeline by Keith McCormick>
- Translating your business objective into a data mining objective by Dean Abbott>
- Assessing the situation by Meta Brown>
- Define business objectives by Tom Khabaza>
- Introduction>
- Appendix A. Business Understanding
- Rolling your own modeling algorithm – Weibull analysis>
- Automating HTML reports and graphs>
- Automating time series forecasts>
- Optimizing K-means cluster solutions>
- Detecting outliers with the jackknife method>
- Implementing champion/challenger model management>
- Quantifying variable importance with Monte Carlo simulation>
- Building iterative Neural Network forecasts>
- Introduction>
- Chapter 8. CLEM Scripting
- Combining generated filters>
- Changing formatting of fields in a Table node>
- Reformatting data for reporting with a Transpose node>
- Creating a classification tree financial summary using aggregate and an Excel Export node>
- Using aggregate to write cluster centers to Excel for conditional formatting>
- Correcting a confusion matrix for an imbalanced target variable by incorporating priors>
- Using classification trees to explore the predictions of a Neural Network>
- How (and why) to validate as well as test>
- Introduction>
- Chapter 7. Modeling – Assessment Evaluation Deployment and Monitoring
- Next-Best-Offer for large datasets>
- Using Auto Classifier to tune models>
- Using KNN to match similar cases>
- Creating bagged logistic regression models>
- Creating a bootstrap sample>
- Using Neural Network for Feature Selection>
- Building models with and without outliers>
- Evaluating balancing with Auto Classifier>
- Introduction>
- Chapter 6. Selecting and Building a Model
- Sequence processing>
- Parsing and performing a conversion on a complex stream>
- Parsing nonstandard dates>
- Changing large numbers of variable names without scripting>
- Multiplying out using Cartesian product merge user source and derive dummy>
- Cartesian product merge using key-less merge by key>
- Shuffle-down (nonstandard aggregation)>
- Merging a lookup table>
- Speeding up merge with caching and optimization settings>
- Introduction>
- Chapter 5. Data Preparation – Integrate and Format
- Creating time-aligned cohorts>
- Using Association Rules for interaction detection/feature creation>
- Creating flag variables for aggregation>
- Transforming high skew and kurtosis variables with a multiple Derive node>
- Grouping categorical values>
- Calculating and comparing conversion rates>
- Building transformations with multiple Derive nodes>
- Introduction>
- Chapter 4. Data Preparation – Construct
- Producing longer Soundex codes>
- Using neuro-fuzzy searching to find similar names>
- Searching for similar records using a Neural Network for inexact matching>
- Using random imputation to match a variable's distribution>
- Imputing missing values randomly from uniform or normal distributions>
- Imputing in-stream mean or median>
- Using a full data model/partial data model approach to address missing data>
- Binning scale variables to address missing data>
- Introduction>
- Chapter 3. Data Preparation – Clean
- Selecting variables using single-antecedent Association Rules>
- Selecting variables using the Means node>
- Selecting variables using the CHAID Modeling node>
- Removing redundant variables using correlation matrices>
- Evaluating the use of sampling for speed>
- Running a Statistics node on anti-join to evaluate the potential missing data>
- Using the Feature Selection node creatively to remove or decapitate perfect predictors>
- Introduction>
- Chapter 2. Data Preparation – Select
- Detecting potential model instability early using the Partition node and Feature Selection node>
- Creating an Outlier report to give to SMEs>
- Using an @NULL multiple Derive to explore missing data>
- Using a single cluster K-means as an alternative to anomaly detection>
- Using CHAID stumps when interviewing an SME>
- Evaluating the need to sample from the initial data>
- Using an empty aggregate to evaluate sample size>
- Introduction>
- Chapter 1. Data Understanding
- Preface
- www.PacktPub.com
- About the Reviewers
- About the Authors
- Foreword
- Credits
- 版權信息
- 封面
- 封面
- 版權信息
- Credits
- Foreword
- About the Authors
- About the Reviewers
- www.PacktPub.com
- Preface
- Chapter 1. Data Understanding
- Introduction>
- Using an empty aggregate to evaluate sample size>
- Evaluating the need to sample from the initial data>
- Using CHAID stumps when interviewing an SME>
- Using a single cluster K-means as an alternative to anomaly detection>
- Using an @NULL multiple Derive to explore missing data>
- Creating an Outlier report to give to SMEs>
- Detecting potential model instability early using the Partition node and Feature Selection node>
- Chapter 2. Data Preparation – Select
- Introduction>
- Using the Feature Selection node creatively to remove or decapitate perfect predictors>
- Running a Statistics node on anti-join to evaluate the potential missing data>
- Evaluating the use of sampling for speed>
- Removing redundant variables using correlation matrices>
- Selecting variables using the CHAID Modeling node>
- Selecting variables using the Means node>
- Selecting variables using single-antecedent Association Rules>
- Chapter 3. Data Preparation – Clean
- Introduction>
- Binning scale variables to address missing data>
- Using a full data model/partial data model approach to address missing data>
- Imputing in-stream mean or median>
- Imputing missing values randomly from uniform or normal distributions>
- Using random imputation to match a variable's distribution>
- Searching for similar records using a Neural Network for inexact matching>
- Using neuro-fuzzy searching to find similar names>
- Producing longer Soundex codes>
- Chapter 4. Data Preparation – Construct
- Introduction>
- Building transformations with multiple Derive nodes>
- Calculating and comparing conversion rates>
- Grouping categorical values>
- Transforming high skew and kurtosis variables with a multiple Derive node>
- Creating flag variables for aggregation>
- Using Association Rules for interaction detection/feature creation>
- Creating time-aligned cohorts>
- Chapter 5. Data Preparation – Integrate and Format
- Introduction>
- Speeding up merge with caching and optimization settings>
- Merging a lookup table>
- Shuffle-down (nonstandard aggregation)>
- Cartesian product merge using key-less merge by key>
- Multiplying out using Cartesian product merge user source and derive dummy>
- Changing large numbers of variable names without scripting>
- Parsing nonstandard dates>
- Parsing and performing a conversion on a complex stream>
- Sequence processing>
- Chapter 6. Selecting and Building a Model
- Introduction>
- Evaluating balancing with Auto Classifier>
- Building models with and without outliers>
- Using Neural Network for Feature Selection>
- Creating a bootstrap sample>
- Creating bagged logistic regression models>
- Using KNN to match similar cases>
- Using Auto Classifier to tune models>
- Next-Best-Offer for large datasets>
- Chapter 7. Modeling – Assessment Evaluation Deployment and Monitoring
- Introduction>
- How (and why) to validate as well as test>
- Using classification trees to explore the predictions of a Neural Network>
- Correcting a confusion matrix for an imbalanced target variable by incorporating priors>
- Using aggregate to write cluster centers to Excel for conditional formatting>
- Creating a classification tree financial summary using aggregate and an Excel Export node>
- Reformatting data for reporting with a Transpose node>
- Changing formatting of fields in a Table node>
- Combining generated filters>
- Chapter 8. CLEM Scripting
- Introduction>
- Building iterative Neural Network forecasts>
- Quantifying variable importance with Monte Carlo simulation>
- Implementing champion/challenger model management>
- Detecting outliers with the jackknife method>
- Optimizing K-means cluster solutions>
- Automating time series forecasts>
- Automating HTML reports and graphs>
- Rolling your own modeling algorithm – Weibull analysis>
- Appendix A. Business Understanding
- Introduction>
- Define business objectives by Tom Khabaza>
- Assessing the situation by Meta Brown>
- Translating your business objective into a data mining objective by Dean Abbott>
- Produce a project plan – ensuring a realistic timeline by Keith McCormick>
- Index 更新時間:2021-07-23 16:01:53