官术网_书友最值得收藏!

What this book covers

Chapter 1, Get Closer to Your Data, explores a dataset and implements hands-on coding with Python for exploratory data analysis using statistical methods and visualization for the dataset.

Chapter 2, Getting Started with Ensemble Machine Learning, explores what ensemble learning is and how it can help in real-life scenarios. Basic ensemble techniques, including averaging, weighted averaging, and max-voting, are explained. These techniques form the basis for ensemble techniques, and an understanding of them will lay the groundwork for readers to move to more advanced stage after reading this chapter.

Chapter 3, Resampling Methods, introduces a handful of algorithms that will be useful when we get into an ensemble of multiple heterogeneous algorithms. This chapter uses scikit-learn to prepare all the algorithms to be used.

Chapter 4, Statistical and Machine Learning Algorithms, helps the readers to get to know various types of resampling methods that are used by machine-learning algorithms. Each resampling method has its advantages and disadvantages, which are explained to the readers. The readers also learn the code to be executed for each type of sampling.

Chapter 5, Bag the Models with Bagging, provides the readers with an understanding of what bootstrap aggregation is and how the bootstrap results can be aggregated, in a process also known as bagging.

Chapter 6, When in Doubt, Use Random Forests, introduces the random forest algorithm. It will introduce to readers how, and what kind of, ensemble techniques are used by Random Forest and how this helps our models avoid overfitting.

Chapter 7Boosting Model Performance with Boosting, introduces boosting and discusses how it helps to improve a model performance by reducing variances and increasing accuracy. This chapter provides information such as the fact that boosting is not robust against outliers and noisy data but is flexible and can be used with a loss function.

Chapter 8, Blend It with Stacking, applies stacking to learn the optimal combination of base learners. This chapter will acquaint readers with stacking, which is also known as stacked generalization.

Chapter 9, Homogeneous Ensemble Using Keras, is a complete code walk-through on a classification case study for recognizing hand-written digits with homogeneous algorithms – in this case, multiple neural network models using Keras.

Chapter 10Heterogeneous Ensemble Classifiers Using H2O, is a complete code walk-through on a classification case study for default prediction with an ensemble of multiple heterogeneous algorithms using scikit-learn.

Chapter 11Heterogeneous Ensemble for Text Classification Using NLP,  is a complete code walk-through on a classification case study to classify sentiment polarity using an ensemble of multiple heterogeneous algorithms. Here, NLP techniques such as semantics are used to improve the accuracy of classification. Then, the mined text information is used to employ ensemble classification techniques for sentiment analysis. In this case study, the H2O library is used for building models.

Chapter 12Homogeneous Ensemble for Multiclass Classification Using Keras, is a complete code walk-through on a classification case study for multiple classification with homogeneous ensemble using data diversity with the tf.keras module from TensorFlow.

主站蜘蛛池模板: 莆田市| 洛川县| 玛曲县| 富锦市| 鄂伦春自治旗| 佛冈县| 义马市| 武定县| 鄢陵县| 六盘水市| 耒阳市| 乌审旗| 汨罗市| 承德市| 义马市| 桂林市| 嫩江县| 新密市| 邯郸县| 辽阳县| 沁阳市| 曲松县| 边坝县| 萍乡市| 右玉县| 建始县| 灯塔市| 中牟县| 日照市| 利津县| 株洲县| 鄄城县| 武定县| 平昌县| 阆中市| 秦皇岛市| 托克托县| 邵东县| 湖北省| 东乌| 佛学|