Ensemble Machine Learning: A beginner’s guide that combines powerful machine learning algorithms to build optimized models
An effective guide to using ensemble techniques to enhance machine learning models
Ensembling is a technique of combining two or more similar or dissimilar machine learning algorithms to create a model that delivers superior prediction power. This book will show you how you can use many weak algorithms to make a strong predictive model. This book contains Python code for different machine learning algorithms so that you can easily understand and implement it in your own systems.
This book covers different machine learning algorithms that are widely used in the practical world to make predictions and classifications. It addresses different aspects of a prediction framework, such as data pre-processing, model training, validation of the model, and more. You will gain knowledge of different machine learning aspects such as bagging (decision trees and random forests), Boosting (Ada-boost) and stacking (a combination of bagging and boosting algorithms).
Then you’ll learn how to implement them by building ensemble models using TensorFlow and Python libraries such as scikit-learn and NumPy. As machine learning touches almost every field of the digital world, you’ll see how these algorithms can be used in different applications such as computer vision, speech recognition, making recommendations, grouping and document classification, fitting regression on data, and more.
By the end of this book, you’ll understand how to combine machine learning algorithms to work behind the scenes and reduce challenges and common problems.
This book is for data scientists, machine learning practitioners, and deep learning enthusiasts who want to implement ensemble techniques and make a deep dive into the world of machine learning algorithms. You are expected to understand Python code and have a basic knowledge of probability theories, statistics, and linear algebra.
Chapter 1. Introduction of Ensemble Learning
Chapter 2. Decision Trees
Chapter 3. Random Forest
Chapter 4. Random Subspace and KNN Bagging
Chapter 5. AdaBoost Classifier
Chapter 6. Gradient Boosting Machines
Chapter 7. XGBoost- extreme gradient boosting
Chapter 8. Stacked Generalization
Chapter 9. Stacked Generalization-Part 2
Chapter 10. Modern Day Machine Learning
Chapter 11. Appendix: Troubleshooting