Book Image

Smarter Decisions - The Intersection of Internet of Things and Decision Science

By : Jojo Moolayil
Book Image

Smarter Decisions - The Intersection of Internet of Things and Decision Science

By: Jojo Moolayil

Overview of this book

With an increasing number of devices getting connected to the Internet, massive amounts of data are being generated that can be used for analysis. This book helps you to understand Internet of Things in depth and decision science, and solve business use cases. With IoT, the frequency and impact of the problem is huge. Addressing a problem with such a huge impact requires a very structured approach. The entire journey of addressing the problem by defining it, designing the solution, and executing it using decision science is articulated in this book through engaging and easy-to-understand business use cases. You will get a detailed understanding of IoT, decision science, and the art of solving a business problem in IoT through decision science. By the end of this book, you’ll have an understanding of the complex aspects of decision making in IoT and will be able to take that knowledge with you onto whatever project calls for it
Table of Contents (15 chapters)
Smarter Decisions – The Intersection of Internet of Things and Decision Science
Credits
About the Author
About the Reviewer
eBooks, discount offers, and more
Preface

Ensemble modeling - XGBoost


XGBoost, that is, Extreme Gradient Boosting, is a very popular machine learning ensemble technique that has helped data scientists across the globe to achieve great results with phenomenal accuracy. XGBoost is built on the principles of ensemble modeling and is an improved version of the Gradient Boosted Machine algorithm. In general, the XgBoost algorithm creates multiple classifiers that are weak learners, which means a model that gives a bit better accuracy than just a random guess. The learner in the ensemble model can be a linear or tree model that is built iteratively with random sampling along with an added weight from the learnings of the previously built model. At each step, a tree is built and the cases where the tree has failed to classify an outcome correctly is assigned a corresponding weight. The next iteration of model building learns from the mistakes of the previous model. At each step, the weight of an incorrect prediction is calculated using...