Book Image

Machine Learning Quick Reference

By : Rahul Kumar
Book Image

Machine Learning Quick Reference

By: Rahul Kumar

Overview of this book

Machine learning makes it possible to learn about the unknowns and gain hidden insights into your datasets by mastering many tools and techniques. This book guides you to do just that in a very compact manner. After giving a quick overview of what machine learning is all about, Machine Learning Quick Reference jumps right into its core algorithms and demonstrates how they can be applied to real-world scenarios. From model evaluation to optimizing their performance, this book will introduce you to the best practices in machine learning. Furthermore, you will also look at the more advanced aspects such as training neural networks and work with different kinds of data, such as text, time-series, and sequential data. Advanced methods and techniques such as causal inference, deep Gaussian processes, and more are also covered. By the end of this book, you will be able to train fast, accurate machine learning models at your fingertips, which you can easily use as a point of reference.
Table of Contents (18 chapters)
Title Page
Copyright and Credits
About Packt
Contributors
Preface
Index

Overfitting


We have already discussed overfitting in detail. However, let's have a recap of what we learned and what overfitting is in a neural network scenario.

By now, we are cognizant of the fact that, when a large number of parameters (in deep learning) are available at our disposal to map and explain an event, more often than not, the model built using these parameters will tend to have a good fit and try to showcase that it has the ability to describe the event properly. However, the real test of any model is always on unseen data, and we were able to assess how the model fares on such unseen data points. We expect our model to have an attribute of generalization and it will enable the model to score on test data (unseen) in alignment with the trained one. But, a number of times our model fails to generalize when it comes to the unseen data, as the model has not learned the insights and causal relationship of the event. In this scenario, one might be able to see the huge gulf of variance...