Book Image

Training Systems using Python Statistical Modeling

By : Curtis Miller
Book Image

Training Systems using Python Statistical Modeling

By: Curtis Miller

Overview of this book

Python's ease-of-use and multi-purpose nature has made it one of the most popular tools for data scientists and machine learning developers. Its rich libraries are widely used for data analysis, and more importantly, for building state-of-the-art predictive models. This book is designed to guide you through using these libraries to implement effective statistical models for predictive analytics. You’ll start by delving into classical statistical analysis, where you will learn to compute descriptive statistics using pandas. You will focus on supervised learning, which will help you explore the principles of machine learning and train different machine learning models from scratch. Next, you will work with binary prediction models, such as data classification using k-nearest neighbors, decision trees, and random forests. The book will also cover algorithms for regression analysis, such as ridge and lasso regression, and their implementation in Python. In later chapters, you will learn how neural networks can be trained and deployed for more accurate predictions, and understand which Python libraries can be used to implement them. By the end of this book, you will have the knowledge you need to design, build, and deploy enterprise-grade statistical models for machine learning using Python and its rich ecosystem of libraries for predictive analytics.
Table of Contents (9 chapters)

Support vector machines

In this section, we will look at SVMs, what they are, and how they classify data. We will discuss important hyperparameters, including how kernel methods are used. Finally, we will see their use by training them on the Titanic dataset.

With SVMs, we seek to find a hyperplane that best separates instances of two classes. These classes are assumed to be linearly separable. All data on one side of the hyperplane is predicted to belong to one class. All others belong to the other class. By best line, we mean that the plane separates classes while at the same time maximizing the distance between the line and the nearest data point, as shown in the following diagram:

The hope is that by doing this, the SVM will generalize well to data that hasn't been seen. There are two hyperparameters that are of interest to SVMs:

  • One is a tolerance parameter of C that...