Book Image

Python Data Science Essentials - Second Edition

By : Luca Massaron, Alberto Boschetti
Book Image

Python Data Science Essentials - Second Edition

By: Luca Massaron, Alberto Boschetti

Overview of this book

Fully expanded and upgraded, the second edition of Python Data Science Essentials takes you through all you need to know to suceed in data science using Python. Get modern insight into the core of Python data, including the latest versions of Jupyter notebooks, NumPy, pandas and scikit-learn. Look beyond the fundamentals with beautiful data visualizations with Seaborn and ggplot, web development with Bottle, and even the new frontiers of deep learning with Theano and TensorFlow. Dive into building your essential Python 3.5 data science toolbox, using a single-source approach that will allow to to work with Python 2.7 as well. Get to grips fast with data munging and preprocessing, and all the techniques you need to load, analyse, and process your data. Finally, get a complete overview of principal machine learning algorithms, graph analysis techniques, and all the visualization and deployment instruments that make it easier to present your results to an audience of both data science experts and business users.
Table of Contents (13 chapters)
Python Data Science Essentials - Second Edition
Credits
About the Authors
About the Reviewer
www.PacktPub.com
Preface

Nonlinear algorithms


Support Vector Machine (SVM) is a powerful and advanced supervised learning technique for classification and regression that can automatically fit linear and nonlinear models.

SVM algorithms have quite a few advantages over other machine learning algorithms:

  • They can handle the majority of supervised problems such as regression, classification, and anomaly detection (anyway, they are actually best at binary classification).

  • Provide a good handling of noisy data and outliers. They tend to overfit less since they only work with some particular examples, the support vectors.

  • Work fine with datasets presenting more features than examples, though, as other machine learning algorithms, also SVM would gain both from dimensionality reduction and feature selection.

As drawbacks, we have to mention these:

  • They provide just estimates, but no probabilities unless you run some time-consuming and computationally intensive probability calibration by means of Platt scaling

  • They scale super...