Book Image

Reproducible Data Science with Pachyderm

By : Svetlana Karslioglu
Book Image

Reproducible Data Science with Pachyderm

By: Svetlana Karslioglu

Overview of this book

Pachyderm is an open source project that enables data scientists to run reproducible data pipelines and scale them to an enterprise level. This book will teach you how to implement Pachyderm to create collaborative data science workflows and reproduce your ML experiments at scale. You’ll begin your journey by exploring the importance of data reproducibility and comparing different data science platforms. Next, you’ll explore how Pachyderm fits into the picture and its significance, followed by learning how to install Pachyderm locally on your computer or a cloud platform of your choice. You’ll then discover the architectural components and Pachyderm's main pipeline principles and concepts. The book demonstrates how to use Pachyderm components to create your first data pipeline and advances to cover common operations involving data, such as uploading data to and from Pachyderm to create more complex pipelines. Based on what you've learned, you'll develop an end-to-end ML workflow, before trying out the hyperparameter tuning technique and the different supported Pachyderm language clients. Finally, you’ll learn how to use a SaaS version of Pachyderm with Pachyderm Notebooks. By the end of this book, you will learn all aspects of running your data pipelines in Pachyderm and manage them on a day-to-day basis.
Table of Contents (16 chapters)
1
Section 1: Introduction to Pachyderm and Reproducible Data Science
5
Section 2:Getting Started with Pachyderm
12
Section 3:Pachyderm Clients and Tools

Reviewing hyperparameter tuning techniques and strategies

Hyperparameter tuning or hyperparameter optimization is a technique that ML professionals use to determine the best parameters to solve a specific ML problem. In different problems, you'd need to tune different types of parameters, such as weights in neural networks, or the number of trees in the Random Forest algorithm, or the learning rate of your model. Ultimately, selecting the best parameters helps you determine which method is best to solve a problem. A data scientist needs to understand the tunable parameters in the algorithm they use to be able to optimize them correctly.

There are a number of ML algorithms that help solve the hyperparameter optimization problem. Let's review the most common ones.

Grid search

Grid search is the simplest algorithm and is sometimes called a brute-force approach to hyperparameter optimization. This method calculates the optimum values of hyperparameters.

In Grid search...