Book Image

MLOps with Red Hat OpenShift

By : Ross Brigoli, Faisal Masood
Book Image

MLOps with Red Hat OpenShift

By: Ross Brigoli, Faisal Masood

Overview of this book

MLOps with OpenShift offers practical insights for implementing MLOps workflows on the dynamic OpenShift platform. As organizations worldwide seek to harness the power of machine learning operations, this book lays the foundation for your MLOps success. Starting with an exploration of key MLOps concepts, including data preparation, model training, and deployment, you’ll prepare to unleash OpenShift capabilities, kicking off with a primer on containers, pods, operators, and more. With the groundwork in place, you’ll be guided to MLOps workflows, uncovering the applications of popular machine learning frameworks for training and testing models on the platform. As you advance through the chapters, you’ll focus on the open-source data science and machine learning platform, Red Hat OpenShift Data Science, and its partner components, such as Pachyderm and Intel OpenVino, to understand their role in building and managing data pipelines, as well as deploying and monitoring machine learning models. Armed with this comprehensive knowledge, you’ll be able to implement MLOps workflows on the OpenShift platform proficiently.
Table of Contents (13 chapters)
Free Chapter
1
Part 1: Introduction
3
Part 2: Provisioning and Configuration
6
Part 3: Operating ML Workloads

Using Jupyter Notebooks in OpenShift

Jupyter Notebooks is the de facto standard environment for data scientists and data engineers to analyze data and build ML models. Since the notebooks provided by the platform run as containers, you will see how your team can start quickly and consistently by adopting the platform. The platform provides a rapid way to develop, train, and test ML models and deploy them onto production. In the ODS platform, the Jupyter Notebooks environments are referred to as workbenches. You will learn about workbenches later in this section. But first, we need to learn how to create these environments.

We’ll start by provisioning S3 object storage for you to access the data required for the model training process. This is part of the platform setup, and data scientists will not have to execute these steps for their day-to-day work.