Book Image

MLOps with Red Hat OpenShift

By : Ross Brigoli, Faisal Masood
Book Image

MLOps with Red Hat OpenShift

By: Ross Brigoli, Faisal Masood

Overview of this book

MLOps with OpenShift offers practical insights for implementing MLOps workflows on the dynamic OpenShift platform. As organizations worldwide seek to harness the power of machine learning operations, this book lays the foundation for your MLOps success. Starting with an exploration of key MLOps concepts, including data preparation, model training, and deployment, you’ll prepare to unleash OpenShift capabilities, kicking off with a primer on containers, pods, operators, and more. With the groundwork in place, you’ll be guided to MLOps workflows, uncovering the applications of popular machine learning frameworks for training and testing models on the platform. As you advance through the chapters, you’ll focus on the open-source data science and machine learning platform, Red Hat OpenShift Data Science, and its partner components, such as Pachyderm and Intel OpenVino, to understand their role in building and managing data pipelines, as well as deploying and monitoring machine learning models. Armed with this comprehensive knowledge, you’ll be able to implement MLOps workflows on the OpenShift platform proficiently.
Table of Contents (13 chapters)
Free Chapter
1
Part 1: Introduction
3
Part 2: Provisioning and Configuration
6
Part 3: Operating ML Workloads

Autoscaling the deployed models

While creating a model server, you will be presented with the option to set the number of replicas. This corresponds to the number of instances of the model servers to be created. This allows you to increase or decrease the serving capacity of your model servers. Figure 5.12 shows this option as Model server replicas:

Figure 5.12 – Add model server

Figure 5.12 – Add model server

However, with this approach, you need to decide on the number of serving instances or replicas at the time of the model server’s creation. OpenShift provides another construct where you can add an automatic scaler that increases or decreases the number of replicas of the model server based on the memory or CPU utilization of the model server instances. This construct is called horizontal pod autoscaling. This allows us to automatically scale workloads to match the demand.

Let’s see how the model server that we defined with the data science project is deployed...