Book Image

Engineering MLOps

By : Emmanuel Raj
Book Image

Engineering MLOps

By: Emmanuel Raj

Overview of this book

Engineering MLps presents comprehensive insights into MLOps coupled with real-world examples in Azure to help you to write programs, train robust and scalable ML models, and build ML pipelines to train and deploy models securely in production. The book begins by familiarizing you with the MLOps workflow so you can start writing programs to train ML models. Then you’ll then move on to explore options for serializing and packaging ML models post-training to deploy them to facilitate machine learning inference, model interoperability, and end-to-end model traceability. You’ll learn how to build ML pipelines, continuous integration and continuous delivery (CI/CD) pipelines, and monitor pipelines to systematically build, deploy, monitor, and govern ML solutions for businesses and industries. Finally, you’ll apply the knowledge you’ve gained to build real-world projects. By the end of this ML book, you'll have a 360-degree view of MLOps and be ready to implement MLOps in your organization.
Table of Contents (18 chapters)
1
Section 1: Framework for Building Machine Learning Models
7
Section 2: Deploying Machine Learning Models at Scale
13
Section 3: Monitoring Machine Learning Models in Production

Old is gold – REST API-based microservices

Old is gold. Plus, it's better to start somewhere where there are various API protocols. The Representational State Transfer (REST) protocol has become a gold standard for many applications over the years, and it's not so very different for ML applications today. The majority of companies prefer developing their ML applications based on the REST API protocol. 

A REST API or RESTful API is based on REST, an architectural method used to communicate mainly in web services development.

RESTful APIs are widely used; companies such as Amazon, Google, LinkedIn, and Twitter use them. Serving our ML models via RESTful APIs has many benefits, such as the following: 

  • Serve predictions on the fly to multiple users.
  • Add more instances to scale up the application behind a load balancer.
  • Possibly combine multiple models using different API endpoints.
  • Separate our model operating environment from the user...