Book Image

Automated Machine Learning

By : Adnan Masood
Book Image

Automated Machine Learning

By: Adnan Masood

Overview of this book

Every machine learning engineer deals with systems that have hyperparameters, and the most basic task in automated machine learning (AutoML) is to automatically set these hyperparameters to optimize performance. The latest deep neural networks have a wide range of hyperparameters for their architecture, regularization, and optimization, which can be customized effectively to save time and effort. This book reviews the underlying techniques of automated feature engineering, model and hyperparameter tuning, gradient-based approaches, and much more. You'll discover different ways of implementing these techniques in open source tools and then learn to use enterprise tools for implementing AutoML in three major cloud service providers: Microsoft Azure, Amazon Web Services (AWS), and Google Cloud Platform. As you progress, you’ll explore the features of cloud AutoML platforms by building machine learning models using AutoML. The book will also show you how to develop accurate models by automating time-consuming and repetitive tasks in the machine learning development lifecycle. By the end of this machine learning book, you’ll be able to build and deploy AutoML models that are not only accurate, but also increase productivity, allow interoperability, and minimize feature engineering tasks.
Table of Contents (15 chapters)
1
Section 1: Introduction to Automated Machine Learning
5
Section 2: AutoML with Cloud Platforms
12
Section 3: Applied Automated Machine Learning

Introducing Microsoft NNI

Microsoft Neural Network Intelligence (NNI) is an open source platform that addresses the three key areas of any automated ML life cycle – automated feature engineering, architectural search (also referred to as neural architectural search or NAS), and hyperparameter tuning (HPI). The toolkit also offers model compression features and operationalization. NNI comes with many hyperparameter tuning algorithms already built in.

A high-level architecture diagram of NNI is as follows:

Figure 3.26 – Microsoft NNI high-level architecture

NNI has several state-of-the-art hyperparameter optimization algorithms built in, and they are called tuners. The list includes TPE, Random Search, Anneal, Naive Evolution, SMAC, Metis Tuner, Batch Tuner, Grid Search, GP Tuner, Network Morphism, Hyperband, BOHB, PPO Tuner, and PBT Tuner.

The toolkit is available on GitHub to be downloaded: https://github.com/microsoft/nni. More information...