Book Image

Hands-On One-shot Learning with Python

By : Shruti Jadon, Ankush Garg
Book Image

Hands-On One-shot Learning with Python

By: Shruti Jadon, Ankush Garg

Overview of this book

One-shot learning has been an active field of research for scientists trying to develop a cognitive machine that mimics human learning. With this book, you'll explore key approaches to one-shot learning, such as metrics-based, model-based, and optimization-based techniques, all with the help of practical examples. Hands-On One-shot Learning with Python will guide you through the exploration and design of deep learning models that can obtain information about an object from one or just a few training samples. The book begins with an overview of deep learning and one-shot learning and then introduces you to the different methods you can use to achieve it, such as deep learning architectures and probabilistic models. Once you've got to grips with the core principles, you'll explore real-world examples and implementations of one-shot learning using PyTorch 1.x on datasets such as Omniglot and MiniImageNet. Finally, you'll explore generative modeling-based methods and discover the key considerations for building systems that exhibit human-level intelligence. By the end of this book, you'll be well-versed with the different one- and few-shot learning methods and be able to use them to build your own deep learning models.
Table of Contents (11 chapters)
1
Section 1: One-shot Learning Introduction
3
Section 2: Deep Learning Architectures
7
Section 3: Other Methods and Conclusion

Understanding directed graphical models

We will now study directed graphical models briefly before we delve into probabilistic models for one-shot learning. A directed graphical model (also known as a Bayesian network) is defined with random variables connected with directed edges, as in the parent-child relationship. One such Bayesian network is shown in the following diagram:

The joint distribution over random variables in this graph S, R, L, W, and T can be broken into multiple distributions by a simple chain rule:

The conditional distributions on the right side of the preceding equation have a large number of parameters. This is because each distribution is conditioned on many variables and each conditioned variable has its own outcome space. This effect is even more prominent if we go further down in the graph when we have a huge set of conditioned variables. Consequently...