Book Image

PyTorch Deep Learning Hands-On

By : Sherin Thomas, Sudhanshu Passi
Book Image

PyTorch Deep Learning Hands-On

By: Sherin Thomas, Sudhanshu Passi

Overview of this book

PyTorch Deep Learning Hands-On is a book for engineers who want a fast-paced guide to doing deep learning work with PyTorch. It is not an academic textbook and does not try to teach deep learning principles. The book will help you most if you want to get your hands dirty and put PyTorch to work quickly. PyTorch Deep Learning Hands-On shows how to implement the major deep learning architectures in PyTorch. It covers neural networks, computer vision, CNNs, natural language processing (RNN), GANs, and reinforcement learning. You will also build deep learning workflows with the PyTorch framework, migrate models built in Python to highly efficient TorchScript, and deploy to production using the most sophisticated available tools. Each chapter focuses on a different area of deep learning. Chapters start with a refresher on how the model works, before sharing the code you need to implement it in PyTorch. This book is ideal if you want to rapidly add PyTorch to your deep learning toolset.
Table of Contents (11 chapters)
10
Index

Introduction to recurrent neural networks

Recurrent neural networks (RNNs) are the de facto implementation for sequential data processing. As the name indicates, RNNs recur through the data holding the information from the previous run and try to find the meaning of the sequence, just like how humans do.

Although the vanilla RNN, the unrolling of a simple RNN cell for each unit in the input, was a revolutionary idea, it failed to provide production-ready results. The major hindrance was the long-term dependency problem. When the length of the input sequence is increased, the network won't be able to remember information from the initial units (words if it is natural language) when it reaches the last cells. We'll see what an RNN cell contains and how it can be unrolled in the upcoming sections.

Several iterations and years of research yielded a few different approaches to RNN architectural design. The state-of-the-art models now use long short-term memory (LSTM...