Book Image

The Deep Learning with PyTorch Workshop

By : Hyatt Saleh
Book Image

The Deep Learning with PyTorch Workshop

By: Hyatt Saleh

Overview of this book

Want to get to grips with one of the most popular machine learning libraries for deep learning? The Deep Learning with PyTorch Workshop will help you do just that, jumpstarting your knowledge of using PyTorch for deep learning even if you’re starting from scratch. It’s no surprise that deep learning’s popularity has risen steeply in the past few years, thanks to intelligent applications such as self-driving vehicles, chatbots, and voice-activated assistants that are making our lives easier. This book will take you inside the world of deep learning, where you’ll use PyTorch to understand the complexity of neural network architectures. The Deep Learning with PyTorch Workshop starts with an introduction to deep learning and its applications. You’ll explore the syntax of PyTorch and learn how to define a network architecture and train a model. Next, you’ll learn about three main neural network architectures - convolutional, artificial, and recurrent - and even solve real-world data problems using these networks. Later chapters will show you how to create a style transfer model to develop a new image from two images, before finally taking you through how RNNs store memory to solve key data issues. By the end of this book, you’ll have mastered the essential concepts, tools, and libraries of PyTorch to develop your own deep neural networks and intelligent apps.
Table of Contents (8 chapters)

Introduction

In the previous chapters, different network architectures were explained – from traditional ANNs, which can solve both classification and regression problems, to CNNs, which are mainly used to solve computer vision problems by performing the tasks of object classification, localization, detection, and segmentation.

In this final chapter, we will explore the concept of RNNs and solve sequential data problems. These network architectures are capable of handling sequential data where context is crucial, thanks to their ability to hold information from previous predictions, which is called memory. This means that, for instance, when analyzing a sentence word by word, RNNs have the ability to hold information about the first word of the sentence when they are handling the last one.

This chapter will explore the LSTM network architecture, which is a type of RNN that can hold both long-term and short-term memory and is especially useful for handling long sequences...