Book Image

Deep Learning for Natural Language Processing

By : Karthiek Reddy Bokka, Shubhangi Hora, Tanuj Jain, Monicah Wambugu
Book Image

Deep Learning for Natural Language Processing

By: Karthiek Reddy Bokka, Shubhangi Hora, Tanuj Jain, Monicah Wambugu

Overview of this book

Applying deep learning approaches to various NLP tasks can take your computational algorithms to a completely new level in terms of speed and accuracy. Deep Learning for Natural Language Processing starts by highlighting the basic building blocks of the natural language processing domain. The book goes on to introduce the problems that you can solve using state-of-the-art neural network models. After this, delving into the various neural network architectures and their specific areas of application will help you to understand how to select the best model to suit your needs. As you advance through this deep learning book, you’ll study convolutional, recurrent, and recursive neural networks, in addition to covering long short-term memory networks (LSTM). Understanding these networks will help you to implement their models using Keras. In later chapters, you will be able to develop a trigger word detection application using NLP techniques such as attention model and beam search. By the end of this book, you will not only have sound knowledge of natural language processing, but also be able to select the best text preprocessing and neural network models to solve a number of NLP issues.
Table of Contents (11 chapters)

Training a Neural Network

So far, we know that once an input is provided to a neural network, it enters the input layer which is an interface that exists to pass on the input to the next layer. If a hidden layer is present, then the inputs are sent to the activation nodes of the hidden layer via weighted connections. The weighted sum of all the inputs received by the activations nodes is calculated by multiplying the inputs with their respective weights and adding these values up along with the bias. The activation function generates an activation value from the weighted sum and this is passed on to the nodes in the next layer. If the next layer is another hidden layer, then it uses the activation values from the previous hidden layer as inputs and repeats the activation process. However, if the proceeding layer is the output layer, then the output is provided by the neural network.

From all of this information, we can conclusively say that there are three parts of the deep learning model...