# How to Visualize Your Recurrent Neural Network with Attention in Keras

*A technical discussion and tutorial*

Neural networks are taking over every part of our lives. In particular — thanks to deep learning — Siri can fetch you a taxi using your voice; and Google can enhance and organize your photos automagically. Here at Datalogue, we use deep learning to structurally and semantically understand data, allowing us to prepare it for use automatically.

Neural networks are massively successful in the domain of computer vision. Specifically, convolutional neural networks(CNNs) take images and extract relevant features from them by using small windows that travel over the image. This understanding can be leveraged to identify objects from your camera (Google Lens) and, in the future, even drive your car (NVIDIA).

The analogous neural network for text data is the recurrent neural network (RNN). This kind of network is designed for sequential data and applies the same function to the words or characters of the text. These models are successful in translation (Google Translate), speech recognition (Cortana) and language generation.

Dr. Yoshua Bengio from Université de Montréal believes that language is going to be the next challenge for neural networks. At Datalogue, we deal with tons of text data, and we are interested in helping the community solve this challenge.

In this tutorial, we will write an RNN in Keras that can translate *human* dates (“November 5, 2016”, “5th November 2016”) into a standard format (“2016–11–05”). In particular, we want to gain some intuition into how the neural network did this. We will leverage the concept of *attention* to generate a map (like that shown in Figure 1) that shows which input characters were important in predicting the output characters.

### What is in this Tutorial

We start off with some technical background material to get everyone on the same page and then move on to programming this thing! Throughout the tutorial, I will provide links to more advanced content.

#### If you want to directly jump to the code:

#### What you need to know

To be able to jump into the coding part of this tutorial, it would be best if you have some familiarity with Python and Keras. You should have some familiarity of linear algebra, in particular that neural networks are just a few weight matrices with some non-linearities applied to it.

The intuition of RNNs and seq2seq models will be explained below.

### Recurrent Neural Networks (RNN)

An RNN is a function that applies the same transformation (known as the RNN ce*ll* or s*tep*) to every element of a sequence. The output of an RNN *layer* is the output of the RNN cell applied on each element of the sequence. In the case of text, these are usually successive words or characters. In addition to this, RNN cells hold an internal memory that summarize the history of the sequence it has seen so far.

The output of the RNN *layer* is an encoded sequence,

, that can be manipulated and passed into another network. RNNs are incredibly flexible in their inputs and outputs:*h*

- Many-to-One: use the complete input sequence to make a single prediction

.*h* - One-to-Many: transforming a single input to generate a sequence
*h**.* - Many-to-Many: transforming the entire input sequence into another sequence.

In theory, the sequences in our training data need not be of the same length. In practice, we pad or truncate them to be of the same length to take advantage of the static computational graph in Tensorflow.

We will focus on #3 “Many-to-Many” also known as sequence-to-sequence (**seq2seq**).

Long sequences can be difficult to learn from in RNNs due to instabilities in gradient calculations during training. To solve this, the RNN cell is replaced by a *gated cell* like the gated recurrent unit (GRU) or the long-short term memory cell (LSTM). To learn more about LSTMs and gated units, I highly recommend reading Christopher Olah’s blog (it’s where I first started understanding the RNN cell). From now on, whenever we talk about RNNs, we are talking about a gated cell. Since Olah’s blog gives an intuitive introduction to LSTMs, we will use that.

### A General Framework for seq2seq: The Encoder-Decoder Setup

Almost all neural network approaches to solving the seq2seq problem involve:

*Encoding*the input sentences into some abstract representation.*Manipulating*this encoding.*Decoding*it to our target sequence.

Our encoders and decoders can be any kind and combination of neural networks. In practice, most people use RNNs for both the encoders and decoders.

Figure 3 shows a simple encoder-decoder setup. The encoding step usually produces a sequence of vectors, *h**, *corresponding to the sequence of characters in the input date,

. In an RNN encoder, each vector is generated by integrating information from the past sequence of vectors.*x*

Before

is passed onto the decoder, we may choose to manipulate it in preparation for the decoder. For example, we might choose to only use the last encoding, as shown in Figure 4, since in theory, it is a summary of the whole sequence.*h*

Intuitively, this is similar to summarizing the whole input date into a single representation and then trying to decode that. While there may be enough information in this summary state for a classification problem like detecting sentiment (Many-to-One), it might be insufficient for an effective translation where it helps to consider the full sequence of hidden states (Figure 5).

However, this is not how humans translate dates: We do not read the whole text and then independently write down the translation at each character. Intuitively, a person would understand that the characters “Jan” correspond to 1st month, “5” corresponds to the day and “2016” corresponds to the year. As we have already seen in Figure 1, this idea known as *attention* can be captured by RNNs and has been applied successfully in caption generation (Xu et al. 2015), speech recognition (Chan et al. 2015) and indeed machine translation (Bahdanau et al. 2014). Most importantly, they produce *interpretable models*.

A slightly more visual example of how the attention mechanism works comes from the Xu et. al, 2015 paper (Figure 6). In the most complicated example of the girl with the teddy, we can see that when generating the word “girl”, the attention mechanism correctly focuses on the girl and not the teddy! Pretty brilliant. Not only does this make for pretty pictures, but it also let the authors diagnose problems in the model.

The creators of SpaCy have an in-depth overview of the encoder-attention-decoder paradigm. For other ways of modifying RNNs you can head over to this Distill article.

This tutorial will feature a single *bidirectional* LSTM as the encoder and the *attention* decoder (more on this later). More concretely, we will implement a simpler version of the model presented in *“Neural machine translation by jointly learning to align and translate.” (**Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio.**2014)*. I’ll walk through some of the math, but I invite you to jump into the appendices of the paper to get your hands dirty!

Now that we have learned about RNNs and the intuition behind the attention mechanism, let us learn how to implement it and subsequently obtain some nice visualizations. All code for subsequent sections is provided at `datalogue/keras-attention`

. The complete implementation of the model below is in `/models/NMT.py`

### The Encoder

Since we are trying to learn about AttentionRNNs, we will skip implementing our own vanilla RNN (LSTM) and use the one that ships with Keras. We can invoke it using:

BLSTM = Bidirectional(LSTM(encoder_units, return_sequences=True))

The parameter `encoder_units`

is the size of the weight matrix. We use `return_sequences=True`

here because we'd like to access the complete encoded sequence rather than the final summary state as described above.

Our BLSTM will consume the *characters *in the input sentence `x=(x1,...,xT)`

and output an encoded sequence `h=(h1,...,hT)`

where

is the number of characters in the date. Note that this is slightly different to the Bahdanau et al. paper where the sentence is a collection of words rather than characters. We also do not refer to the encoded sequence as the *T**annotations *like in the original paper.

### The Decoder

Now for the interesting part: the decoder. For any given character at position

in the sequence, our decoder accepts the encoded sequence *t*`h=(h1,...,hT)`

as well as the previous hidden state

(shared within the decoder cell) and character *st-1*

. Our decoder layer will output *yt-1*`y=(y1,...,yT)`

(the characters in the standardized date). Our overall architecture is summarized in Figure 7.

#### Equations

As shown in Figure 6, the decoder is quite complicated. So let’s break it down into the steps executed by the decoder cell when trying to predict character

In the following equations, the capital letter variables represent trainable parameters (Note that I have dropped the bias terms for brevity.)*t.*

- Calculate the attention probabilities
`α=(α1,…,αT)`

based on the encoded sequence and the internal hidden state of the decoder cell,*st-1**.*These are shown in Equation 1 and Equation 2.

2. Calculate the *context *vector which is the weighted sum of the encoded sequence with the attention probabilities. Intuitively, this vector summarizes the importance of the different encoded characters in predicting the

-th character.*t*

3. We then update our hidden state. If you are familiar with the equations of an LSTM cell, these might be ring a bell as the reset gate,

, update gate, *r*

, and the proposal state. We use the reset gate to control how much information from the previous hidden state *z**st-1** *is used to create a proposal hidden state. The update gate controls how we much of the proposal we use in the new hidden state

. (Confused? See step by step walk through of LSTM equations)*st*

4. Calculate the

-th character using a simple one layer neural network using the context, hidden state, and previous character. This is a modification from the paper which uses a *t**maxout* layer. Since we are trying to keep things as simple as possible this works fine!

Equations 1–8 are applied for every character in the encoded sequence to produce a decoded sequence

which represents the probability of a translated character at each position.*y*

#### Code

Our custom layer is implemented in `models/custom_recurrent.py`

. This part is somewhat complicated in particular because of the manipulations we need to make for vectorized operations acting on the complete encoded sequence. It will make more sense when you think about it. I promise it will become easier the more you look at the equations and the code simultaneously.

A minimal custom Keras layer has to implement a few methods: `__init__`

, `compute_ouput_shape`

, `build`

and `call`

. For completeness, we also implement `get_config`

which allows you to load the model back into memory easily. In addition to these, a Keras recurrent layer implements a `step`

method that will hold all the computations of our cell.

First let us break down boiler-plate layer code:

`__init__`

is what is called when the Layer is first instantiated. It sets functions that will eventually initialize the weights, regularizers, and constraints. Since the output of our layer is a sequence, we hard code`self.return_sequences=True`

.`build`

is called when we run`Model.compile(…)`

. Since our model is quite complicated, you can see there are a ton of weights to initialize here. The call`self.add_weight`

automatically handles initializing the weights and setting them as trainable within the model. Weights with the subscript`a`

are used to calculate the context vector (step 1 and 2). Weights with the subscript`r`

,`z`

,`p`

will calculate the new hidden states from step 3. Finally, weights with the subscript`o`

will calculate the output of the layer.- Some convenience functions are implemented as well: (a)
`compute_output_shape`

will calculate output shapes for any given input; (b)`get_config`

let’s us load the model using just a saved file (once we are done training)

Now for the cell logic:

- By default, each execution of the cell only has information from the previous time step. Since we need to access the entire encoded sequence within the cell, we need to save it somewhere. Therefore, we make a simple modification in
`call`

. The only way I could find to do this was to set the sequence being fed into the cell as`self.x`

so that we can access it later:

Now we need to think vectorized: The `_time_distributed_dense`

function calculates the last term of Equation 1 for all the elements of the encoded sequence.

- We now walk through the most important part of the code, that is in
`step`

which executes the cell logic. Recall that`step`

is applied to every element of the input sequence.

In this cell we want to access the previous character `ytm`

and hidden state `stm`

which is obtained from `states`

in line 4.

Think vectorized: we manipulate `stm`

to repeat it for the number of characters we have in our input sequence.

On lines 11–18 we implement a version of equation 1 that does the calculations on all the characters in the sequence at once.

In lines 24–28 we have implemented Equation 2 in the vectorized form for the whole sequence. We use `repeat`

to allow us to divide every time step by the respective sums.

To calculate the context vector, we need to keep in mind that `self.x_seq`

and `at`

have a “batch dimension” and therefore we need to use `batch_dot`

to avoid doing the multiplication over that dimension. The squeeze operation just removes left-over dimensions. This is done in lines 33–37.

The next few lines of code are a more straightforward implementation of equations 4 –8.

Now we think a bit ahead: We would like to calculate those fancy attention maps in Figure 1. To do this, we need a “toggle” that returns the attention probabilities `at`

.

### Training

#### Data

Any good learning problem should have training data. In this case, it’s easy enough thanks to the Faker library which can generate fake dates with ease. I also use the Babel library to generate dates in different languages and formats as inspired by rasmusbergpalm/normalization. The script `data/generate.py`

will generate some fake data and I won’t bore you with the details but invite you to poke around or to make it better.

The script also generates a vocabulary that will convert characters into integers so that the neural network can understand it. Also included is a script in `data/reader.py`

to read and prepare data for the neural network to consume.

#### Model

This simple model with a bidirectional LSTM and the decoder we wrote above is implemented in `models/NMT.py`

. You can run it using `python run.py`

where I have set some default arguments (Readme has more information). I’d recommend training on a machine with a GPU as it can be prohibitively slow on a CPU-only machine.

If you want to skip the training part, I have provided some weights in `weights/`

### Visualization

Now the easy part. For the visualizer implemented in `visualizer.py`

, we need to load the weights in twice: Once with the predictive model, and the other to obtain the probabilities. Since we implemented the model architecture in `models/NMT.py`

we can simply call the function twice:

from models.NMT import simpleNMT

predictive_model = simpleNMT(...)

predictive_model.load_weights(...,return_probabilities=False)

probability_model = simpleNMT(...,return_probabilities=True)

probability_model.load_weights(...)

To simply use the implemented visualizer, you can type:

`python visualizer.py -h`

to see available command line arguments.

#### Example visualizations

Let us now examine the attentions we generated from `probability_model`

. The `predictive_model`

above returns the translated date that you see on the y-axis. On the x-axis is our input date. The map shows us which input characters (on the x-axis) were used in the prediction of the output character on the y-axis. The brighter the white, the more weight that character had. Here are some I thought were quite interesting.

A correct example which doesn’t pay attention to unnecessary information like days of the week:

And here is an example of an incorrect translation because we submitted our example text in a novel order: “January 2016 05” was translated to “2016–01–02” rather than “2016–01–05”.

We can see that the model has mistakenly interpreted the characters “20” from 2016 as the day of the month. The activations are weak (some even correctly on the actual date of “5”) which can provide us with some insight into how to better train this model.

### Conclusion

I hope this tutorial has shown you how to solve a machine learning problem from start to finish. Furthermore, I hope it helps you in trying to visualize seq2seq problems with recurrent neural networks. If there is something I missed or you think I could do better, please feel free to reach out via twitter or make an issue on our github, I’d love to chat!

### Acknowledgements

Zaf is an intern at Datalogue partially supported by an NSERC Experience Award. The Datalogue Team contributed for code review and reading the proofs of this post.