Attention Intuition
Imagine we are sitting in a room. We have a red cup of coffee, and a notebook in front of us. When we first sit down, the red cup stands out. So it attracts our attention “involuntarily” to notice the red cup first.
After drinking the coffee, we tell ourselves that “we need to focus on the notebook now”. So we voluntarily and consciously pull our attention to the notebook. Because we are consciously doing it, the attention strength is stronger.
Query-Key-Value (QKV)
When objects enter a machine eye, in our head, they will have a key (a short code), a value (e.g., their pixel values). Based on the machine brain’s “voluntary attention”, the brain will issue a query “what should I see if I want to work?”. They query will be run through all objects’ keys, and based on their similarity (or relavance), each object’s value get assigned to a relavance score, then gets added up, and outputted as the combined “attention”.
More formally, the combined attention is
\[\begin{gather*} f(q, k1, v1, ...) = \sum_i \alpha(q, k_i) v_i \end{gather*}\]where the attention weight $\alpha_i$ for the ith
key value pair is:
Now, let’s talk about how to calculate the attention score a(q, k_i)
. There are two types: additive attention, and scaled dot-product attention.
Additive (Bahdanau) Attention
When keys and the query have different lengths, we can use the additive attention. Additive attention projects keys and the query into the same length using two linear layers.
The above can be implemented as a single multi-layer perceptron. Below is from the seq2seq tutorial on PyTorch
- key vector
k
isdk
long - query vector
q
isdq
long - Say we have
h
as hidden dimension - Learnable weight matrices $W_v$
(h, 1)
, $W_k$(h, dk)
, $W_q$(h, q)
score how weighted queries and keys match with each other.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
from torch import nn
import torch
class BahdanauAttention(nn.Module):
def __init__(self, key_size, query_size, hidden_size, dropout_p):
super().__init__()
self.Wk = nn.Linear(key_size, hidden_size, bias=False)
self.Wq = nn.Linear(query_size, hidden_size, bias=False)
self.Wv = nn.Linear(hidden_size, 1, bias=False) # a vector
self.dropout = nn.Dropout(dropout_p)
def forward(self, queries, keys, values):
"""
queries: (batch_size, query_num, query_size)
keys: (batch_size,total_num_key_value_pairs, key_size)
values: (batch_size,total_num_key_value_pairs, value_size)
"""
# Project queries and keys onto the same hidden dim
queries = self.Wq(queries) # (batch_size, query_num, hidden_size)
keys = self.Wk(keys) # (batch_size,total_num_key_value_pairs,hidden_size)
# Broadcasting to add queries and keys together
queries = queries.unsqueeze(2) # (batch_size, query_num, 1, hidden_size)
keys = keys.unsqueeze(1) # (batch_size, 1, total_num_key_value_pairs, hidden_size)
features = queries + keys # (batch_size, query_num, total_num_key_value_pairs, hidden_size)
features = torch.tanh(features)
scores = self.Wv(features) # (batch_size, query_num, total_num_key_value_pairs, 1)
scores = scores.squeeze(-1) # (batch_size, query_num, total_num_key_value_pairs)
# Use masked_softmax here with a pre-designated length
self.attention_weights = nn.functional.softmax(scores)
# torch.bmm is batch-matrix-multiplication
# (batch_size, query_num, value_size), so we get all queries, weighted
attention = torch.bmm(self.dropout(self.attention_weights), values)
return attention
value_size = 2
key_size = 3
query_size = 4
hidden_size = 5
attention = BahdanauAttention(key_size=key_size, query_size=query_size, hidden_size=hidden_size, dropout_p=0.1)
batch_size = 1
query_num = 2
total_num_key_value_pairs = 3
torch.manual_seed(42)
queries = torch.rand((batch_size, query_num, query_size))
keys = torch.rand((batch_size, total_num_key_value_pairs, key_size))
values = torch.rand((batch_size,total_num_key_value_pairs, value_size))
attention(queries, keys, values)
Scaled Dot-Product (Luong) Attention
When keys and queries do have the same length, dot-multiplying them together is faster to give a “relavance” score. Assume Queries is num_queries x hidden_length (d)
, keys key_pair_num x hidden_length
, values key_pair_num x value_length
. Below, we denote the length of keys hidden_length
as $d_k$
Note that if every pair of elements in keys and queries are independent with [mean=0, var=1]
, their product $QK^T$ has a zero mean, and a variance d
. We normalize this product and choose it to be our attention score a
, so its variance is always 1.
The attention mask, a.k.a look-ahead mask M
is applied before softmax, after calculating the attention score a
(not shown in the illustration). In the attention is all you need paper, the look-ahead mask is applied to make sure future words are not considered in the current step. The “attention is all you need” paper worded this point only with the basic intent, which I found confusing at the first time 😭
1
"We need to prevent leftward information flow in the decoder to preserve the auto-regressive property. We implement this inside of scaled dot-product attention by masking out (setting to −∞) all values in the input of the softmax which correspond to illegal connections"
Then the attention weight is:
\[\begin{gather*} \alpha = softmax(\frac{QK^T}{\sqrt{d_k}} + M)V \end{gather*}\]All together the code looks like:
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
class DotProductAttention(torch.nn.Module):
def forward(
self,
q: torch.Tensor,
k: torch.Tensor,
v: torch.Tensor,
attn_mask: torch.Tensor = None,
key_padding_mask: torch.Tensor = None,
):
"""
Args:
q (torch.Tensor): [batch_size, query_num, qk_dim] or [batch_size, head_num, query_num, qk_dim]
k (torch.Tensor): [batch_size, kv_num, qk_dim] or [batch_size, head_num, query_num, qk_dim]
v (torch.Tensor): [batch_size, kv_num, v_dim] or [batch_size, head_num, query_num, qk_dim]
attn_mask (torch.Tensor): or look-ahead mask, [query_num, kv_num]. 1 means "mask out"
Later, they are multiplied by large negative values -1e9. so values can be ignored in softmax.
key_padding_mask (torch.Tensor): [batch_size, kv_num]. 1 means "mask out"
Returns:
attention: [batch_size, query_num, v_dim] or [batch_size, head_num, query_num, qk_dim]
"""
q_kT_scaled = (q @ k.transpose(-2, -1)) / torch.sqrt(
torch.tensor(k.shape[-1], dtype=torch.float32)
)
if attn_mask is not None:
q_kT_scaled.masked_fill_(attn_mask.bool(), float("-inf"))
if key_padding_mask is not None:
key_padding_mask = key_padding_mask.unsqueeze(1)
if q_kT_scaled.ndim == 4:
key_padding_mask = key_padding_mask.unsqueeze(2)
# [batch_size, query_num, kv_num]
q_kT_scaled = q_kT_scaled.masked_fill(
key_padding_mask,
float("-inf"),
)
attention_weight = torch.nn.functional.softmax(q_kT_scaled, dim=-1)
attention = attention_weight @ v
# TODO In this implementation, there's a drop out
# https://ricojia.github.io/2022/03/27/deep-learning-attention-mechanism/#scaled-dot-product-luong-attention
return attention
Look-Ahead-Mask
Omits Paddding For Each Query’s Attention
At timestep t
, Padding at in an input sentence start at t
. When we train the transformer for a translation task, the decoder gets its last output as its input. So at t
, we want to omit <PAD>
from then on.
The look ahead mask looks like:
\[\begin{gather*} \begin{bmatrix} 0 & -10^{-9} & -10^{-9}& -10^{-9} & \dots& -10^{-9} \\ 0 & 0 & -10^{-9} & -10^{-9} & \dots & -10^{-9} \\ 0 & 0 & 0 & -10^{-9} & \dots & -10^{-9} \\ \vdots \\ 0 & 0 & 0 & 0 & \dots & 0 \\ \end{bmatrix} \end{gather*}\]After Applying an look-ahead mask, the attention we get is a weighted sum of all rows in $V$:
\[\begin{gather*} softmax(\frac{QK^T}{\sqrt{d_k}}) V = \begin{bmatrix} q_1^T k_1 v_1 \\ q_2^T k_1 v_1 + q_2^T k_2 v_2\\ \vdots \\ q_n^T k_1 v_1 + q_n^T k_2 v_2 + \dots + q_n^T k_2 v_2 \\ \end{bmatrix} \end{gather*}\]- Note that the output attention is
[num_queries, value_dimension]
. In decoder, only the first self-attention needs this look-ahead mask. The attention’s inputs are outputs from the last timesteps, sonum_queries=sentece_length
. So effectively, the output attention at each timestep does not consider outputs from a later timestep. Later timesteps are just padding, which will pollute our attention
Visualization of Attention
One great feature about attention is its visibility. Below is an example from the PyTorch NLP page
The input sentence is “il n est pas aussi grand que son pere”.
To interpret:
- When outputting “he”, most attention was given to “il”, “n”, “est”
- When outputting “is”, most attention was given to “aussi”, “grand”, “que” (which is interesting because
is
should beest
) - When outputting “not”, most attention was given to “aussi”, “pas”, “que”
- The output “his father” focuses on “son père,” which matches the intended translation.
Bahdanau Encoder-Decoder Structure
Click to expand
In 2014, Bahdanau et al. proposed an encoder-decoder structure **on top of the additive attention**. To illustrate, we have a neural machine translation example (NMT): **translate French input "Jane visite l'Afrique en septembre" to English**. For attention pooling, we talked about scaled dot-product attention pooling and additive attention pooling in the previous sections. - Encoder: we are using a [bi-directional RNN encoder](../2022/2022-03-15-deep-learning-rnn3-lstm.markdown) to generate embeddings of french sentences. Now, our input "Jane visite l'Afrique en septembre" will complete its forward and backward passes. - At each time `t`, the bidirectional RNN encoder outputs **a hidden state** $a^{(t)}$ (which is the key and value at the same time.) - $\alpha^{(t, t')}$: amount of attention output at time `t`, $y^{(t)}$ should put to hidden state at time `t'`, $a^{(t)}$ - Decoder we have **another single-drectional RNN decoder** to generate the word probabilities in the vocab space. - Here, we denote the hidden states as $s^{(t)}$. That's the **query** - Before outputting `