PyTorch re-implementation of transformers.
The transformer_layers.py
file contains the main components of the transformer architecture.
The class Self_Attention
defines the self-attention layer, with the possibility to adapt the attention mask to variable input sequences with padding (valid_lens
), or stop the tokens from attending future tokens (forward_mask
variable).
The class MultiHeadAttention
creates multiple attention heads in parallel.
The class FeedForward
defines a linear layer followed by non-linearity applied to each encoded token (through the self-attention layer).
The class Block
defines a whole transformer block, containing a multi-head attention, a feedforward layer, and layer-normalization layer.
Finally, a class PositionalEncoding
creates the positional encoding necessary to provide sequence information concatenated to the input to the transformer.
The train_GPT.py
file trains the transformer architecture on a GPT-like task, consisting of predicting the next token given a sequence of tokens. After training, the GPT network should be able to generate Shakespeare-like text (use generate_GPT.py
).
The seq2seq.ipynb
notebook uses the transformer architecture for sequence to sequence learning (e.g., translation). Here, a simple sequence to reversed-sequence is implemented to test whether the model works. A python script version is provided in train_seq2seq_reverse.py
. I also test the model on a more complex task, i.e., french-to-english translation. The code is provided in train_seq2seq_translate.py
.
To test the trained model, use generate_seq2seq_reverse.py
for sequence reversal and generate_seq2seq_translate.py
.
To adapt the transformer architecture for this task, we provide additional classes for cross attention in the file transformer_layers.py
:
The class Cross_Attention
defines the cross-attention layer between the encoder output and the decoder embedded tokens. It adapts the cross attention mask to the variable input and target sequences (valid_lens_x
and valid_lens_y
).
The class MultiHeadCrossAttention
uses the cross attention with multiple heads.
The class Decoder_Block
stacks self-attention, cross-attention and feedforward layers into a decoder-block.
The class Seq2Seq
constructs the encoder decoder transformer with cross attention layers.
I also provide the code to predict the embedding on a pre-trained GPT2 from Huggingface (predictions_gpt2.py
). The embeddings from our own train transformer model are provided by the file predictions_seq2seq_translate.py
, using hooks to save activations.