This project is about implementing attention mechanism.
- What is the attention mechanism?
- How to apply attention to RNNs
- What is a transformer?
- How to create an encoder-decoder transformer model
- What is GPT?
- What is BERT?
- What is self-supervised learning?
- How to use BERT for specific NLP tasks
- What is SQuAD? GLUE?
| File | Description |
|---|---|
| 0-rnn_encoder.py | RNNEncoder class that inherits from tensorflow.keras.layers.Layer to encode for machine translation. |
| 1-self_attention.py | SelfAttention class that inherits from tensorflow.keras.layers.Layer to calculate the attention for machine translation. |
| 2-rnn_decoder.py | RNNDecoder class that inherits from tensorflow.keras.layers.Layer to decode for machine translation:. |
| 4-positional_encoding.py | Calculates the positional encoding for a transformer. |
| 5-sdp_attention.py | Calculates the scaled dot product attention. |
| 6-multihead_attention.py | MultiHeadAttention class that inherits from tensorflow.keras.layers.Layer to perform multi head attention. |
| 7-transformer_encoder_block.py | EncoderBlock class that inherits from tensorflow.keras.layers.Layer to create an encoder block for a transformer. |
| 8-transformer_decoder_block.py | DecoderBlock class that inherits from tensorflow.keras.layers.Layer to create an encoder block for a transformer. |
| 9-transformer_encoder.py | Encoder class that inherits from tensorflow.keras.layers.Layer to create the encoder for a transformer. |
| 10-transformer_decoder.py | Decoder class that inherits from tensorflow.keras.layers.Layer to create the decoder for a transformer. |
| 11-transformer.py | Transformer class that inherits from tensorflow.keras.Model to create a transformer network. |