Stay tuned
.
References:
-
Transformers explained: https://e2eml.school/transformers.html ↩
-
Positional Embeddings: https://kazemnejad.com/blog/transformer_architecture_positional_encoding/ ↩
-
Positional Embeddings: https://timodenk.com/blog/linear-relationships-in-the-transformers-positional-encoding/ ↩
-
POS Summation v/s Concat: https://github.com/tensorflow/tensor2tensor/issues/1591 ↩