Deep-dive on the seminal paper on Transformer models - "Attention is all you need" - https://arxiv.org/abs/1706.03762