Understanding the Self-Attention Mechanism in Transformers | Tidily.ai