25 January 2023
Positional embeddings are key to the success of transformer models like BERT and GPT, but the way they work is often left unexplored. In this deep-dive, I want to break down the problem they're intended to solve and establish an intuitive feel for how they achieve it.