I propose to add a question on implementing a self attention block with rotary positional embedding. Link: [RoFormer: Enhanced Transformer with Rotary Position Embedding](https://arxiv.org/abs/2104.09864 ) (Not sure if this question is already part of some PR)