Rotary Position Embedding (RoPE) is a technique that encodes positional information in transformers by rotating query and key vectors. Unlike earlier approaches that mixed position and semantic information, RoPE applies different rotation angles based on token distance and hidden state dimensions. Tokens closer together receive

8m read timeFrom towardsdatascience.com
Post cover image
Table of contents
IntroductionRoPE IntuitionBefore RoPERotation IntuitionAngle of RotationThe Angle FormulaConclusion

Sort: