{"@context":"https://schema.org","@type":"CreativeWork","@id":"https://forgecascade.org/public/capsules/01894394-acc2-4b29-8764-f7a9bc5d3b46","identifier":"01894394-acc2-4b29-8764-f7a9bc5d3b46","url":"https://forgecascade.org/public/capsules/01894394-acc2-4b29-8764-f7a9bc5d3b46","name":"Rotary Position Embedding (RoPE): Relative Position via Rotation","text":"RoPE (Su et al. 2021) encodes position via rotation matrices applied to query/key pairs. Relative position is automatically captured in the dot product. Used in LLaMA, PaLM, Mistral, Qwen. Enables length generalization beyond training context via NTK-aware scaling.","keywords":["rope","position-encoding","llama","transformers"],"about":[],"citation":[],"isPartOf":{"@type":"Dataset","name":"Forge Cascade Knowledge Graph","url":"https://forgecascade.org"},"publisher":{"@type":"Organization","name":"Forge Cascade","url":"https://forgecascade.org"},"dateCreated":"2026-04-12T05:54:52.018503Z","dateModified":"2026-05-09T01:48:30.979964Z","additionalProperty":[{"@type":"PropertyValue","name":"trust_level","value":40},{"@type":"PropertyValue","name":"verification_status","value":"unverified"},{"@type":"PropertyValue","name":"provenance_status","value":"valid"},{"@type":"PropertyValue","name":"evidence_level","value":"ungraded"},{"@type":"PropertyValue","name":"content_hash","value":"df280964aa127d9ad0fab7001bde7aee63179d3b0b7d4ef5da3883269bf6a9f7"}]}