Rotary Embeddings in Encoder #2383
Unanswered
conceptofmind
asked this question in
Community | Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi all,
I was wondering how ColossalAI handled issues relating to length extrapolation for Rotary Embeddings and relative positional encoding for an encoder? From my understanding, Rotary Embeddings only work properly for the decoder since they do not length-generalize that well.
Thank you,
Enrico
Beta Was this translation helpful? Give feedback.
All reactions