请问RWKV Context长度可以无限外推么。 #179
Answered
by
Triang-jyed-driung
BrightXiaoHan
asked this question in
Q&A
-
RWKV Context长度可以无限外推么,对于超出训练长度的文本有没有性能的损失呢? |
Beta Was this translation helpful? Give feedback.
Answered by
Triang-jyed-driung
Sep 5, 2023
Replies: 1 comment 1 reply
-
Hello! The context length of RWKV is theoretically infinity, while in reality, it is limited by the context length on which it was trained, as well as numerical precision, and the size of the hidden state. |
Beta Was this translation helpful? Give feedback.
1 reply
Answer selected by
BrightXiaoHan
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Hello! The context length of RWKV is theoretically infinity, while in reality, it is limited by the context length on which it was trained, as well as numerical precision, and the size of the hidden state.
As a result, you might encounter some performance decrease when the context length is longer than that being trained.
See 128k context length here:
https://huggingface.co/xiaol/rwkv-7B-world-novel-128k
(Possibly the open-source model with the longest context length)