Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

gpt模型的注意力头数n_head会影响模型在哪些方面的性能?怎样有效提升模型处理多轮对话的准确率? #98

Open
oerifjmerefver opened this issue Sep 5, 2022 · 0 comments

Comments

@oerifjmerefver
Copy link

oerifjmerefver commented Sep 5, 2022

我增加了隐藏层n_layer的数量,模型处理单轮对话的准确率提升了很多,但是处理多轮对话的准确率没有多少提升
然后我注意到了注意力头数n_head,但是我并不确定它能否有效提升多轮对话的准确率。我尝试将它从12修改为96,发现准确率依然没有特别的提升

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant