Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

baseline score #84

Open
Lorraine-Kwok opened this issue Apr 23, 2024 · 3 comments
Open

baseline score #84

Lorraine-Kwok opened this issue Apr 23, 2024 · 3 comments

Comments

@Lorraine-Kwok
Copy link

Having a reference point for the baseline model's scores would be incredibly beneficial for my team and me as we develop our approach and compare its performance. If it's possible, could you please provide us with the baseline model's scores on the validation dataset, or direct us to where we could find this information?

@ChonghaoSima
Copy link
Contributor

Please see here

@Lorraine-Kwok
Copy link
Author

Thank you

@Lorraine-Kwok
Copy link
Author

Hi,

I've been looking at the language scores reported in your results and noticed that the baseline's language score is much lower compared to the GPT's high score. Additionally, the GPT scores for both sampled data and test data are almost identical, leading to very similar final scores.

Could you shed some light on the following?

Why is there such a big difference in language scores between the baseline and GPT?
How can the GPT scores be so close for sampled and test data?
It seems a bit odd, and I'm trying to make sense of it. Any clarification would be appreciated.

Thanks!

@Lorraine-Kwok Lorraine-Kwok reopened this Apr 24, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants