Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Problems with mteb_meta for german evaluation #847

Open
achibb opened this issue May 29, 2024 · 6 comments
Open

Problems with mteb_meta for german evaluation #847

achibb opened this issue May 29, 2024 · 6 comments
Labels
bug Something isn't working

Comments

@achibb
Copy link
Contributor

achibb commented May 29, 2024

Hi everyone, I am having troubles with generating the mteb_meta for German, with just running the script.

I am currently trying to format results but it does not seem to work straight away with "mteb_meta.py" - any idea? I just get a blank metadata file:

tags:

mteb
model-index:
name: gbert-large

results:

it gives this for every dataset:

WARNING:mteb.evaluation.MTEB:Passing task names as strings is deprecated and will be removed in the next release. Please use tasks = mteb.get_tasks(tasks=[...]) method to get tasks instead.
INFO:main:Skipping AmazonCounterfactualClassification as split test not present.
WARNING:mteb.evaluation.MTEB:Passing task names as strings is deprecated and will be removed in the next release. Please use tasks = mteb.get_tasks(tasks=[...]) method to get tasks instead.
INFO:main:Skipping AmazonReviewsClassification as split test not present.

Do I need to modify something on the code?

@imenelydiaker
Copy link
Contributor

imenelydiaker commented May 29, 2024

Due to recent updates #826 and #806:

For the WARNING message, if you're using a python script then your code should look like this:

import mteb
from sentence_transformers import SentenceTransformer

# Define the sentence-transformers model name
model_name = "average_word_embeddings_komninos"
# or directly from huggingface:
# model_name = "sentence-transformers/all-MiniLM-L6-v2"

model = SentenceTransformer(model_name)
tasks = mteb.get_tasks(tasks=["Banking77Classification"])
evaluation = mteb.MTEB(tasks=tasks)
results = evaluation.run(model, output_folder=f"results/{model_name}")

@imenelydiaker
Copy link
Contributor

imenelydiaker commented May 29, 2024

Else for these 2 messages:

INFO:main:Skipping AmazonCounterfactualClassification as split test not present.
INFO:main:Skipping AmazonReviewsClassification as split test not present.

It's likely a bug on our side, we'll check, thank you for reporting!

@imenelydiaker imenelydiaker added the bug Something isn't working label May 29, 2024
@KennethEnevoldsen
Copy link
Contributor

KennethEnevoldsen commented May 29, 2024

@imenelydiaker I believe it is due to the new results format introduced in #759. mteb_meta.py will need to be rewritten for the new format.

We should probably make it a CLI with a test (otherwise it is impossible to know if it breaks).

@KennethEnevoldsen
Copy link
Contributor

@achibb, we have currently updated the CLI and well as the benchmark lists. I believe the new CLI should suit your purpose

@achibb
Copy link
Contributor Author

achibb commented Jun 11, 2024 via email

@imenelydiaker
Copy link
Contributor

imenelydiaker commented Jun 11, 2024

Thank you very much! Will test the next days and feedback. I was wondering can I also compute something for the German benchmark for other models like mdeberta, and somehow add it to the leaderboard ?

Yes you can evaluate any model and submit results to this repo via PR so they can be added to the leaderboard (check the guide on opening a PR on HF here).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

3 participants