-
Notifications
You must be signed in to change notification settings - Fork 231
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Benchmark] Support Video MCQ with TaskMeAnything-v1-video-random as an example #359
base: main
Are you sure you want to change the base?
Conversation
…LMEvalKit into add_tma_video
Hi, @weikaih04. We found that TaskMeAnything_v1_imageqa_random has a minor typo. The field of tsv file spelled category incorrectly, resulting in the final calculation result can only output the overall score, but not the category. |
@junming-yang Hi, thank you for pointing this out! I’ve uploaded new versions of ImageQA and VideoQA and changed the md5, it should work now. |
hey @junming-yang do you think it's ready to merge? |
Hi, @JieyuZ2 , |
Thanks! |
Hi,
For this PR
The usage of video_mcq.py is the same as image_mcq.py:
Just convert the benchmark to a TSV file, and encode the MP4 video to base64. I have provided the function named mp4_to_base64 in vlmeval/dataset/utils/video_mcq_utils.py.
I added the TaskMeAnything-v1-video-random video benchmark as an example for video_mcq.py and tested it on Paligemma (ImageQA model) and Video-LLaVA (VideoQA model), and it works well.