Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Correct the configuration of LLaVA-CoT #705

Merged
merged 1 commit into from
Dec 31, 2024

Conversation

XuGW-Kevin
Copy link
Contributor

@XuGW-Kevin XuGW-Kevin commented Dec 31, 2024

First of all, we sincerely appreciate the work of VLMEvalKit and the tremendous contributions it has made to the entire VLM community! However, the current configuration of LLaVA-CoT (e.g., max_new_tokens) is incorrect, leading to significant deviations in the benchmark test results. This PR aims to correct the configuration of LLaVA-CoT.

@kennymckormick kennymckormick merged commit 6e1a59a into open-compass:main Dec 31, 2024
1 check failed
kennymckormick pushed a commit to TobiasLee/VLMEvalKit that referenced this pull request Jan 1, 2025
kennymckormick added a commit that referenced this pull request Jan 1, 2025
* update vlrewardbench

* pre-commit fix

* formatter

* [Improvement] Better `AUTO_SPLIT` and model split for InternVL2

* [Minor] Improve CC-OCR Import

* [Model] Support QVQ

* [Model] Update Molmo Eval to Match Official Implementation (#648)

* add molmo prompts

* fix lint format

* [Fix] Refine Qwen-VL2 device assignment

* [Fix] Fix RealWorldQA md5

* update MMMU_DEV_VAL tsv

* [Fix] Fix confusing image width&height (#704)

Co-authored-by: Yuan Ye <[email protected]>

* Update llama_vision.py (#705)

* [Fix] Fix Lint

* Fix Lint

* Fix Lint

---------

Co-authored-by: kennymckormick <[email protected]>
Co-authored-by: jamespark3922 <[email protected]>
Co-authored-by: CMeteor <[email protected]>
Co-authored-by: Yuan Ye <[email protected]>
Co-authored-by: Guowei Xu <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants