Skip to content

Add VL-RewardBench#703

Merged
kennymckormick merged 15 commits intoopen-compass:mainfrom
TobiasLee:vlrewardbench
Jan 1, 2025
Merged

Add VL-RewardBench#703
kennymckormick merged 15 commits intoopen-compass:mainfrom
TobiasLee:vlrewardbench

Conversation

@TobiasLee
Copy link
Copy Markdown
Contributor

Hi there,

Thanks for your awesome project, which helps a lot for LMM evaluation & development!

This PR incorporates our recently released VL-RewardBench.
Example script:

python run.py --data VL-RewardBench --model GPT4o 

Saved results for GPT4O-MINI:

"hallucination","reasoning","general","Macro Accuracy","Overall Consistency"
"0.4552736982643525","0.6477987421383647","0.4371584699453552","0.5134103034493575","0.5016"

and GPT4O:

"hallucination","reasoning","general","Macro Accuracy","Overall Consistency"
"0.7076101468624834","0.6509433962264151","0.4918032786885246","0.616785607259141","0.6616"

The results are consistent with our reported with small variance.

@kennymckormick kennymckormick merged commit 276d90a into open-compass:main Jan 1, 2025
@kennymckormick
Copy link
Copy Markdown
Member

Evaluation Results of GPT4o-20241120


hallucination 0.753004
reasoning 0.676101
general 0.535519
Macro Accuracy 0.654875
Overall Consistency 0.7016


Mercury7353 pushed a commit to Mercury7353/VLMEvalKit that referenced this pull request Apr 28, 2025
* update vlrewardbench

* pre-commit fix

* formatter

* [Improvement] Better `AUTO_SPLIT` and model split for InternVL2

* [Minor] Improve CC-OCR Import

* [Model] Support QVQ

* [Model] Update Molmo Eval to Match Official Implementation (open-compass#648)

* add molmo prompts

* fix lint format

* [Fix] Refine Qwen-VL2 device assignment

* [Fix] Fix RealWorldQA md5

* update MMMU_DEV_VAL tsv

* [Fix] Fix confusing image width&height (open-compass#704)

Co-authored-by: Yuan Ye <yuany2@chinatelecom.cn>

* Update llama_vision.py (open-compass#705)

* [Fix] Fix Lint

* Fix Lint

* Fix Lint

---------

Co-authored-by: kennymckormick <dhd.efz@gmail.com>
Co-authored-by: jamespark3922 <jspark96@cs.washington.edu>
Co-authored-by: CMeteor <CMeteor@users.noreply.github.com>
Co-authored-by: Yuan Ye <yuany2@chinatelecom.cn>
Co-authored-by: Guowei Xu <113534787+XuGW-Kevin@users.noreply.github.com>
Koii2k3 pushed a commit to wjnwjn59/VLMEvalKit that referenced this pull request Nov 13, 2025
* update vlrewardbench

* pre-commit fix

* formatter

* [Improvement] Better `AUTO_SPLIT` and model split for InternVL2

* [Minor] Improve CC-OCR Import

* [Model] Support QVQ

* [Model] Update Molmo Eval to Match Official Implementation (open-compass#648)

* add molmo prompts

* fix lint format

* [Fix] Refine Qwen-VL2 device assignment

* [Fix] Fix RealWorldQA md5

* update MMMU_DEV_VAL tsv

* [Fix] Fix confusing image width&height (open-compass#704)

Co-authored-by: Yuan Ye <yuany2@chinatelecom.cn>

* Update llama_vision.py (open-compass#705)

* [Fix] Fix Lint

* Fix Lint

* Fix Lint

---------

Co-authored-by: kennymckormick <dhd.efz@gmail.com>
Co-authored-by: jamespark3922 <jspark96@cs.washington.edu>
Co-authored-by: CMeteor <CMeteor@users.noreply.github.com>
Co-authored-by: Yuan Ye <yuany2@chinatelecom.cn>
Co-authored-by: Guowei Xu <113534787+XuGW-Kevin@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants