-
Notifications
You must be signed in to change notification settings - Fork 593
[VLM] Accuracy Evaluation #2393
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[VLM] Accuracy Evaluation #2393
Conversation
|
MLCommons CLA bot All contributors have signed the MLCommons CLA ✍️ ✅ |
multimodal/vl2l/src/mlperf_inference_multimodal_vl2l/evaluation.py
Outdated
Show resolved
Hide resolved
multimodal/vl2l/src/mlperf_inference_multimodal_vl2l/evaluation.py
Outdated
Show resolved
Hide resolved
multimodal/vl2l/src/mlperf_inference_multimodal_vl2l/evaluation.py
Outdated
Show resolved
Hide resolved
multimodal/vl2l/src/mlperf_inference_multimodal_vl2l/evaluation.py
Outdated
Show resolved
Hide resolved
multimodal/vl2l/src/mlperf_inference_multimodal_vl2l/evaluation.py
Outdated
Show resolved
Hide resolved
|
As it stands, |
|
The results were after including |
multimodal/vl2l/src/mlperf_inference_multimodal_vl2l/evaluation.py
Outdated
Show resolved
Hide resolved
multimodal/vl2l/src/mlperf_inference_multimodal_vl2l/evaluation.py
Outdated
Show resolved
Hide resolved
multimodal/vl2l/src/mlperf_inference_multimodal_vl2l/evaluation.py
Outdated
Show resolved
Hide resolved
multimodal/vl2l/src/mlperf_inference_multimodal_vl2l/evaluation.py
Outdated
Show resolved
Hide resolved
multimodal/vl2l/src/mlperf_inference_multimodal_vl2l/evaluation.py
Outdated
Show resolved
Hide resolved
multimodal/vl2l/src/mlperf_inference_multimodal_vl2l/evaluation.py
Outdated
Show resolved
Hide resolved
We should try on the actual |
wangshangsam
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM! Thanks a lot, John!
|
This PR seems ready. @hanyunfan @mrmhodak @arjunsuresh I'm wondering if you could take a look, approve and merge it? Thanks! |
|
@arjunsuresh Could you merge this PR? Thanks a lot! |
…f-inference into jcalderon/vlm-accuracy-eval
…f-inference into jcalderon/vlm-accuracy-eval
|
@anandhu-eng Could you help to check the CLA checker |
|
Hi @johncalesp @wangshangsam , could you merge this PR or please feel free to do a commit with this change. This should trigger the GitHub action. I'm not able to do a direct commit to the PR branch |
Commit to trigger the GitHub Actions in inference PR
|
Looked like pushing an empty commit would work. @hanyunfan could you help to approve and merge this PR? |
hanyunfan
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
|
@wangshangsam Done |
This PR serves as a proposal to evaluate the accuracy of the VLM.
Notes
calculate_exact_matchwas left in case we need during development, but it may be deleted later)As of now, running 1k samples, I get: