-
Notifications
You must be signed in to change notification settings - Fork 1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
how to evaluate model outputs on testset #2
Comments
Hi, Our evaluation server for the test set is now available on EvalAI. We welcome all submissions and look forward to your participation! Best wishes! |
Thanks for your reply! We have evaluated our model Marco-VL-Plus on EvalAI, and the accuracy results on val and test dataset are 43.44 and 40.69 respectively. Would you please consider showing our result in your repo? val result: test result: |
Thank you for your reply, and for your interest in the CMMMU benchmark. Regarding your inquiry about submitting your model to the leaderboard, there are a few details we need to confirm with you:
We appreciate your support and contribution to our work once again. Best, |
Thank you for your great job in CMMMU benchmark. |
Great project! And would love to see whether you guys can provide the test answer key for a subset of the health and science partition. Thanks! |
No description provided.
The text was updated successfully, but these errors were encountered: