[bug] MT-Bench only evaluate the second turn. #1085
Unanswered
notoschord
asked this question in
Q&A
Replies: 2 comments 8 replies
-
Hello, you can set Besides, if you want to test alpacaeval2, just use this config is totally same as official https://github.com/open-compass/opencompass/blob/main/configs/eval_subjective_alpacaeval.py Feel free to contact me if you have any other problems on subjective evaluation for llm. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Thank you for your answer. It can evaluate the first turn separately. |
Beta Was this translation helpful? Give feedback.
8 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
When MTBenchDataset is loaded, multi_turn is always
True
, so only second turn answers will be evaluated.Beta Was this translation helpful? Give feedback.
All reactions