In https://huggingface.co/datasets/drive-bench/arena
the json file corresponding to corruptions only have the question_type: robust_qas
This is a serious issue as the eval.py expects these question types:
self.results = {
"perception": {
"MCQ": {"gpt": [], "accuracy": []},
"VQA": {"gpt": [], "language": []}
},
"prediction": {
"VQA": {"gpt": [], "language": []}
},
"planning": {
"VQA": {"gpt": [], "language": []}
},
"behavior": {
"MCQ": {"gpt": [], "accuracy": []}}
}
The json files for corruptions need to be updated.
