- Authors: Shankar Kantharaj, Xuan Long Do, Rixie Tiffany Ko Leong, Jia Qing Tan, Enamul Hoque, Shafiq Joty
- Paper Link: OpenCQA
- Chart images: The OpenCQA images are available in the chart_image folder.
- Chart metadata: The full annotations are available in the etc/data(full_summary_article) folder. They are saved as (train/val/test)_extended.json files for the train/val/test split. Each sample in the json split has the following structure:
{
image_no: [image_file,
title,
article,
summary,
question,
abstractive_answer,
extractive_answer
],
....
....
....
}
in which title, article, summary, question, abstractive_answer, extractive_answer
stand for the chart title, the full article, the chart summary, the open-ended question, the question's abstractive answer (modified version of extractive answer), and the question's extractive answer.
-
Bounding-boxes and OCRs: All the bounding boxes and OCRs are available in bboxes. Each chart image has a list of bounding boxes. Each bounding box sample has the OCR
sentence
, and bbox format:[x_center, y_center, width, height]
. -
Note: Chart image
X.png
has the bounding box data namedX.json
.
- Experimental types:
- Extractive experiments: Use
extractive_answer
as the ground-truth answer. - Generative experiments: Use
anstractive_answer
as the ground-truth answer.
- Extractive experiments: Use
- Experimental setups:
- With Article:
- Input:
question + title + OCRs + article
. - Output:
extractive_/abstractive_answer
.
- Input:
- With Summary:
- Input:
question + title + OCRs + summary
. - Output:
extractive_/abstractive_answer
.
- Input:
- Without Summary:
- Input:
question + title + OCRs
. - Output:
extractive_/abstractive_answer
.
- Input:
- With Article:
Please refer to VL-T5
Please refer to T5
Please refer to BART
Please refer to BERTQA
Please refer to DOC2GRND
Please refer to ELECTRA
Please refer to GPT2
The metrics used in this work are listed in evaluation_metrics. For each metric, we have steps.txt
which presents the steps to setup and run the metric.
If you have any questions about this work, please contact Enamul Hoque using the following email address: enamulh@yorku.ca.
Please cite our paper if you use our models or dataset in your research.
@inproceedings{kantharaj-etal-2022-opencqa,
title = "{O}pen{CQA}: Open-ended Question Answering with Charts",
author = "Kantharaj, Shankar and
Do, Xuan Long and
Leong, Rixie Tiffany and
Tan, Jia Qing and
Hoque, Enamul and
Joty, Shafiq",
booktitle = "Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2022",
address = "Abu Dhabi, United Arab Emirates",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.emnlp-main.811",
pages = "11817--11837",
abstract = "Charts are very popular to analyze data and convey important insights. People often analyze visualizations to answer open-ended questions that require explanatory answers. Answering such questions are often difficult and time-consuming as it requires a lot of cognitive and perceptual efforts. To address this challenge, we introduce a new task called OpenCQA, where the goal is to answer an open-ended question about a chart with descriptive texts. We present the annotation process and an in-depth analysis of our dataset. We implement and evaluate a set of baselines under three practical settings. In the first setting, a chart and the accompanying article is provided as input to the model. The second setting provides only the relevant paragraph(s) to the chart instead of the entire article, whereas the third setting requires the model to generate an answer solely based on the chart. Our analysis of the results show that the top performing models generally produce fluent and coherent text while they struggle to perform complex logical and arithmetic reasoning.",
}