Ambiguous questions are a challenge for Question Answering models, as they require answers that cover multiple interpretations of the original query. To this end, these models are required to generate long-form answers that often combine conflicting pieces of information. Although recent advances in the field have shown strong capabilities in generating fluent responses, certain research questions remain unanswered. Does model/data scaling improve the answers' quality? Do automated metrics align with human judgment? To what extent do these models ground their answers in evidence? In this study, we aim to thoroughly investigate these aspects, and provide valuable insights into the limitations of the current approaches.
You can finetune T5-base on the ASQA dataset by running the following command:
BASE_MODEL=t5-base OPEN_BOOK=true python finetune.py
Similarly, you can finetune BART-large with:
BASE_MODEL=bart-large MODEL_NAME=facebook/bart-large OPEN_BOOK=true python finetune.py
All available environment variables are:
BASE_MODEL
: the base model being used (defaults to "t5-base")MODEL_NAME
: the HuggingFace name of the model (defaults to BASE_MODEL)TOKENIZER_NAME
: the HuggingFace name of the tokenizer (defaults to MODEL_NAME)DATASET_HF_USER
: the HuggingFace user that hosts the dataset to train on (defaults to "din0s")DATASET_NAME
: the HuggingFace dataset trian on (defaults to "asqa")OPEN_BOOK
: whether to finetune for the open-book scenario or not (defaults to "false")
To replicate our human evaluation study, you can use the notebook create_pairwise_comparisons.ipynb.
This project is built on top of the ASQA dataset. For more information, please refer to the ASQA repository. The following setup instructions come from the original codebase.
To download the ASQA dataset, run:
mkdir dataset
gsutil cp -R gs://gresearch/ASQA/data/ASQA.json dataset
Note: this requires gsutil.
-
You might want to setup a virtual environment before installation.
-
Install PyTorch by following the instructions here.
-
Install python packages and download the Roberta checkpoint by running:
sh install.sh
chmod +x ./eval.sh
./eval.sh ${RESULTS_PATH} ${EXP_NAME}
The final results will show on the screen and will also be generated in ./results/${EXP_NAME}/final_eval_results.json.