UPDF AI

QA4QG: Using Question Answering to Constrain Multi-Hop Question Generation

Dan Su,Peng Xu,Pascale Fung

2022 · DOI: 10.1109/ICASSP43922.2022.9747008
IEEE International Conference on Acoustics, Speech, and Signal Processing · 7 Citations

TLDR

This work proposes a novel framework, QA4QG, a QA-augmented BART-based framework for MQG that augments the standard BART model with an additional multi-hop QA module to further con-strain the generated question.

Abstract

Multi-hop question generation (MQG) aims to generate com-plex questions which require reasoning over multiple pieces of information of the input passage. Most existing work on MQG has focused on exploring graph-based networks to equip the traditional Sequence-to-sequence framework with reasoning ability. However, these models do not take full advantage of the constraint between questions and answers. Furthermore, studies on multi-hop question answering (QA) suggest that Transformers can replace the graph structure for multi-hop reasoning. Therefore, in this work, we propose a novel framework, QA4QG, a QA-augmented BART-based framework for MQG. It augments the standard BART model with an additional multi-hop QA module to further con-strain the generated question. Our results on the HotpotQA dataset show that QA4QG outperforms all state-of-the-art models, with an increase of 8 BLEU-4 and 8 ROUGE points compared to the best results previously reported. Our work suggests the advantage of introducing pre-trained language models and QA module for the MQG task.

Cited Papers
Citing Papers