End-to-End generation of Multiple-Choice questions using Text-to-Text transfer Transformer models

作者:

Highlights:

摘要

The increasing worldwide adoption of e-learning tools and widespread increase of online education has brought multiple challenges, including the ability of generating assessments at the scale and speed demanded by this environment. In this sense, recent advances in language models and architectures like the Transformer, provide opportunities to explore how to assist educators in these tasks. This study focuses on using neural language models for the generation of questionnaires composed of multiple-choice questions, based on English Wikipedia articles as input. The problem is addressed using three dimensions: Question Generation (QG), Question Answering (QA), and Distractor Generation (DG). A processing pipeline based on pre-trained T5 language models is designed and a REST API is implemented for its use. The DG task is defined using a Text-To-Text format and a T5 model is fine-tuned on the DG-RACE dataset, showing an improvement to ROUGE-L metric compared to the reference for the dataset. A discussion about the lack of an adequate metric for DG is presented and the cosine similarity using word embeddings is considered as a complement. Questionnaires are evaluated by human experts reporting that questions and options are generally well formed, however, they are more oriented to measuring retention than comprehension.

论文关键词:Multiple-Choice Question Generation,Distractor Generation,Question Answering,Question Generation,Reading Comprehension

论文评审过程:Received 28 December 2021, Revised 10 June 2022, Accepted 20 July 2022, Available online 22 July 2022, Version of Record 31 July 2022.

论文官网地址:https://doi.org/10.1016/j.eswa.2022.118258