Transformer QG on SQuAD

HLQG is Proposed by Ying-Hong Chan & Yao-Chung Fan. (2019). A Re-current BERT-based Model for Question Generation.

This is a Reproduce Version

More detail: p208p2002/Transformer-QG-on-SQuAD

Usage

Input Format

C' = [c1, c2, ..., [HL], a1, ..., a|A|, [HL], ..., c|C|]

Input Example

Harry Potter is a series of seven fantasy novels written by British author, [HL]J. K. Rowling[HL].

Who wrote Harry Potter?

Data setting

We report two dataset setting as Follow

SQuAD

SQuAD NQG

Available models

  • BART
  • GPT2
  • T5

Expriments

We report score with NQG Scorer which is using in SQuAD NQG.

If not special explanation, the size of the model defaults to "base".

SQuAD

Model Bleu 1 Bleu 2 Bleu 3 Bleu 4 METEOR ROUGE-L
BART-HLSQG 54.67 39.26 30.34 24.15 25.43 52.64
GPT2-HLSQG 49.31 33.95 25.41 19.69 22.29 48.82
T5-HLSQG 54.29 39.22 30.43 24.26 25.56 53.11

SQuAD NQG

Model Bleu 1 Bleu 2 Bleu 3 Bleu 4 METEOR ROUGE-L
BERT-HLSQG (Chan et al.) 49.73 34.60 26.13 20.33 23.88 48.23
BART-HLSQG 54.12 38.19 28.84 22.35 24.55 51.03
GPT2-HLSQG 49.82 33.69 24.71 18.63 21.90 47.60
T5-HLSQG 53.13 37.60 28.62 22.38 24.48 51.20
Downloads last month
17
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train p208p2002/bart-squad-nqg-hl

Collection including p208p2002/bart-squad-nqg-hl