|
--- |
|
language: |
|
- en |
|
license: apache-2.0 |
|
tags: |
|
- t5-large |
|
- text2text-generation |
|
- conversational question rewriting |
|
datasets: |
|
- CANARD |
|
metrics: |
|
- BLEU |
|
widget: |
|
- text: 'Rewrite the question according to the given context to make the dialog fluent |
|
using anaphora and ellipsis. |
|
|
|
|
|
question: What else happened during 1977-1981 other than Superstar Billy Graham''s |
|
return? |
|
|
|
|
|
context: Superstar Billy Graham |
|
|
|
Return to WWWF (1977-1981) |
|
|
|
Why did he return to the WWWF? |
|
|
|
an agreement with promoter Vincent J. McMahon (Senior |
|
|
|
What was his agreement with McMahon? |
|
|
|
I don''t know. |
|
|
|
How did people respond to his return? |
|
|
|
I don''t know.' |
|
- text: 'Rewrite the question according to the given context to make the dialog fluent |
|
using anaphora and ellipsis. |
|
|
|
|
|
question: why did Billy Graham personally sued Zahorian and the WWF? |
|
|
|
|
|
context: Superstar Billy Graham |
|
|
|
Disputes with the McMahons |
|
|
|
what disputes did he have? |
|
|
|
Graham personally sued Zahorian and the WWF,' |
|
inference: |
|
parameters: |
|
max_length: 100 |
|
base_model: t5-large |
|
model-index: |
|
- name: t5-large-coqr-canard |
|
results: |
|
- task: |
|
type: text2text-generation |
|
name: conversational question rewriting |
|
dataset: |
|
name: CANARD |
|
type: CANARD |
|
split: test |
|
metrics: |
|
- type: BLEU |
|
value: 77.8 |
|
name: BLEU |
|
--- |
|
|
|
# t5-large-coqr-canard |
|
|
|
This model is a fine-tuned version of [t5-large](https://huggingface.co/t5-large) on the [CANARD](https://sites.google.com/view/qanta/projects/canard) dataset. |
|
It achieves the following results on the test set: |
|
- Loss: 0.3064 |
|
- Bleu: 77.1979 |
|
- Generation Length: 9.576 |
|
|
|
## Model description |
|
|
|
CANARD dataset rewrites the original questions in conversations to make them context-independent (understandable w/o context). |
|
On the contrary, this model is trained to rewrite context-independent questions to conversational questions, aiming to create fluent dialog with anaphora and ellipsis. |
|
|
|
Input: |
|
``` |
|
Rewrite the question according to the given context to make the dialog fluent using anaphora and ellipsis. |
|
|
|
question: How did people respond to Superstar Billy Graham's return? |
|
|
|
context: Superstar Billy Graham |
|
Return to WWWF (1977-1981) |
|
Why did he return to the WWWF? |
|
an agreement with promoter Vincent J. McMahon (Senior |
|
What was his agreement with McMahon? |
|
I don't know. |
|
``` |
|
|
|
Target: |
|
``` |
|
How did people respond to his return? |
|
``` |
|
|
|
## Training procedure |
|
|
|
### Training hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
- learning_rate: 0.001 |
|
- train_batch_size: 64 |
|
- eval_batch_size: 64 |
|
- seed: 42 |
|
- distributed_type: multi-GPU |
|
- num_devices: 8 |
|
- total_train_batch_size: 512 |
|
- total_eval_batch_size: 512 |
|
- optimizer: Adafactor |
|
- lr_scheduler_type: linear |
|
- num_epochs: 1.0 |
|
|
|
### Training results |
|
|
|
| Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len | |
|
|:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:| |
|
| No log | 1.0 | 62 | 0.2987 | 77.2361 | 9.4534 | |
|
|
|
|
|
### Framework versions |
|
|
|
- Transformers 4.20.1 |
|
- Pytorch 1.11.0+cu113 |
|
- Datasets 2.6.1 |
|
- Tokenizers 0.12.1 |
|
|