zhuqi commited on
Commit
49774e1
1 Parent(s): 4b23a45

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +100 -0
README.md ADDED
@@ -0,0 +1,100 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ tags:
6
+ - t5-large
7
+ - text2text-generation
8
+ - conversational question rewriting
9
+ datasets:
10
+ - CANARD
11
+ metrics:
12
+ - BLEU
13
+
14
+ model-index:
15
+ - name: t5-large-coqr-canard
16
+ results:
17
+ - task:
18
+ type: text2text-generation
19
+ name: conversational question rewriting
20
+ dataset:
21
+ type: CANARD
22
+ name: CANARD
23
+ split: test
24
+ metrics:
25
+ - type: BLEU
26
+ value: 77.8
27
+ name: BLEU
28
+
29
+ widget:
30
+ - text: "Rewrite the question according to the given context to make the dialog fluent using anaphora and ellipsis.\n\nquestion: What else happened during 1977-1981 other than Superstar Billy Graham's return?\n\ncontext: Superstar Billy Graham\nReturn to WWWF (1977-1981)\nWhy did he return to the WWWF?\nan agreement with promoter Vincent J. McMahon (Senior\nWhat was his agreement with McMahon?\nI don't know.\nHow did people respond to his return?\nI don't know."
31
+ - text: "Rewrite the question according to the given context to make the dialog fluent using anaphora and ellipsis.\n\nquestion: why did Billy Graham personally sued Zahorian and the WWF?\n\ncontext: Superstar Billy Graham\nDisputes with the McMahons\nwhat disputes did he have?\nGraham personally sued Zahorian and the WWF,"
32
+
33
+ inference:
34
+ parameters:
35
+ max_length: 100
36
+
37
+ ---
38
+
39
+ # t5-large-coqr-canard
40
+
41
+ This model is a fine-tuned version of [t5-large](https://huggingface.co/t5-large) on the [CANARD](https://sites.google.com/view/qanta/projects/canard) dataset.
42
+ It achieves the following results on the test set:
43
+ - Loss: 0.3064
44
+ - Bleu: 77.1979
45
+ - Generation Length: 9.576
46
+
47
+ ## Model description
48
+
49
+ CANARD dataset rewrites the original questions in conversations to make them context-independent (understandable w/o context).
50
+ On the contrary, this model is trained to rewrite context-independent questions to conversational questions, aiming to create fluent dialog with anaphora and ellipsis.
51
+
52
+ Input:
53
+ ```
54
+ Rewrite the question according to the given context to make the dialog fluent using anaphora and ellipsis.
55
+
56
+ question: How did people respond to Superstar Billy Graham's return?
57
+
58
+ context: Superstar Billy Graham
59
+ Return to WWWF (1977-1981)
60
+ Why did he return to the WWWF?
61
+ an agreement with promoter Vincent J. McMahon (Senior
62
+ What was his agreement with McMahon?
63
+ I don't know.
64
+ ```
65
+
66
+ Target:
67
+ ```
68
+ How did people respond to his return?
69
+ ```
70
+
71
+ ## Training procedure
72
+
73
+ ### Training hyperparameters
74
+
75
+ The following hyperparameters were used during training:
76
+ - learning_rate: 0.001
77
+ - train_batch_size: 64
78
+ - eval_batch_size: 64
79
+ - seed: 42
80
+ - distributed_type: multi-GPU
81
+ - num_devices: 8
82
+ - total_train_batch_size: 512
83
+ - total_eval_batch_size: 512
84
+ - optimizer: Adafactor
85
+ - lr_scheduler_type: linear
86
+ - num_epochs: 1.0
87
+
88
+ ### Training results
89
+
90
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
91
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|
92
+ | No log | 1.0 | 62 | 0.2987 | 77.2361 | 9.4534 |
93
+
94
+
95
+ ### Framework versions
96
+
97
+ - Transformers 4.20.1
98
+ - Pytorch 1.11.0+cu113
99
+ - Datasets 2.6.1
100
+ - Tokenizers 0.12.1