kaejo98 commited on
Commit
7209849
1 Parent(s): 113ad01

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -8
README.md CHANGED
@@ -7,18 +7,14 @@ model-index:
7
  results: []
8
  ---
9
 
10
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
- should probably proofread and complete it, then remove this comment. -->
12
 
13
- # bart-base_question_generation
14
 
15
- This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset.
16
 
17
- ## Model description
18
 
19
- More information needed
20
 
21
- ## Intended uses & limitations
22
 
23
  The model takes context as an input sequence, and will generate a full question sentence as an output sequence. There are two ways the model can be queried produce the questions:
24
  - <b> Casual-Generation </b>: where the model is tasked to generate questions answerable by a given passage. The input should be follow the structure or format: '\<generate_questions\> paragraph: put your passage text here'. <br/>
@@ -33,6 +29,11 @@ Example: <br/>
33
 
34
  The input sequence can then be encoded and passed as the input_ids argument in the model's generate() method.
35
 
 
 
 
 
 
36
  ## Training and evaluation data
37
 
38
  The dataset used to train the model comprises the training datasets from:
@@ -41,8 +42,10 @@ The dataset used to train the model comprises the training datasets from:
41
  - DROP (Discrete Reasoning Over Paragraphs): https://allenai.org/data/drop
42
  - SciQ
43
 
 
 
44
  ## Training procedure
45
-
46
  ### Training hyperparameters
47
 
48
  The following hyperparameters were used during training:
@@ -55,6 +58,8 @@ The following hyperparameters were used during training:
55
  - lr_scheduler_warmup_ratio: 0.25
56
  - num_epochs: 5
57
 
 
 
58
  ### Framework versions
59
 
60
  - Transformers 4.23.1
 
7
  results: []
8
  ---
9
 
 
 
10
 
 
11
 
12
+ # BART-base Question Generation
13
 
14
+ This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on different questions and answering dataset. It was trained to generation question using two different approaches, <b> Casual-Generation </b> and <b> Context-based-Generation </b>.
15
 
 
16
 
17
+ ## Model description
18
 
19
  The model takes context as an input sequence, and will generate a full question sentence as an output sequence. There are two ways the model can be queried produce the questions:
20
  - <b> Casual-Generation </b>: where the model is tasked to generate questions answerable by a given passage. The input should be follow the structure or format: '\<generate_questions\> paragraph: put your passage text here'. <br/>
 
29
 
30
  The input sequence can then be encoded and passed as the input_ids argument in the model's generate() method.
31
 
32
+ ## limitations
33
+ The model was trained on only a limited amount of data hence questions might be poor quality. In addition the questions generated have style similar to that of the training data.
34
+
35
+
36
+
37
  ## Training and evaluation data
38
 
39
  The dataset used to train the model comprises the training datasets from:
 
42
  - DROP (Discrete Reasoning Over Paragraphs): https://allenai.org/data/drop
43
  - SciQ
44
 
45
+ After preprocessing the data from the above listed datasets, we had 408372 examples for training the model and 25k for development and 18k for testing.
46
+
47
  ## Training procedure
48
+ The model is trained (finetuned) for 5 epochs with the hyperparameters listed below:
49
  ### Training hyperparameters
50
 
51
  The following hyperparameters were used during training:
 
58
  - lr_scheduler_warmup_ratio: 0.25
59
  - num_epochs: 5
60
 
61
+ At the end of 5 epochs, the Evaluation loss was: 1.6493076086044312 and the training loss was: 0.9671.
62
+
63
  ### Framework versions
64
 
65
  - Transformers 4.23.1