pszemraj commited on
Commit
af4f81b
1 Parent(s): 53e90a8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -5
README.md CHANGED
@@ -1,10 +1,39 @@
1
  ---
2
  license: apache-2.0
 
 
3
  tags:
4
- - generated_from_trainer
5
- model-index:
6
- - name: checkpoints
7
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  ---
9
 
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -12,7 +41,7 @@ should probably proofread and complete it, then remove this comment. -->
12
 
13
  # checkpoints
14
 
15
- This model is a fine-tuned version of [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) on an unknown dataset.
16
 
17
  ## Model description
18
 
 
1
  ---
2
  license: apache-2.0
3
+ language:
4
+ - en
5
  tags:
6
+ - t5
7
+ - qa
8
+ - askscience
9
+ - lfqa
10
+ - information retrieval
11
+ datasets:
12
+ - vblagoje/lfqa
13
+ metrics:
14
+ - rouge
15
+ widget:
16
+ - text: "why aren't there more planets in our solar system?"
17
+ example_title: "solar system"
18
+ - text: "question: what is a probability distribution? context: I am just learning about statistics."
19
+ example_title: "probability distribution"
20
+ - text: "question: how does exercise help us lose weight? context: I started working out two weeks ago and already feel a lot better, and started to think about it and became deeply confused."
21
+ example_title: "pumpen"
22
+ - text: "what is a neural network?"
23
+ example_title: "deep learning"
24
+ - text: "How can computers understand human language?"
25
+ example_title: "NLP"
26
+
27
+ inference:
28
+ parameters:
29
+ max_length: 64
30
+ no_repeat_ngram_size: 2
31
+ encoder_no_repeat_ngram_size: 4
32
+ repetition_penalty: 3.51
33
+ length_penalty: 0.8
34
+ num_beams: 4
35
+ early_stopping: True
36
+
37
  ---
38
 
39
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
41
 
42
  # checkpoints
43
 
44
+ This model is a fine-tuned version of [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) on the `vblagoje/lfqa` dataset, with training duration of 2 epochs. For a (_somewhat_) apples-to-apples comparison with [t5-base](https://huggingface.co/pszemraj/t5-base-askscience) on the standard eli5 dataset.
45
 
46
  ## Model description
47