pszemraj commited on
Commit
3f80286
1 Parent(s): bc3209c

update generation examples and params

Browse files
Files changed (1) hide show
  1. README.md +15 -15
README.md CHANGED
@@ -1,9 +1,10 @@
1
  ---
2
  license: apache-2.0
3
  tags:
4
- - generated_from_trainer
5
- model-index:
6
- - name: t5-v1_1-base-ft-jflAUG
 
7
 
8
  widget:
9
  - text: "Anna and Mike is going skiing"
@@ -26,24 +27,27 @@ ta ohow to remove trents in these nalitives from time series"
26
  example_title: "dangling modifier"
27
  - text: "There car broke down so their hitching a ride to they're class."
28
  example_title: "compound-1"
 
 
29
 
30
  inference:
31
  parameters:
32
- no_repeat_ngram_size: 2
33
  max_length: 64
34
  min_length: 4
35
  num_beams: 4
36
- repetition_penalty: 3.51
37
- length_penalty: 0.8
38
  early_stopping: True
39
  ---
40
 
41
 
42
  # t5-v1_1-base-ft-jflAUG
43
 
44
- - **GOAL:** build a more robust and generalized grammar and spelling correction model with minimal impact on the semantics of correct sentences (I.e. it does not change things that do not need to be changed.
45
- - this grammar correction model (at least from preliminary testing) can handle large amounts of errors in the source text (i.e. from audio transcription) and still produce cohesive results.
46
- - This model is a fine-tuned version of [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) on an expanded version of the [JFLEG dataset](https://aclanthology.org/E17-2037/).
 
47
 
48
  ## Model description
49
 
@@ -58,9 +62,8 @@ inference:
58
 
59
  ## Training and evaluation data
60
 
61
- More information needed
62
-
63
- ## Training procedure
64
 
65
  ### Training hyperparameters
66
 
@@ -77,9 +80,6 @@ The following hyperparameters were used during training:
77
  - lr_scheduler_warmup_ratio: 0.05
78
  - num_epochs: 5
79
 
80
- ### Training results
81
-
82
-
83
 
84
  ### Framework versions
85
 
 
1
  ---
2
  license: apache-2.0
3
  tags:
4
+ - grammar
5
+ - spelling
6
+ - punctuation
7
+ - error-correction
8
 
9
  widget:
10
  - text: "Anna and Mike is going skiing"
 
27
  example_title: "dangling modifier"
28
  - text: "There car broke down so their hitching a ride to they're class."
29
  example_title: "compound-1"
30
+ - text: "Which part of Zurich was you going to go hiking in when we were there for the first time together? ! ?"
31
+ example_title: "chatbot on Zurich"
32
 
33
  inference:
34
  parameters:
35
+ no_repeat_ngram_size: 4
36
  max_length: 64
37
  min_length: 4
38
  num_beams: 4
39
+ repetition_penalty: 1.51
40
+ length_penalty: 1
41
  early_stopping: True
42
  ---
43
 
44
 
45
  # t5-v1_1-base-ft-jflAUG
46
 
47
+ > **GOAL:** a more robust and generalized grammar and spelling correction model with minimal impact on the semantics of correct sentences (i.e. it does not change things that do not need to be changed).
48
+
49
+ - this model _(at least from preliminary testing)_ can handle large amounts of errors in the source text (i.e. from audio transcription) and still produce cohesive results.
50
+ - a fine-tuned version of [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) on an expanded version of the [JFLEG dataset](https://aclanthology.org/E17-2037/).
51
 
52
  ## Model description
53
 
 
62
 
63
  ## Training and evaluation data
64
 
65
+ - trained as text-to-text
66
+ - JFLEG dataset + additional selected and/or generated grammar corrections
 
67
 
68
  ### Training hyperparameters
69
 
 
80
  - lr_scheduler_warmup_ratio: 0.05
81
  - num_epochs: 5
82
 
 
 
 
83
 
84
  ### Framework versions
85