Commit
•
754bb8f
1
Parent(s):
e679e9e
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Model Description:
|
2 |
+
|
3 |
+
This model is a T5-base model which is pre-trained on the C4 dataset and fine-tuned on the JFLEG dataset. T5 is an encoder-decoder model pre-trained on a multi-task mixture of unsupervised and supervised tasks and for which each task is converted into a text-to-text format.
|
4 |
+
|
5 |
+
The T5 model was presented in **Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer** by Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J. Liu.
|
6 |
+
|
7 |
+
## Pre-Processing:
|
8 |
+
For this task of grammar correction, we’ll use the prefix “grammar: “ to each of the input sentences. This is done because T5 models are able to perform multiple tasks like translation and summarization with a single model, and a unique prefix is used for each task so that the model learns which task to perform.
|
9 |
+
|
10 |
+
|
11 |
+
|
12 |
+
|