farelzii commited on
Commit
0c5262a
1 Parent(s): b413f6e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -5
README.md CHANGED
@@ -39,18 +39,23 @@ parameters:
39
 
40
  # Grammar-Synthesis-Enhanced: FLAN-t5
41
 
42
- <a href="https://colab.research.google.com/gist/pszemraj/5dc89199a631a9c6cfd7e386011452a0/demo-flan-t5-large-grammar-synthesis.ipynb">
43
  <img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open In Colab"/>
44
  </a>
45
 
46
- A fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) for grammar correction on an expanded version of the [JFLEG](https://paperswithcode.com/dataset/jfleg) dataset and further fine-tuned using the [C4 200M](https://www.tensorflow.org/datasets/community_catalog/huggingface/c4) dataset. [Demo](https://huggingface.co/spaces/pszemraj/FLAN-grammar-correction) on HF spaces.
47
 
 
48
 
49
- ## Usage in Python
50
 
51
- > There's a colab notebook that already has this basic version implemented (_click on the Open in Colab button_)
52
 
53
- After `pip install transformers` run the following code:
 
 
 
 
54
 
55
  ```python
56
  from transformers import pipeline
 
39
 
40
  # Grammar-Synthesis-Enhanced: FLAN-t5
41
 
42
+ <a href="https://colab.research.google.com/gist/pszemraj/5dc89199a631a9c6cfd7e386011452a0/demo-flan-t5-large-grammar-synthesis.ipynb">
43
  <img src="https://colab.research.google.com/assets/colab-badge.svg" alt="Open In Colab"/>
44
  </a>
45
 
46
+ This model is a fine-tuned version of [pszemraj/flan-t5-large-grammar-synthesis](https://huggingface.co/pszemraj/flan-t5-large-grammar-synthesis) using the C4 200M dataset for the NaraSpeak Bangkit 2024 ENTR-H130 application.
47
 
48
+ ## T5 Model Overview
49
 
50
+ The T5 (Text-To-Text Transfer Transformer) model, introduced by Google Research, is a transformer-based model that treats every NLP task as a text-to-text problem. This unified approach allows T5 to excel at a variety of tasks, such as translation, summarization, and question answering, by converting inputs and outputs into text format.
51
 
52
+ ### Transformer Architecture
53
 
54
+ Transformers are a type of deep learning model designed for sequence-to-sequence tasks. They utilize a mechanism called "attention" to weigh the influence of different words in a sequence, allowing the model to focus on relevant parts of the input when generating each word in the output. This architecture is highly parallelizable and has proven effective in NLP tasks.
55
+
56
+ ## Usage in Python
57
+
58
+ After `pip install transformers`, run the following code:
59
 
60
  ```python
61
  from transformers import pipeline