NaraSpeak_GEC_V1 / README.md
farelzii's picture
Update README.md
6de1f33 verified
metadata
languages:
  - en
license:
  - cc-by-nc-sa-4.0
  - apache-2.0
tags:
  - grammar
  - spelling
  - punctuation
  - error-correction
  - grammar synthesis
  - FLAN
  - C4
datasets:
  - C4
widget:
  - text: >-
      Me go to the store yesterday and buy many thing. I saw a big dog but he no
      bark at me. Then I walk home and eat my lunch, it was delicious sandwich.
      After that, I watch TV and see a funny show about cat who can talk. I
      laugh so hard I cry. Then I go to bed but I no can sleep because I too
      excited about the cat show.
    example_title: Long-Text
  - text: >-
      Me and my family go on a trip to the mountains last week. We drive for
      many hours and finally reach our cabin. The cabin was cozy and warm, with
      a fireplace and big windows. We spend our days hiking and exploring the
      forest. At night, we sit by the fire and tell story. It was a wonderful
      vacation.
    example_title: Long-Text
  - text: >-
      so em if we have an now so with fito ringina know how to estimate the tren
      given the ereafte mylite trend we can also em an estimate is nod s i again
      tort watfettering an we have estimated the trend an called wot to be
      called sthat of exty right now we can and look at wy this should not hare
      a trend i becan we just remove the trend an and we can we now estimate
      tesees ona effect of them exty
    example_title: Transcribed Audio Example
  - text: >-
      My coworker said he used a financial planner to help choose his stocks so
      he wouldn't loose money.
    example_title: incorrect word choice
  - text: >-
      good so hve on an tadley i'm not able to make it to the exla session on
      monday this week e which is why i am e recording pre recording an this
      excelleision and so to day i want e to talk about two things and first of
      all em i wont em wene give a summary er about ta ohow to remove trents in
      these nalitives from time series
    example_title: lowercased audio transcription output
parameters:
  max_length: 128
  min_length: 4
  num_beams: 8
  repetition_penalty: 1.21
  length_penalty: 1
  early_stopping: true

Grammar-Synthesis-Enhanced: FLAN-t5

Open In Colab

This model is a fine-tuned version of pszemraj/flan-t5-large-grammar-synthesis using the C4 200M dataset for the NaraSpeak Bangkit 2024 ENTR-H130 application.

T5 Model Overview

The T5 (Text-To-Text Transfer Transformer) model, introduced by Google Research, is a transformer-based model that treats every NLP task as a text-to-text problem. This unified approach allows T5 to excel at a variety of tasks, such as translation, summarization, and question answering, by converting inputs and outputs into text format.

Transformer Architecture

Transformers are a type of deep learning model designed for sequence-to-sequence tasks. They utilize a mechanism called "attention" to weigh the influence of different words in a sequence, allowing the model to focus on relevant parts of the input when generating each word in the output. This architecture is highly parallelizable and has proven effective in NLP tasks.

Usage in Python

After pip install transformers, run the following code:

from transformers import pipeline

corrector = pipeline(
              'text2text-generation',
              'farelzii/GEC_Test_v1',
              )
raw_text = 'i can has cheezburger'
results = corrector(raw_text)
print(results)