--- license: apache-2.0 datasets: - jfleg language: - en pipeline_tag: text2text-generation tags: - grammar correction --- # Model This model utilizes the [Flan-T5-base](https://huggingface.co/google/flan-t5-base) pre-trained model and has been fine-tuned using the [JFLEG](https://huggingface.co/datasets/jfleg) dataset with the assistance of the [Happy Transformer](https://github.com/EricFillion/happy-transformer) framework. Its primary objective is to correct a wide range of potential grammatical errors that sentences might contain including issues with punctuation, typos, prepositions, and more. # Usage with Transformers ``` from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("Sajid030/t5-base-grammar-synthesis") model = AutoModelForSeq2SeqLM.from_pretrained("Sajid030/t5-base-grammar-synthesis") text = "One person if do n't have good health that means so many things they could lost ." inputs = tokenizer("grammar:"+text, truncation=True, return_tensors='pt') output = model.generate(inputs['input_ids']) correction=tokenizer.batch_decode(output, skip_special_tokens=True) print("".join(correction)) #Correction: If one person doesn't have good health, so many things could be lost. ``` # Usage with HappyTransformers ``` from happytransformer import HappyTextToText, TTSettings happy_tt = HappyTextToText("T5", "Sajid030/t5-base-grammar-synthesis") args = TTSettings() sentence = "Much many brands and sellers still in the market." result = happy_tt.generate_text("grammar: "+ sentence, args=args) print(result.text) # Many brands and sellers are still in the market. ```