Harmanjotkaur1804 commited on
Commit
0c6e7ed
1 Parent(s): 38cc14d

Upload 2 files

Browse files
Files changed (2) hide show
  1. app.py +34 -0
  2. requirements (2).txt +3 -0
app.py ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import streamlit as st
2
+
3
+
4
+ st.title("Correct Grammar with Transformers 🦄")
5
+ st.write("")
6
+ st.write("Input your text here!")
7
+
8
+ default_value = "Mike and Anna is skiing"
9
+ sent = st.text_area("Text", default_value, height = 50)
10
+ num_return_sequences = st.sidebar.number_input('Number of Return Sequences', min_value=1, max_value=3, value=1, step=1)
11
+
12
+ ### Run Model
13
+ from transformers import T5ForConditionalGeneration, T5Tokenizer
14
+ import torch
15
+ torch_device = 'cuda' if torch.cuda.is_available() else 'cpu'
16
+ tokenizer = T5Tokenizer.from_pretrained('deep-learning-analytics/GrammarCorrector')
17
+ model = T5ForConditionalGeneration.from_pretrained('deep-learning-analytics/GrammarCorrector').to(torch_device)
18
+
19
+ def correct_grammar(input_text,num_return_sequences=num_return_sequences):
20
+ batch = tokenizer([input_text],truncation=True,padding='max_length',max_length=len(input_text), return_tensors="pt").to(torch_device)
21
+ results = model.generate(**batch,max_length=len(input_text),num_beams=2, num_return_sequences=num_return_sequences, temperature=1.5)
22
+ #answer = tokenizer.batch_decode(results[0], skip_special_tokens=True)
23
+ return results
24
+
25
+ ##Prompts
26
+ results = correct_grammar(sent, num_return_sequences)
27
+
28
+ generated_sequences = []
29
+ for generated_sequence_idx, generated_sequence in enumerate(results):
30
+ # Decode text
31
+ text = tokenizer.decode(generated_sequence, clean_up_tokenization_spaces=True, skip_special_tokens=True)
32
+ generated_sequences.append(text)
33
+
34
+ st.write(generated_sequences)
requirements (2).txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ torch
2
+ sentencepiece
3
+ transformers