Vaishakhh commited on
Commit
54acd71
1 Parent(s): e45f695

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +18 -12
app.py CHANGED
@@ -8,29 +8,29 @@ Original file is located at
8
 
9
  """
10
 
11
- from sentence_splitter import SentenceSplitter, split_text_into_sentences
12
- from parrot import Parrot
13
- import warnings
14
- warnings.filterwarnings("ignore")
15
-
16
- #parrot = Parrot(model_tag="prithivida/parrot_paraphraser_on_T5")
17
  import os
 
18
  from parrot import Parrot
19
  import torch
20
  import warnings
21
  import nltk
22
-
23
- #!pip install sentence-splitter
24
-
25
  warnings.filterwarnings("ignore")
 
26
  splitter = SentenceSplitter(language='en')
 
 
27
  from transformers import PegasusForConditionalGeneration, PegasusTokenizer
28
  from transformers import AutoTokenizer
29
  from transformers import AutoModelForSeq2SeqLM
30
- import pandas as pd
31
  from parrot.filters import Adequacy
32
  from parrot.filters import Fluency
33
  from parrot.filters import Diversity
 
 
 
34
  adequacy_score = Adequacy()
35
  fluency_score = Fluency()
36
  diversity_score= Diversity()
@@ -38,7 +38,11 @@ device= "cuda:0"
38
  adequacy_threshold = 0.90
39
  fluency_threshold = 0.90
40
  diversity_ranker="levenshtein"
41
- do_diverse=True
 
 
 
 
42
  model_name = 'tuner007/pegasus_paraphrase'
43
  torch_device = 'cuda' if torch.cuda.is_available() else 'cpu'
44
  tokenizer = PegasusTokenizer.from_pretrained(model_name)
@@ -48,7 +52,7 @@ def get_max_str(lst):
48
  return max(lst, key=len)
49
  def get_response(input_text,num_return_sequences=10,num_beams=10):
50
  batch = tokenizer.prepare_seq2seq_batch([input_text],truncation=True,padding='longest',max_length=30,return_tensors='pt').to(torch_device)
51
- translated = model_pegasus.generate(**batch,max_length=30,num_beams=num_beams, num_return_sequences=num_return_sequences, num_beam_groups=num_beams, diversity_penalty=0.5, temperature=1.5)
52
  tgt_text = tokenizer.batch_decode(translated, skip_special_tokens=True)
53
  try:
54
  adequacy_filtered_phrases = adequacy_score.filter(input_text,tgt_text, adequacy_threshold, device)
@@ -63,6 +67,8 @@ def get_response(input_text,num_return_sequences=10,num_beams=10):
63
  return get_max_str(adequacy_filtered_phrases)
64
  except:
65
  return(get_max_str(tgt_text))
 
 
66
 
67
  import gradio as gr
68
 
 
8
 
9
  """
10
 
11
+ # importing the libraries
 
 
 
 
 
12
  import os
13
+ import pandas as pd
14
  from parrot import Parrot
15
  import torch
16
  import warnings
17
  import nltk
18
+ import warnings
 
 
19
  warnings.filterwarnings("ignore")
20
+
21
  splitter = SentenceSplitter(language='en')
22
+ from sentence_splitter import SentenceSplitter, split_text_into_sentences
23
+ from parrot import Parrot
24
  from transformers import PegasusForConditionalGeneration, PegasusTokenizer
25
  from transformers import AutoTokenizer
26
  from transformers import AutoModelForSeq2SeqLM
27
+
28
  from parrot.filters import Adequacy
29
  from parrot.filters import Fluency
30
  from parrot.filters import Diversity
31
+
32
+ # Adding the metrics
33
+
34
  adequacy_score = Adequacy()
35
  fluency_score = Fluency()
36
  diversity_score= Diversity()
 
38
  adequacy_threshold = 0.90
39
  fluency_threshold = 0.90
40
  diversity_ranker="levenshtein"
41
+ do_diverse=False
42
+ #num_beam_groups=num_beams, diversity_penalty=0.5
43
+
44
+ # adding the model
45
+
46
  model_name = 'tuner007/pegasus_paraphrase'
47
  torch_device = 'cuda' if torch.cuda.is_available() else 'cpu'
48
  tokenizer = PegasusTokenizer.from_pretrained(model_name)
 
52
  return max(lst, key=len)
53
  def get_response(input_text,num_return_sequences=10,num_beams=10):
54
  batch = tokenizer.prepare_seq2seq_batch([input_text],truncation=True,padding='longest',max_length=30,return_tensors='pt').to(torch_device)
55
+ translated = model_pegasus.generate(**batch,max_length=30,num_beams=num_beams, num_return_sequences=num_return_sequences, temperature=1.5)
56
  tgt_text = tokenizer.batch_decode(translated, skip_special_tokens=True)
57
  try:
58
  adequacy_filtered_phrases = adequacy_score.filter(input_text,tgt_text, adequacy_threshold, device)
 
67
  return get_max_str(adequacy_filtered_phrases)
68
  except:
69
  return(get_max_str(tgt_text))
70
+
71
+ # Deploying the model
72
 
73
  import gradio as gr
74