Pippoz commited on
Commit
b408af1
1 Parent(s): d3a925a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -4
app.py CHANGED
@@ -11,10 +11,6 @@ import torch
11
  # return model, tokenizer
12
  generator = pipeline('text-generation', model="facebook/opt-1.3b", skip_special_tokens=True)
13
 
14
- @st.cache(allow_output_mutation=True)
15
- def define_model(prompt):
16
- answer = generator(prompt)
17
- return answer
18
 
19
 
20
  #@st.cache(allow_output_mutation=True)
@@ -29,6 +25,7 @@ def define_model(prompt):
29
 
30
  prompt= st.text_area('Your prompt here',
31
  '''Hello, I'm am conscious and''')
 
32
 
33
  #answer = opt_model(prompt, model, tokenizer,)
34
  #lst = ['ciao come stai sjfsbd dfhsdf fuahfuf feuhfu wefwu ']
 
11
  # return model, tokenizer
12
  generator = pipeline('text-generation', model="facebook/opt-1.3b", skip_special_tokens=True)
13
 
 
 
 
 
14
 
15
 
16
  #@st.cache(allow_output_mutation=True)
 
25
 
26
  prompt= st.text_area('Your prompt here',
27
  '''Hello, I'm am conscious and''')
28
+ answer = generator(prompt, max_length=60)
29
 
30
  #answer = opt_model(prompt, model, tokenizer,)
31
  #lst = ['ciao come stai sjfsbd dfhsdf fuahfuf feuhfu wefwu ']