DHEIVER commited on
Commit
c22a077
1 Parent(s): 9add10d

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -10
app.py CHANGED
@@ -1,22 +1,18 @@
1
  import gradio as gr
2
- from transformers import AutoTokenizer, AutoModelForCausalLM
3
- import torch
4
 
5
- # Load the tokenizer and model
6
- tokenizer = AutoTokenizer.from_pretrained("bigscience/bloom-560m")
7
- model = AutoModelForCausalLM.from_pretrained("bigscience/bloom-560m")
8
 
9
- # Define a function to generate text
10
  def generate_text(prompt):
11
- input_ids = tokenizer.encode(prompt, return_tensors="pt")
12
- output = model.generate(input_ids, max_length=100, num_return_sequences=1)
13
- generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
14
  return generated_text
15
 
16
  # Create a Gradio interface
17
  interface = gr.Interface(
18
  fn=generate_text,
19
- inputs=gr.Textbox("text", label="Digite seu texto aqui:", lines=5), # Increase lines for easier text input.
20
  outputs=gr.Textbox("text", label="Texto Gerado:")
21
  )
22
 
 
1
  import gradio as gr
2
+ from transformers import pipeline
 
3
 
4
+ # Load the pipeline
5
+ pipe = pipeline("text-generation", model="bigscience/bloom-560m")
 
6
 
7
+ # Define a function to generate text using the pipeline
8
  def generate_text(prompt):
9
+ generated_text = pipe(prompt, max_length=100)[0]['generated_text']
 
 
10
  return generated_text
11
 
12
  # Create a Gradio interface
13
  interface = gr.Interface(
14
  fn=generate_text,
15
+ inputs=gr.Textbox("text", label="Digite seu texto aqui:", lines=5),
16
  outputs=gr.Textbox("text", label="Texto Gerado:")
17
  )
18