Non-playing-Character commited on
Commit
06427dc
·
verified ·
1 Parent(s): 078187c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +13 -58
app.py CHANGED
@@ -1,63 +1,18 @@
1
  import gradio as gr
2
- from transformers import BertTokenizer, BertForSequenceClassification
3
- import torch
4
 
5
- # Load the tokenizer and model
6
- tokenizer = BertTokenizer.from_pretrained('RAGFillerModel1')
7
- model = BertForSequenceClassification.from_pretrained('RAGFillerModel1', num_labels=30)
8
 
9
- # Define the labels
10
- labels = [
11
- "That's an interesting question... let me see.",
12
- "Hmm, I need to consider that for a moment.",
13
- "Let me think about how best to address that.",
14
- "Well, I think it really depends on a few factors...",
15
- "Good thought! I need a moment to process that.",
16
- "You know, I've never really thought about it that way before.",
17
- "Okay, let me break that down for a second.",
18
- "That's a tough one... give me a second to gather my thoughts.",
19
- "I want to make sure I give you the right answer, so let me think.",
20
- "Let me reflect on that... there are a few angles to consider.",
21
- "Alright, if I remember correctly, it goes something like this...",
22
- "That's a good point, and I think the answer is...",
23
- "Good question! Let me take a moment to unpack that.",
24
- "Hmm, there's a lot to consider here. Give me a second.",
25
- "Let me think about that... it's not a straightforward answer.",
26
- "Interesting... I need to gather my thoughts on this.",
27
- "Well, let me consider the various aspects before I answer.",
28
- "Alright, let's break this down a bit before I answer.",
29
- "Good thought! I want to make sure I address it properly.",
30
- "Hmm, let's delve into that a bit more; I'll need a moment.",
31
- "Great question! I want to provide a thoughtful response, so let me think.",
32
- "That's a fascinating angle... let me think it through.",
33
- "I'll need a moment to come up with an answer.",
34
- "I'll take a quick moment to weigh my options.",
35
- "I appreciate the question; let me think it through.",
36
- "Let me take a step back and think that over.",
37
- "Let me mull that over for just a moment.",
38
- "I want to consider that carefully; let me pause for a second.",
39
- "Let's explore that further; I need a moment to think.",
40
- "I'd like to think that over before answering."
41
- ]
42
 
43
- def classify_text(text):
44
- # Tokenize the input text
45
- inputs = tokenizer(text, padding=True, truncation=True, return_tensors='pt')
46
- outputs = model(**inputs)
 
 
47
 
48
- predictions = torch.nn.functional.softmax(outputs.logits, dim=-1)
49
- # Convert predictions to numpy array
50
- predictions = predictions.cpu().detach().numpy()
51
-
52
- labeled_predictions = {labels[i]: predictions[0][i] for i in range(len(labels))}
53
- max_label = labels[predictions[0].argmax()]
54
- max_probability = predictions[0].max()
55
-
56
- result = {max_label: max_probability}
57
- return result
58
-
59
- # Create a Gradio interface
60
- iface = gr.Interface(fn=classify_text, inputs="text", outputs="label")
61
-
62
- # Launch the interface
63
- iface.launch()
 
1
  import gradio as gr
2
+ from transformers import pipeline
 
3
 
4
+ pipeline = pipeline(task="image-classification", model="julien-c/hotdog-not-hotdog")
 
 
5
 
6
+ def predict(input_img):
7
+ predictions = pipeline(input_img)
8
+ return input_img, {p["label"]: p["score"] for p in predictions}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
 
10
+ gradio_app = gr.Interface(
11
+ predict,
12
+ inputs=gr.Image(label="Select hot dog candidate", sources=['upload', 'webcam'], type="pil"),
13
+ outputs=[gr.Image(label="Processed Image"), gr.Label(label="Result", num_top_classes=2)],
14
+ title="Hot Dog? Or Not?",
15
+ )
16
 
17
+ if __name__ == "__main__":
18
+ gradio_app.launch()