NadiaBedhiafi commited on
Commit
478aa63
1 Parent(s): ee7a6fa

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +19 -47
app.py CHANGED
@@ -1,29 +1,15 @@
1
  import gradio as gr
2
- from huggingface_hub import InferenceClient
3
-
4
- # Use a pipeline as a high-level helper
5
- from transformers import pipeline
6
-
7
- messages = [
8
- {"role": "user", "content": "Who are you?"},
9
- ]
10
- pipe = pipeline("text-generation", model="arcee-ai/Arcee-Spark")
11
- pipe(messages)
12
-
13
-
14
- # Load model directly
15
  from transformers import AutoTokenizer, AutoModelForCausalLM
 
16
 
 
17
  tokenizer = AutoTokenizer.from_pretrained("arcee-ai/Arcee-Spark")
18
  model = AutoModelForCausalLM.from_pretrained("arcee-ai/Arcee-Spark")
19
 
20
- /*
21
- client = InferenceClient("HuggingFaceH4/zephyr-7b-beta")
22
-
23
-
24
  def respond(
25
  message,
26
- history: list[tuple[str, str]],
27
  system_message,
28
  max_tokens,
29
  temperature,
@@ -39,39 +25,25 @@ def respond(
39
 
40
  messages.append({"role": "user", "content": message})
41
 
42
- response = ""
43
 
44
- for message in client.chat_completion(
45
- messages,
46
- max_tokens=max_tokens,
47
- stream=True,
48
- temperature=temperature,
49
- top_p=top_p,
50
- ):
51
- token = message.choices[0].delta.content
52
 
53
- response += token
54
- yield response
55
 
56
- """
57
- For information on how to customize the ChatInterface, peruse the gradio docs: https://www.gradio.app/docs/chatinterface
58
- """
59
- demo = gr.ChatInterface(
60
- respond,
61
- additional_inputs=[
62
- gr.Textbox(value="You are a friendly Chatbot.", label="System message"),
63
- gr.Slider(minimum=1, maximum=2048, value=512, step=1, label="Max new tokens"),
64
- gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
65
- gr.Slider(
66
- minimum=0.1,
67
- maximum=1.0,
68
- value=0.95,
69
- step=0.05,
70
- label="Top-p (nucleus sampling)",
71
- ),
72
  ],
 
 
73
  )
74
- */
75
 
 
76
  if __name__ == "__main__":
77
- demo.launch()
 
1
  import gradio as gr
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  from transformers import AutoTokenizer, AutoModelForCausalLM
3
+ from huggingface_hub import InferenceClient
4
 
5
+ # Load tokenizer and model
6
  tokenizer = AutoTokenizer.from_pretrained("arcee-ai/Arcee-Spark")
7
  model = AutoModelForCausalLM.from_pretrained("arcee-ai/Arcee-Spark")
8
 
9
+ # Function to generate response
 
 
 
10
  def respond(
11
  message,
12
+ history,
13
  system_message,
14
  max_tokens,
15
  temperature,
 
25
 
26
  messages.append({"role": "user", "content": message})
27
 
28
+ #response = ""
29
 
30
+ response = client.chat_completion(messages, max_tokens=max_tokens, temperature=temperature, top_p=top_p)
 
 
 
 
 
 
 
31
 
32
+ yield response # Placeholder yield for response
 
33
 
34
+ # Define Gradio interface
35
+ demo = gr.Interface(
36
+ fn=respond,
37
+ inputs=[
38
+ gr.Textbox(label="System message", default="You are a friendly Chatbot."),
39
+ gr.Slider(label="Max new tokens", minimum=1, maximum=2048, step=1, default=512),
40
+ gr.Slider(label="Temperature", minimum=0.1, maximum=4.0, step=0.1, default=0.7),
41
+ gr.Slider(label="Top-p (nucleus sampling)", minimum=0.1, maximum=1.0, step=0.05, default=0.95),
 
 
 
 
 
 
 
 
42
  ],
43
+ outputs=gr.Textbox(label="Response"),
44
+ title="Chatbot Demo",
45
  )
 
46
 
47
+ # Launch Gradio interface
48
  if __name__ == "__main__":
49
+ demo.launch()