TheMaisk commited on
Commit
eefd049
1 Parent(s): fca2eb7

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -11,17 +11,17 @@ def generate(
11
  temperature=0.4,
12
  top_p=0.95,
13
  top_k=50,
14
- max_new_tokens=2000,
15
  ):
16
  pipe = load_model(model_name)
17
- # Need to add additional options later.
18
  if template_name == "Falcon 1B Template":
19
  message_template = [
20
  {"role": "user", "content": "Hello!"},
21
  {"role": "assistant", "content": "Hello! How can I assist you today?"},
22
  {"role": "user", "content": user_input},
23
  ]
24
- else: # Default to "TinyLlama Template"
25
  message_template = [
26
  {
27
  "role": "system",
@@ -30,7 +30,6 @@ def generate(
30
  {"role": "user", "content": user_input},
31
  ]
32
 
33
- # Set tokenize correctly. Otherwise ticking the box breaks it.
34
  prompt = pipe.tokenizer.apply_chat_template(message_template, tokenize=False, add_generation_prompt=True)
35
  outputs = pipe(prompt, max_new_tokens=max_new_tokens, do_sample=True,
36
  temperature=temperature, top_k=top_k, top_p=top_p, repetition_penalty=1.10)
@@ -38,7 +37,7 @@ def generate(
38
 
39
  model_choices = ["TinyLlama/TinyLlama-1.1B-Chat-v1.0", "ericzzz/falcon-rw-1b-chat"]
40
  template_choices = ["TinyLlama Template", "Falcon Template"]
41
- # What at the best options?
42
  g = gr.Interface(
43
  fn=generate,
44
  inputs=[
@@ -48,12 +47,13 @@ g = gr.Interface(
48
  gr.components.Slider(minimum=0, maximum=1, value=0.4, label="Temperature"),
49
  gr.components.Slider(minimum=0, maximum=1, value=0.95, label="Top p"),
50
  gr.components.Slider(minimum=0, maximum=100, step=1, value=50, label="Top k"),
51
- gr.components.Slider(minimum=1, maximum=1024, step=1, value=2000, label="Max tokens"),
52
  ],
53
  outputs=[gr.Textbox(lines=10, label="Output")],
54
  title="Hugging Face Transformers Model",
55
  description="A simple interface for generating text with a Hugging Face Transformers model.",
 
56
  concurrency_limit=1
57
  )
58
 
59
- g.launch(max_threads=2)
 
11
  temperature=0.4,
12
  top_p=0.95,
13
  top_k=50,
14
+ max_new_tokens=256,
15
  ):
16
  pipe = load_model(model_name)
17
+ # Hier weitere Optionen hinzufügen
18
  if template_name == "Falcon 1B Template":
19
  message_template = [
20
  {"role": "user", "content": "Hello!"},
21
  {"role": "assistant", "content": "Hello! How can I assist you today?"},
22
  {"role": "user", "content": user_input},
23
  ]
24
+ else: # Standardmäßig "TinyLlama Template"
25
  message_template = [
26
  {
27
  "role": "system",
 
30
  {"role": "user", "content": user_input},
31
  ]
32
 
 
33
  prompt = pipe.tokenizer.apply_chat_template(message_template, tokenize=False, add_generation_prompt=True)
34
  outputs = pipe(prompt, max_new_tokens=max_new_tokens, do_sample=True,
35
  temperature=temperature, top_k=top_k, top_p=top_p, repetition_penalty=1.10)
 
37
 
38
  model_choices = ["TinyLlama/TinyLlama-1.1B-Chat-v1.0", "ericzzz/falcon-rw-1b-chat"]
39
  template_choices = ["TinyLlama Template", "Falcon Template"]
40
+
41
  g = gr.Interface(
42
  fn=generate,
43
  inputs=[
 
47
  gr.components.Slider(minimum=0, maximum=1, value=0.4, label="Temperature"),
48
  gr.components.Slider(minimum=0, maximum=1, value=0.95, label="Top p"),
49
  gr.components.Slider(minimum=0, maximum=100, step=1, value=50, label="Top k"),
50
+ gr.components.Slider(minimum=1, maximum=1024, step=1, value=256, label="Max tokens"),
51
  ],
52
  outputs=[gr.Textbox(lines=10, label="Output")],
53
  title="Hugging Face Transformers Model",
54
  description="A simple interface for generating text with a Hugging Face Transformers model.",
55
+ theme='syddharth/gray-minimal', # Hinzugefügtes Theme
56
  concurrency_limit=1
57
  )
58
 
59
+ g.launch(max_threads=2)