rodrigomasini commited on
Commit
0eecf08
1 Parent(s): 9f15b87

Update app/main.py

Browse files
Files changed (1) hide show
  1. app/main.py +6 -8
app/main.py CHANGED
@@ -12,18 +12,16 @@ PWD = os.getenv("PWD")
12
  def main(args):
13
  demo = gr.ChatInterface(
14
  fn=chat,
15
- examples=["hello", "how are you?", "What is Large Language Model?"],
16
- title="Gradio 🤝 TGI",
17
- description="This space is a template that you can fork/duplicate for your own usage. "
18
  "This space let you build LLM powered idea on top of [Gradio](https://www.gradio.app/) "
19
  "and open LLM served locally by [TGI(Text Generation Inference)](https://huggingface.co/docs/text-generation-inference/en/index). "
20
  "Below is a placeholder Gradio ChatInterface for you to try out Mistral-7B backed by the power of TGI's efficiency. \n\n"
21
  "To use this space for your own usecase, follow the simple steps below:\n"
22
- "1. [Duplicate](https://huggingface.co/spaces/chansung/gradio_together_tgi/blob/main/app/main.py?duplicate=true) this space. \n"
23
  "2. Set which LLM you wish to use (i.e. mistralai/Mistral-7B-Instruct-v0.2). \n"
24
- "3. Inside [app/main.py](https://huggingface.co/spaces/chansung/gradio_together_tgi/blob/main/app/main.py), write Gradio application. \n"
25
- "4. (Bonus➕) [app/gen](https://huggingface.co/spaces/chansung/gradio_together_tgi/tree/main/app/gen) provides handy utility functions "
26
- "to aynchronously generate text by interacting with the locally served LLM.",
27
  multimodal=False
28
  )
29
 
@@ -33,7 +31,7 @@ def main(args):
33
  ).launch(auth=(USERNAME, PWD), server_name="0.0.0.0", server_port=args.port)
34
 
35
  if __name__ == "__main__":
36
- parser = argparse.ArgumentParser(description="This is my Gradio app's description")
37
  parser.add_argument("--port", type=int, default=7860, help="Port to expose Gradio app")
38
 
39
  args = parser.parse_args()
 
12
  def main(args):
13
  demo = gr.ChatInterface(
14
  fn=chat,
15
+ examples=["Explain the enteerprise adoption challenges", "How can we identify a fraud transaction?", "Por que os grandes modelos de linguagem de AI halucinam?"],
16
+ title="Chat and LLM server in the same application",
17
+ description="This space is a template that we can duplicate for your own usage. "
18
  "This space let you build LLM powered idea on top of [Gradio](https://www.gradio.app/) "
19
  "and open LLM served locally by [TGI(Text Generation Inference)](https://huggingface.co/docs/text-generation-inference/en/index). "
20
  "Below is a placeholder Gradio ChatInterface for you to try out Mistral-7B backed by the power of TGI's efficiency. \n\n"
21
  "To use this space for your own usecase, follow the simple steps below:\n"
22
+ "1. Duplicate this space. \n"
23
  "2. Set which LLM you wish to use (i.e. mistralai/Mistral-7B-Instruct-v0.2). \n"
24
+ "3. Inside app/main.py write Gradio application. \n",
 
 
25
  multimodal=False
26
  )
27
 
 
31
  ).launch(auth=(USERNAME, PWD), server_name="0.0.0.0", server_port=args.port)
32
 
33
  if __name__ == "__main__":
34
+ parser = argparse.ArgumentParser(description="A MAGIC example by ConceptaTech")
35
  parser.add_argument("--port", type=int, default=7860, help="Port to expose Gradio app")
36
 
37
  args = parser.parse_args()