File size: 2,856 Bytes
cf9ec28
d7cd17a
6cde855
d7cd17a
cf9ec28
 
992da9d
7ea7a0f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
62afa46
 
 
 
 
 
 
 
 
 
 
 
d7cd17a
62afa46
 
 
 
 
 
b2cbbb9
f2d54d3
b2cbbb9
 
 
 
 
 
 
 
 
 
 
 
f2d54d3
2653cec
 
 
 
 
 
 
62afa46
 
a9aba1c
2653cec
 
 
62afa46
 
2653cec
 
 
 
 
 
d7cd17a
2653cec
843623c
62afa46
2653cec
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
from groq import Groq 
import gradio as gr 
import os


client = Groq(
    api_key =os.getenv('api_key_gorq')
)
def response_from_llam3(query):
  messages = [
          {
          "role" : "system",
          "content": "You are an helpul Assistant who has plently of Knowledge on Ayur Veda. If the message is Hi or any greeting say namste how can i assist you "
          },
        { 
          "role": "user",
          "content": "What is the answer to {}".format(query)
          }
      ]
    
  response = client.chat.completions.create(
      messages = messages,
      model = "llama3-70b-8192"

  )
  return response.choices[0].message.content

def response_from_mistral(query):
  messages = [
          {
          "role" : "system",
          "content": "You are an helpul Assistant who has plently of Knowledge on Ayur Veda. If the message is Hi or any greeting say namste how can i assist you "
          },
        { 
          "role": "user",
          "content": "What is the answer to {}".format(query)
          }
      ]
    
  response = client.chat.completions.create(
      messages = messages,
      model = "mixtral-8x7b-32768"

  )
  return response.choices[0].message.content  
# iface = gr.Interface(

#     fn=response_from_llam3,
#     inputs="text",
#     outputs="text",
#        examples=[
#             ['What is importance of fasting according to Ayurveda?'],
#             ['What are the medicinal values of Tusli?'],
#             ['What are the three different doshas?'],
#             ['What is the ideal diet according to ayurveda?']
#             ],
#         cache_examples=False,
#     )
# iface.launch()

def chat_with_models(text):
    llama_response = response_from_llam3(text)
    mistral_response =response_from_mistral(text)
    
    return llama_response, mistral_response


with gr.Blocks() as demo:
    gr.Markdown("<h1>๐Ÿš€ Mistral 7B vs LLama3 8B ๐Ÿฆ™</h1>")
    gr.Markdown("<h3> ๐Ÿ•น๏ธ Type your questions or prompts related to Ayurveda and see how each model responds to the same input ๐Ÿ‘พ </h3>")
    with gr.Row():
        input_text = gr.Textbox(label="Enter your prompt here:", placeholder="Type something...", lines=2)
        submit_button = gr.Button("Submit")
    output_llama = gr.Textbox(label="Llama 3 8B ๐Ÿ‘พ", placeholder="", lines=10, interactive=False)
    output_mistral = gr.Textbox(label="Mistral 7B ๐ŸŒ ", placeholder="", lines=10, interactive=False)
    examples=[
            ['What is importance of fasting according to Ayurveda?'],
            ['What are the medicinal values of Tusli?'],
            ['What are the three different doshas?'],
            ['What is the ideal diet according to ayurveda?']
            ],
    
    submit_button.click(fn=chat_with_models, inputs=input_text, outputs=[output_llama,output_mistral])

if __name__ == "__main__":
    demo.launch()