llm / app.py
tommy24's picture
Update app.py
172da29
raw
history blame
3.11 kB
# import gradio as gr
# from langchain.llms import LlamaCpp
# from langchain import PromptTemplate, LLMChain
# from langchain.llms import GPT4All
# from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
# # import requests
# # url = "https://huggingface.co/TheBloke/Nous-Hermes-13B-GGML/resolve/main/nous-hermes-13b.ggmlv3.q4_0.bin"
# # response = requests.get(url)
# # with open("nous-hermes-13b.ggmlv3.q4_0.bin", "wb") as f:
# # f.write(response.content)
# print("DONE")
# def func(user):
# template = """
# Your name is John and not a assistant but more like a chatbot. Respond precise not more words and act like a human. for example: user: How are you? You: I'm doing good how about you?. user: hello You: Hello how you doing?. Don't say How can I assist you today?.
# Question: {question}
# Answer: """
# prompt = PromptTemplate(template=template, input_variables=["question"])
# local_path = (
# "./nous-hermes-13b.ggmlv3.q4_0.bin"
# )
# # # Callbacks support token-wise streaming
# # callbacks = [StreamingStdOutCallbackHandler()]
# # Verbose is required to pass to the callback manager
# llm = LlamaCpp(model_path="./nous-hermes-13b.ggmlv3.q4_0.bin", n_ctx=2048)
# llm_chain = LLMChain(prompt=prompt, llm=llm)
# question = user
# llm_chain.run(question)
# return llm_chain.run(question)
# iface = gr.Interface(fn=func, inputs="text", outputs="text")
# iface.launch()
# import gradio as gr
# from langchain.llms import LlamaCpp
# from langchain import PromptTemplate, LLMChain
# from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler
# print("DONE")
# def func(user):
# template = """
# Your name is John and not a assistant but more like a chatbot. Respond precise not more words and act like a human. for example: user: How are you? You: I'm doing good how about you?. user: hello You: Hello how you doing?. Don't say How can I assist you today?.
# Question: {question}
# Answer: """
# prompt = PromptTemplate(template=template, input_variables=["question"])
# local_path = "./nous-hermes-13b.ggmlv3.q4_0.bin"
# llm = LlamaCpp(model_path=local_path)
# llm_chain = LLMChain(prompt=prompt, llm=llm, streaming=True) # Enable streaming mode
# question = user
# llm_chain.run(question)
# return llm_chain.run(question)
# iface = gr.Interface(fn=func, inputs="text", outputs="text")
# iface.launch()
import gradio as gr
from gpt4allj import Model
# Load the local model
model = Model('./ggml-gpt4all-j-v1.3-groovy.bin')
# Define a function that generates the model's response given a prompt
def generate_response(prompt):
response = model.generate(prompt)
return response
# Create a Gradio interface with a text input and an output text box
iface = gr.Interface(
fn=generate_response,
inputs="text",
outputs="text",
title="GPT-4 AllJ",
description="Generate responses using GPT-4 AllJ model."
)
# Run the Gradio interface
iface.launch()