|
import streamlit as st |
|
from langchain.llms import HuggingFaceHub |
|
|
|
|
|
def generate_answer(query): |
|
llm = HuggingFaceHub( |
|
repo_id = "huggingfaceh4/zephyr-7b-alpha", |
|
model_kwargs={"temperature": 0.5, "max_length": 64,"max_new_tokens":512} |
|
) |
|
prompt = f""" |
|
<|system|> |
|
You are an AI assistant that follows instruction extremely well. Please be truthful and give direct answers |
|
</s> |
|
<|user|> |
|
{query}</s> |
|
<|assistant|> |
|
""" |
|
result = llm.predict(prompt) |
|
return result |
|
|
|
|
|
|
|
st.set_page_config(page_title = "LangChain Demo", page_icon = ":robot:") |
|
st.header("LangChain Demo") |
|
|
|
|
|
|
|
def get_text(): |
|
input_text = st.text_input("You: ", key="input") |
|
return input_text |
|
|
|
|
|
user_input = get_text() |
|
response = generate_answer(user_input) |
|
|
|
submit = st.button("Generate") |
|
|
|
|
|
if submit: |
|
st.subheader("Answer: ") |
|
st.write(response) |
|
|
|
|
|
|