Spaces:
Sleeping
Sleeping
File size: 4,444 Bytes
dcc1634 6f2d9e1 9a8052a 6f2d9e1 9a8052a 6f2d9e1 ab52ca3 6f2d9e1 ab52ca3 dcc1634 6f2d9e1 dcc1634 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 |
from typing import Set
from backend.core import run_llm
import streamlit as st
from streamlit_chat import message
from PIL import Image
from io import BytesIO
import base64
# def add_bg_from_local(image_file):
# with open(image_file, "rb") as image_file:
# encoded_string = base64.b64encode(image_file.read())
# st.markdown(
# f"""
# <style>
# .stApp {{
# background-image: url(data:{"jpeg"};base64,{encoded_string.decode()});
# background-size: cover
# }}
# </style>
# """,
# unsafe_allow_html=True
# )
# background_image = "bg2.jpeg"
# add_bg_from_local(background_image)
st.header("LangChain π¦π Documentation - Helper ChatBot")
# profile_image = Image.open("langchain background removed.png")
# st.sidebar.image(profile_image, width=200)
st.markdown(
"""
<style>
[data-testid=stSidebar] [data-testid=stImage]{
text-align: center;
display: block;
margin-left: auto;
margin-right: auto;
width: 100%;
}
</style>
""", unsafe_allow_html=True
)
with st.sidebar:
st.image("langchain background removed.png", width = 200)
st.sidebar.markdown(
"""
# LangChain Helper Bot π€
Welcome to the LangChain Helper Bot!
**What does it offer ?**
π Easy scraping of LangChain documentation.
π Instead of going through tedious LangChain documentation, just submit your query here about any LangChain topic, and you'll get the answer!
---
**How to use ?**
Just type your LangChain-related query in the input box on the main page, and the bot will fetch precise information for you.
"""
)
with st.sidebar:
st.caption(
"""The official documentation of langchain can be found at
https://python.langchain.com/docs/get_started
"""
)
st.markdown("---")
st.markdown(
'<h6>Made in  <img src="https://streamlit.io/images/brand/streamlit-mark-color.png" alt="Streamlit logo" height="16">  by <a href="https://www.linkedin.com/in/anujmaha/">@anujmaha</a></h6>',
unsafe_allow_html=True,
)
st.markdown(
'π§ : anujsmahajan1998@gmail.com',
)
st.markdown(
'<div style="margin-top: 0.75em;"><img src="https://cdn.buymeacoffee.com/buttons/default-orange.png" alt="Buy Me A Coffee" height="41" width="174"></a></div>',
unsafe_allow_html=True,
)
if "user_prompt_history" not in st.session_state:
st.session_state["user_prompt_history"] = []
if "chat_answers_history" not in st.session_state:
st.session_state["chat_answers_history"] = []
if "chat_history" not in st.session_state:
st.session_state["chat_history"] = []
def create_sources_string(source_urls: Set[str]) -> str:
if not source_urls:
return ""
sources_list = list(source_urls)
sources_list.sort()
sources_string = "sources:\n"
for i, source in enumerate(sources_list):
sources_string += f"{i+1}. {source}\n"
return sources_string
prompt = st.text_input("Prompt", placeholder="Enter your Langchain related query here... (eg, what is ConversationalRetrievalChain ?)")
if prompt:
with st.spinner("Generating response..."):
generated_response = run_llm(
query=prompt, chat_history=st.session_state["chat_history"]
)
sources = set(
[doc.metadata["source"] for doc in generated_response["source_documents"]]
)
formatted_response = (
f"{generated_response['answer']} \n\n {create_sources_string(sources)}"
)
st.session_state.user_prompt_history.append(prompt)
st.session_state.chat_answers_history.append(formatted_response)
st.session_state.chat_history.append((prompt, generated_response["answer"]))
if st.session_state["chat_answers_history"]:
for generated_response, user_query in zip(
st.session_state["chat_answers_history"],
st.session_state["user_prompt_history"],
):
message(
user_query,
is_user=True,
avatar_style="adventurer",
seed=123,
)
# message(generated_response)
st.write(
f'<div style="word-wrap: break-word;">{generated_response}</div>',
unsafe_allow_html=True,
)
|