Spaces:
Running
Running
File size: 7,459 Bytes
a0c907f eb42e7a 1b37f68 8e71274 bee792c 8e71274 bee792c e0b8a6a bee792c 24c14af 8e71274 a0c907f bde0da6 24c14af bde0da6 24c14af bde0da6 8e71274 eb42e7a 8e71274 bb9374d 8e71274 bb9374d 8e71274 918a154 24c14af 918a154 eb42e7a 918a154 eb42e7a 918a154 ed9ad5e 918a154 eb42e7a ed9ad5e 918a154 ed9ad5e fa0ffc9 49f0f1d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 |
from langchain import PromptTemplate
import os
from dotenv import load_dotenv
load_dotenv()
exp_query = 'Generate top 5 questions that I can ask about this data. Questions should be very precise and short, ideally less than 10 words.'
waitText_initialize = 'Preparing the documents, please wait...'
# initialize_prompt = """Write a short welcome message to the user. Describe the data with a comprehensive overview including short summary.\
# If this data is about a person, mention his name instead of using pronouns. After describing the overview, you should mention top 3 example questions that the user can ask about this data.\
# \n\nYour response should be short and precise. Format of your response should be Summary:\n{Description and Summary} \n\n Example Questions:\n{Example Questions}"""
initialize_prompt = """
Describe an overview of this context including a concise summary. Also generate 3 example questions that can be asked about this data.
"""
user_avatar = 'https://cdn-icons-png.flaticon.com/512/6861/6861326.png'
# user_avatar = None
bot_avatar = 'https://cdn-icons-png.flaticon.com/512/1782/1782384.png'
nustian_exps = ['Tell me about NUSTIAN',
'Who is the NUSTIAN regional lead for Silicon Valley?',
'Tell me details about NUSTIAN coaching program.',
'How can we donate to NUSTIAN fundraiser?',
'Who is the president of NUSTIAN?',
"What are top five missions of NUSTIAN?",
]
stdlQs_rb_info = 'Standalone question is a new rephrased question generated based on your original question and chat history'
stdlQs_rb_choices = ['Retrieve relavant docs using original question, send original question to LLM (Chat history not considered)'\
, 'Retrieve relavant docs using standalone question, send original question to LLM'\
, 'Retrieve relavant docs using standalone question, send standalone question to LLM']
llamaPromptTemplate = """
<s>[INST] <<SYS>>
Use the following pieces of context to answer the question at the end. If you don't know the answer, just say that you don't know, don't try to make up an answer.
<</SYS>>
{context}
Question: {question} [/INST]
"""
promptLlama=PromptTemplate(input_variables=['context', 'question'], template=llamaPromptTemplate)
bam_models_old = sorted(['bigscience/bloom',
'salesforce/codegen2-16b',
'codellama/codellama-34b-instruct',
'tiiuae/falcon-40b',
'ibm/falcon-40b-8lang-instruct',
'google/flan-t5-xl',
'google/flan-t5-xxl',
'google/flan-ul2',
'eleutherai/gpt-neox-20b',
'togethercomputer/gpt-neoxt-chat-base-20b',
'ibm/granite-13b-sft',
'ibm/granite-13b-sft-cft',
'ibm/granite-3b-code-v1',
'meta-llama/llama-2-13b',
'meta-llama/llama-2-13b-chat',
'meta-llama/llama-2-13b-chat-beam',
'meta-llama/llama-2-70b',
'meta-llama/llama-2-70b-chat',
'meta-llama/llama-2-7b',
'meta-llama/llama-2-7b-chat',
'mosaicml/mpt-30b',
'ibm/mpt-7b-instruct',
'bigscience/mt0-xxl',
'bigcode/starcoder',
'google/ul2'])
openAi_models = ['gpt-3.5-turbo (openai)', 'gpt-3.5-turbo-16k (openai)', 'gpt-4 (openai)', 'text-davinci-003 (Legacy - openai)', 'text-curie-001 (Legacy - openai)', 'babbage-002 (openai)']
model_dd_info = 'Make sure your credentials are submitted before changing the model. You can also input any OpenAI model name or Watsonx/BAM model ID.'
OaiDefaultModel = 'gpt-3.5-turbo (openai)'
WxDefaultModel = 'meta-llama/llama-2-70b-chat (watsonx)'
BamDefaultModel = 'meta-llama/llama-2-70b-chat (bam)'
url_tb_info = 'Upto 100 domain webpages will be crawled for each URL. You can also enter online PDF files.'
url_tb_ph = 'https://example.com, https://another.com, https://anyremotedocument.pdf'
md_title_general = """
## Chat with your documents and websites<br>
Step 1) Enter your credentials, and click Submit.<br>
Step 2) Upload your documents and/or enter URLs, then click Load Data.<br>
Step 3) Once data is loaded, click Initialize Chatbot (at the bottom of the page) to start talking to your data.<br>
Your documents should be semantically similar (covering related topics or having the similar meaning) in order to get the best results.
You may also play around with Advanced Settings, like changing the model name and parameters.
"""
md_title_nustian = """
## Chat with NUSTIAN website<br>
Step 1) Submit your credentials.<br>
Step 2) Click Initialize Chatbot to start sending messages.<br>
You may also play around with Advanced Settings, like changing the model name and parameters.
"""
md_title_arslan = """
## Talk to Arslan<br>
Welcome to Arslan Ahmed's Chatbot!<br>
This is LLM-based question-answer application built using Retrieval Augmented Generation (RAG) approach with Langchain, implementing Generative AI technology.\
He has developed this application to help people get quick answers on frequently asked questions and topics, rather than waiting for his personal reply.\
Currently, this chatbot is trained on Arslan's resume and LinkedIn profile, with plans to incorporate additional data in the future.<br><br>
By default, this chatbot is powered by OpenAI's Large Language Model gpt-3.5-turbo. For those interested to explore, there are options under Advanced Settings to change the model and its parameters.
"""
welcomeMsgArslan = """Summary: The document provides a comprehensive overview of Arslan Ahmed\'s professional background and expertise as a data scientist.\
It highlights his experience in various industries and his proficiency in a wide range of data analysis tools and techniques.\
The document also mentions his involvement in research projects, publications, and academic achievements.\
\n\nExample Questions:
1. What are some of the key projects that Arslan has worked on as a data scientist?
2. What tools and technologies did Arslan Ahmed utilize in his data science work at IBM?
3. Tell me about Arslan's educational background.
"""
welcomeMsgDefault = """Hello and welcome! I'm your personal data assistant. Ask me anything about your data and I'll try my best to answer."""
def welcomeMsgUser(user):
return f"""Hi, Welcome to personal chatbot of {user}. I am trained on the documents {user} has provided me. Ask me anything about {user} and I'll try my best to answer."""
gDrFolder=(os.getenv("GDRIVE_FOLDER_URL",'')).replace('?usp=sharing','')
class TtydMode():
def __init__(self, name='', title='', type='', dir=None, files=[], urls=[], vis=False, welMsg='', def_k=4, gDrFolder=''):
self.name = name
self.title = title # markdown title for the top display
self.type = type # userInputDocs, fixedDocs, personalBot
self.inputDir=dir
self.file_list=files
self.url_list=urls
self.gDriveFolder=gDrFolder
self.uiAddDataVis = vis # load data from user - this will be true for type = userInputDocs
self.welcomeMsg = welMsg #welcome msg constant - if not provided LLM will generate it
self.k = def_k # default k docs to retrieve
mode_general = TtydMode(name='general', title=md_title_general, type='userInputDocs', vis=True)
mode_nustian = TtydMode(name='nustian', title=md_title_nustian, type='fixedDocs', urls=['https://nustianusa.org', 'https://nustian.ca'])
mode_arslan = TtydMode(name='arslan', title=md_title_arslan, type='personalBot', welMsg=welcomeMsgArslan, def_k=8, gDrFolder=gDrFolder) |