from llama_index import GPTVectorStoreIndex, SimpleDirectoryReader | |
import os | |
from llama_index.node_parser import SimpleNodeParser | |
from llama_index import StorageContext, load_index_from_storage | |
import gradio as gr | |
import openai | |
os.environ['OPENAI_API_KEY'] = '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' | |
# rebuild storage context | |
storage_context = StorageContext.from_defaults(persist_dir="index_dir") | |
# load index | |
index = load_index_from_storage(storage_context) | |
# strat a search engine | |
query_engine = index.as_query_engine() | |
# APP | |
# def get_model_reply_no_prev_context(question): | |
# response = query_engine.query(question) | |
# final_response = response.response[1:] | |
# return final_response | |
# def get_model_reply_no_prev_context(question): | |
# final_response = question | |
# return final_response | |
# title = "Knowledge Center at Penta Building Group" | |
# description = """ | |
# The program is trained to answer questions based on the documentation of 'Lessons Learned' from previous projects! | |
# """ | |
# article = "Your feedback matters!If you like it, contact me at mgupta70@asu.edu" | |
# gr.Interface( | |
# fn=get_model_reply_no_prev_context, | |
# inputs="textbox", | |
# outputs="text", | |
# title=title, | |
# description=description, | |
# article=article, | |
# examples=[["Which code is to be used while planning a pedestrian walkway?"], ["What is AHJ?"]], live=True | |
# ).launch() | |
import gradio as gr | |
def sketch_recognition(img): | |
pass# Implement your sketch recognition model here... | |
gr.Interface(fn=sketch_recognition, inputs="sketchpad", outputs="label").launch() | |