meta2023 / llama-chat.py
philmui's picture
init
5449492
raw
history blame
886 Bytes
import chainlit as cl
import logging
import sys
from dotenv import find_dotenv, load_dotenv
load_dotenv(find_dotenv())
logging.basicConfig(stream=sys.stdout, level=logging.INFO)
logging.getLogger().addHandler(logging.StreamHandler(stream=sys.stdout))
import llama_index
from llama_index.core import set_global_handler
# set_global_handler("wandb", run_args={"project": "meta-10k"})
# wandb_callback = llama_index.core.global_handler
from .globals import (
DEFAULT_QUESTION1,
DEFAULT_QUESTION2,
gpt35_model,
gpt4_mode
)
@cl.on_message
async def main(message: cl.Message):
# Your custom logic goes here...
# Send a response back to the user
await cl.Message(
content=f"Received: {message.content}",
).send()
@cl.on_chat_start
async def start():
await cl.Message(
content="How can I help you about Meta's 2023 10K?"
).send()