# -*- coding: utf-8 -*- """ Created on Tue Oct 22 08:19:29 2024 @author: kadenyo """ # chatbot.py # Import necessary modules from langchain_core.prompts import ChatPromptTemplate from langchain_community.llms import Ollama import streamlit as st # Define a prompt template for the chatbot prompt=ChatPromptTemplate.from_messages( [ ("system","You are a helpful assistant. Please response to the questions"), ("user","Question:{question}") ] ) # Set up the Streamlit framework st.title('Langchain Chatbot With LLAMA2 model') # Set the title of the Streamlit app input_text=st.text_input("Ask your question!") # Create a text input field in the Streamlit app # Initialize the Ollama model llm=Ollama(model="llama2") # Create a chain that combines the prompt and the Ollama model chain=prompt|llm # Invoke the chain with the input text and display the output if input_text: st.write(chain.invoke({"question":input_text}))