tinyllama / app.py
Prajith04's picture
Update app.py
c926cd5 verified
raw
history blame
581 Bytes
import streamlit as st
from langchain_community.llms import Ollama
# Initialize the language model
llm = Ollama(model="tinyllama")
# Streamlit UI elements
st.title("Language Model Invocation")
st.write("Enter a prompt to get a response from the language model.")
# Text input for prompt
prompt = st.text_input("Enter a prompt:")
# Button to invoke the model
if st.button("Submit"):
if prompt:
# Generate the response
response = llm.invoke(prompt)
st.write("Response:")
st.write(response)
else:
st.write("Please enter a prompt.")