ChatNiti-v1 / app.py
Hotair8914's picture
Create app.py
e9dcad6 verified
raw
history blame
1.12 kB
import gradio as gr
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
# Load pre-trained model and tokenizer
def load_model(model_name):
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)
device = "cuda" if torch.cuda.is_available() else "cpu"
model = model.to(device)
return tokenizer, model, device
# Function to generate chat responses
def chat_with_niti(message, history):
tokenizer, model, device = load_model("facebook/mbart-large-50")
input_ids = tokenizer.encode(message, return_tensors="pt").to(device)
output = model.generate(
input_ids,
max_length=100,
temperature=0.7,
num_return_sequences=1,
pad_token_id=tokenizer.eos_token_id
)
response = tokenizer.decode(output[0], skip_special_tokens=True)
return response
# Create Gradio chat interface
demo = gr.ChatInterface(
fn=chat_with_niti,
title="Niti - Your AI Chatbot",
description="Ask Niti anything in Hindi, Hinglish, or English!"
)
# Launch the interface
demo.launch()