# #Import the required Libraries # import gradio as gr # import pickle # import pandas as pd # import numpy as np # import transformers # # Load from the Hugging Face model hub # model = transformers.AutoModel.from_pretrained('AmpomahChief/sentiment_analysis_on_covid_tweets') # # Define a function that takes in input and passes it through the model # def predict(inputs): # input_ids = transformers.BertTokenizer.from_pretrained('AmpomahChief/sentiment_analysis_on_covid_tweets').encode(inputs, return_tensors='pt') # output = model(input_ids)[0] # return output # # Create a Gradio interface for the model # interface = gr.Interface(fn=predict, inputs=gr.Textbox(prompt="Input text:"), outputs=gr.Textbox(prompt="Model output:")) # # Launch the interface # interface.launch() import gradio as gr # Creating a gradio app using the inferene API App = gr.Interface.load("huggingface/AmpomahChief/sentiment_analysis_on_covid_tweets", title="COVID 19 tweets sentiment analysis", description ="This is a sentiment analysis on COVID 19 tweets using pretrained model on hugging face", allow_flagging=False, examples=[["Input your text here"]] ) App.launch() # import gradio as gr # from transformers import pipeline # import transformers # Model = transformers.AutoModel.from_pretrained('AmpomahChief/sentiment_analysis_on_covid_tweets') # pipeline = pipeline(task="image-classification", model=Model) # def predict(inputs): # input_ids = transformers.BertTokenizer.from_pretrained('AmpomahChief/sentiment_analysis_on_covid_tweets').encode(inputs, return_tensors='pt') # output = model(input_ids)[0] # return output # # Create a Gradio interface for the model # interface = gr.Interface(fn=predict, inputs=gr.Textbox(prompt="Input text:"), outputs=gr.Textbox(prompt="Model output:")) # # Launch the interface # interface.launch()