Shriharshan's picture
Update app.py
99c773c
from transformers import VisionEncoderDecoderModel, ViTFeatureExtractor, PreTrainedTokenizerFast
model = VisionEncoderDecoderModel.from_pretrained("nlpconnect/vit-gpt2-image-captioning")
vit_feature_extractor = ViTFeatureExtractor.from_pretrained("google/vit-base-patch16-224-in21k")
tokenizer = PreTrainedTokenizerFast.from_pretrained("distilgpt2")
def vit2distilgpt2(img):
pixel_values = vit_feature_extractor(images=img, return_tensors="pt").pixel_values
encoder_outputs = model.generate(pixel_values.to('cpu'), num_beams=5, num_return_sequences=3)
generated_sentences = tokenizer.batch_decode(encoder_outputs, skip_special_tokens=True)
return generated_sentences
import gradio as gr
inputs = [
gr.inputs.Image(type="pil", label="Original Images")
]
outputs = [
gr.outputs.Textbox(label="Caption 1"),
gr.outputs.Textbox(label="Caption 2"),
gr.outputs.Textbox(label="Caption 3")
]
title = "Image Captioning using ViT + GPT2"
description = "ViT and GPT2 are used to generate Image Caption for the uploaded image. COCO DataSet is used for Training"
examples = [
["Image1.png"],
["Image2.png"],
["Image3.png"]
]
gr.Interface(
vit2distilgpt2,
inputs,
outputs,
title=title,
description=description,
examples=examples,
theme="huggingface",
).launch(debug=True, enable_queue=True)