| | from dotenv import find_dotenv, load_dotenv |
| | from transformers import pipeline |
| | import os |
| | import requests |
| | import streamlit as st |
| |
|
| | load_dotenv(find_dotenv()) |
| | HUGGINGFACE_API_TOKEN = os.getenv("HUGGINGFACE_API_TOKEN") |
| |
|
| | pipe = pipeline("image-to-text", model="Salesforce/blip-image-captioning-base") |
| |
|
| | |
| | def img_to_text(url): |
| | text = pipe(url)[0]["generated_text"] |
| | print(text) |
| | return text |
| |
|
| | def text_to_speech(message): |
| | API_URL = "https://api-inference.huggingface.co/models/espnet/kan-bayashi_ljspeech_vits" |
| | headers = {"Authorization": f"Bearer {HUGGINGFACE_API_TOKEN}"} |
| | payloads = { |
| | "inputs":message |
| | } |
| |
|
| | response = requests.post(API_URL, headers=headers, json=payloads) |
| | with open('audio.flac', 'wb') as file: |
| | file.write(response.content) |
| | |
| | def main(): |
| | st.set_page_config(page_title="Image to Text", page_icon="🎙️") |
| |
|
| | st.header("Image to Text") |
| | |
| | image = "narrator.jpeg" |
| | left_co, cent_co, last_co = st.columns(3) |
| | with cent_co: |
| | st.image(image=image) |
| | uploaded_file = st.file_uploader("Choose an image: ", type=["jpg", "jpeg", "png"]) |
| |
|
| | if uploaded_file is not None: |
| | print(uploaded_file) |
| | bytes_data = uploaded_file.getvalue() |
| | with open(uploaded_file.name, "wb") as file: |
| | file.write(bytes_data) |
| | st.image(uploaded_file, caption='Uploaded image', use_column_width=True) |
| | scenario=img_to_text(uploaded_file.name) |
| | text_to_speech(scenario) |
| |
|
| | with st.expander("scenatio"): |
| | st.write(scenario) |
| | |
| | st.audio("audio.flac") |
| |
|
| |
|
| | if __name__== "__main__": |
| | main() |