#!/usr/bin/env python from __future__ import annotations import os import gradio as gr import torch from app_inference import create_inference_demo from app_training import create_training_demo from app_upload import create_upload_demo from inference import InferencePipeline from trainer import Trainer TITLE = '# [Tune-A-Video](https://tuneavideo.github.io/) Training UI' ORIGINAL_SPACE_ID = 'Tune-A-Video-library/Tune-A-Video-Training-UI' SPACE_ID = os.getenv('SPACE_ID', ORIGINAL_SPACE_ID) SHARED_UI_WARNING = f'''# Attention - This Space doesn't work in this shared UI. You can duplicate and use it with a paid private T4 GPU. (Please note that there seems to be an issue with training on the A10G GPU now. The model doesn't learn anything when trained on A10G. Training on T4 works perfectly fine and inference works fine on both.)
Duplicate Space
''' if os.getenv('SYSTEM') == 'spaces' and SPACE_ID != ORIGINAL_SPACE_ID: SETTINGS = f'Settings' else: SETTINGS = 'Settings' CUDA_NOT_AVAILABLE_WARNING = f'''# Attention - Running on CPU.
You can assign a GPU in the {SETTINGS} tab if you are running this on HF Spaces. You can use "T4 small/medium" or "A10G small/large" to run this demo.
''' HF_TOKEN_NOT_SPECIFIED_WARNING = f'''# Attention - The environment variable `HF_TOKEN` is not specified. Please specify your Hugging Face token with write permission as the value of it.
You can check and create your Hugging Face tokens here. You can specify environment variables in the "Repository secrets" section of the {SETTINGS} tab.
''' HF_TOKEN = os.getenv('HF_TOKEN') def show_warning(warning_text: str) -> gr.Blocks: with gr.Blocks() as demo: with gr.Box(): gr.Markdown(warning_text) return demo pipe = InferencePipeline(HF_TOKEN) trainer = Trainer(HF_TOKEN) with gr.Blocks(css='style.css') as demo: if SPACE_ID != ORIGINAL_SPACE_ID: show_warning(SHARED_UI_WARNING) elif not torch.cuda.is_available(): show_warning(CUDA_NOT_AVAILABLE_WARNING) gr.Markdown(TITLE) with gr.Tabs(): with gr.TabItem('Train'): create_training_demo(trainer, pipe) with gr.TabItem('Test'): create_inference_demo(pipe, HF_TOKEN) with gr.TabItem('Upload'): gr.Markdown(''' - You can use this tab to upload models later if you choose not to upload models in training time or if upload in training time failed. ''') create_upload_demo(HF_TOKEN) if not HF_TOKEN: show_warning(HF_TOKEN_NOT_SPECIFIED_WARNING) demo.queue(max_size=1).launch(share=False)