--- title: AdvancedLivePortrait-WebUI emoji: 🚀 colorFrom: red colorTo: yellow sdk: gradio sdk_version: 5.5.0 app_file: app.py pinned: true license: apache-2.0 --- # AdvancedLivePortrait-WebUI Dedicated gradio based WebUI started from [ComfyUI-AdvancedLivePortrait](https://github.com/PowerHouseMan/ComfyUI-AdvancedLivePortrait).
You can edit the facial expression from the image. https://github.com/user-attachments/assets/cce42397-e898-4efb-9de8-735e050a10ce ## Notebook You can try it in Colab - [![colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/jhj0517/AdvancedLivePortrait-WebUI/blob/master/notebooks/advanced_live_portrait_webui.ipynb) # Installation And Running ### Prerequisite 1. `3.9` <= `python` <= `3.12` : https://www.python.org/downloads/release/python-3110/ 2. **(Opitonal, only if you're using Nvidia GPU)** CUDA 12.4 : https://developer.nvidia.com/cuda-12-4-0-download-archive?target_os=Windows 3. (Optional, only needed if you use Video Driven) `FFmpeg`: https://ffmpeg.org/download.html
After installing `FFmpeg`, make sure to add the FFmpeg/bin folder to your **system PATH**! ## Run Locally 1. git clone this repository ``` git clone https://github.com/jhj0517/AdvancedLivePortrait-WebUI.git ``` 2. Install dependencies ( Use `requirements-cpu.txt` if you're not using Nvidia GPU. ) ``` pip install -r requirements.txt ``` 3. Run app ``` python app.py ``` ## Run with PowerShell There're PowerShell scripts for each purpose : [`Install.ps1`](https://github.com/jhj0517/AdvancedLivePortrait-WebUI/blob/master/Install.ps1), [`Start-WebUI.ps1`](https://github.com/jhj0517/AdvancedLivePortrait-WebUI/blob/master/Start-WebUI.ps1) and [`Update.ps1`](https://github.com/jhj0517/AdvancedLivePortrait-WebUI/blob/master/Update.ps1).
They do the same things as above with `venv`, creating, activating `venv` and running the app etc. If you're using Windows, right-click the script and then click on ***Run with PowerShell***. ## Run with Docker 1. git clone this repository ``` git clone https://github.com/jhj0517/AdvancedLivePortrait-WebUI.git ``` 2. Build the image ``` docker compose -f docker/docker-compose.yaml build ``` 3. Run the container ``` docker compose -f docker/docker-compose.yaml up ``` 4. Connect to `http://localhost:7860/` in browser. Update the [`docker-compose.yaml`](https://github.com/jhj0517/AdvancedLivePortrait-WebUI/blob/master/docker/docker-compose.yaml) to match your environment if you're not using an Nvidia GPU. ### 🌐 Translation Any PRs for language translation for [`translation.yaml`](https://github.com/jhj0517/AdvancedLivePortrait-WebUI/blob/master/i18n/translation.yaml) would be greatly appreciated! ## ❤️ Acknowledgement 1. LivePortrait paper comes from ```bibtex @article{guo2024liveportrait, title = {LivePortrait: Efficient Portrait Animation with Stitching and Retargeting Control}, author = {Guo, Jianzhu and Zhang, Dingyun and Liu, Xiaoqiang and Zhong, Zhizhou and Zhang, Yuan and Wan, Pengfei and Zhang, Di}, journal = {arXiv preprint arXiv:2407.03168}, year = {2024} } ``` 2. The models are safetensors that have been converted by kijai. : https://github.com/kijai/ComfyUI-LivePortraitKJ 3. [ultralytics](https://github.com/ultralytics/ultralytics) is used to detect the face. 4. This WebUI is started from [ComfyUI-AdvancedLivePortrait](https://github.com/PowerHouseMan/ComfyUI-AdvancedLivePortrait), various facial expressions like AAA, EEE, Eyebrow, Wink are found by PowerHouseMan. 5. [RealESRGAN](https://github.com/xinntao/Real-ESRGAN) is used for image restoration.